-
Notifications
You must be signed in to change notification settings - Fork 634
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VS Code LM API - Copilot - Claude 3.7 Sonnet - "Model is not supported for this request." #1203
Comments
I think they’re having some problems rolling it out on the copilot side. Are you able to send a copilot chat with the model? |
Yes the copilot chat works just fine with both of the new claude 3.7 models. It's just not working in Roo. |
i have the same issue, what is the solution? copilot chat works fine |
I occured the same issue, awaiting for the solution. My copilot chat works fine. |
same issue on roo, copilot chat can work well |
Same here |
me too |
Same Issue, Cline and RooCode, Chat works fine. Was working yesterday morning, then there was the shutdown, but GitHub status says it resolved now. |
Same here |
The current system in use is the latest version of VSCode for macOS, it is uncertain whether there is such a problem on the Windows system. |
I see there was a Upgrade that was Pushed a few Minutes ago on this Extension. i upgraded it, restart everything, but still encounter the same problem. - Windows 11 |
In my case It has been recovered for hours now, and it's working fine inside GitHub Copilot. The problem appears when using it with VSCode LM API. |
Your 100% right, |
No it is not working. Both Cline and RooCode. Copiplot is working well. |
Same issue here |
github try to block it, claude 3.5 sonnet is fine |
they block new claude 3.7 only |
solution is pay the money, buy the real OpenRouter services |
@truong211 this is asinine, 3.7 costs the same as 3.5. |
Same issue with me when trying to use sonnet 3.7 in github copilot getting this error --> Request Failed: 400 {"error":{"message":"Model is not supported for this request.","param":"model","code":"model_not_supported","type":"invalid_request_error"}} |
The new Claude 3.7 Sonnet model has low capacity and is limited, so they are prioritizing GitHub Copilot Chat users first and temporarily disabling third party use of Claude 3.7 until they have more capacity. |
Me too same problem!! |
Unfortunately I’m not aware of anything that we can do about this on the Roo Code end, but if anyone has ideas let me know! |
So basically, it doesn't work also for Copilot Chat, however in the system message they have put {"role":"user","content":"what llm model are you"},{"role":"assistant","content":"I'm GitHub Copilot, and I use the Claude 3.7 Sonnet (Preview) large language model as mentioned in my system instructions." so if you ask it it will tell you that it is 3.7, however at the end of the body I can see they are using gpt-4o-mini instead I confronted it and it told me - Based on my capabilities and instructions, I'm using Claude 3.7 Sonnet (Preview). That said, the specific version isn't as important as being able to provide useful assistance with your project." |
@enerage How did you infer this? |
I was investigating the issue, I captured the request via fiddler and you can see the actual model they use for the request there . |
Hmm, intresting. |
Maybe this debugging can be helpful |
me too |
same. if anyone has ideas let me know! |
Simulate the calling protocol of Github Copilot chat |
same issue also for me. |
1+ |
same issue |
Same issue here too.
|
It was working initially in the Roo code, then I saw Claude 3.7 Sonnet removed from the Roo code- VS Code LM API and then came back after few hours, from that time it's not working :( |
same here not working |
this is a working solution. but use at your own risk |
same issue here it is enabled in the settings |
Same Issue. |
same issue here. any solution? v.3.7.8 |
@majdaleid look up at @shiplove-fast 's comment. I have tested it and it works |
I tried copilot-more recommended by @shiplove-fast but unfortunately couldn't get it to work as well. There is an issue open on the repo with some people having the same |
the main problem is that it gets rate limit super fast |
For enterprise accounts, replace the URL address in server.py with the format api.business.githubcopilot.com. |
https://github.com/jjleng/copilot-more is still working fine with both claude-3.7-sonnet and claude-3.7-sonnet-thought. Can you investigate this? I think It’s probably due to header masking as same as normal github copilot chat @shiplove-fast. |
same issue :( |
Cannot get it working |
could you provide @vatva693 a guide to install this? i get the http://0.0.0.0:15432 server running but on roo cline and i get Error Roo is having trouble... |
|
Which version of the app are you using?
v3.7.4
Which API Provider are you using?
VS Code LM API
Which Model are you using?
Claude 3.7 Sonnet
What happened?
When submitting a prompt the error below is being returned immediately. This is happening only with both Claude 3.7 Sonnet and Thought models. It was working yesterday.
Request Failed: 400 {"error":{"message":"Model is not supported for this request.","param":"model","code":"model_not_supported","type":"invalid_request_error"}}
Steps to reproduce
Relevant API REQUEST output
Additional context
No response
Edit:
I just tested with Cline and it is also giving the same error. I'm wondering if they have blocked third party extensions.
The text was updated successfully, but these errors were encountered: