-
Notifications
You must be signed in to change notification settings - Fork 810
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
deepseek-r1:32b or 14b almost cannot work in roo cline, because it never calls the tools leading to continuous errors #28
Comments
I also encountered this issue, which might be a problem with the model's capability. I once called qwen2-coder-32b, and it worked fine in roocline. However, deepseek-r1:32b would report an error after multiple repeated invalid conversations. |
I could not make it work even with the deepseek-r1:70b |
I have the same issue using cline and the R1-32B model locally on ollama using ollama published R1 model. |
For me the issue was context length, the default context length is around 4K for Ollama. You can fix this by creating a Modelfile with the following:
and then create a model from this with a unique name using ollama. Choose this model in cline and it should start working. You can also use LM studio which allows to increase context length by holding alt over the model before loading it. These two approaches will make it work. |
Yes, thank you, it worked and became much better. For 14B I set num_ctx 25000 and it is able to complete the simple task, albeit after a long time. R1 models seem to not understand that they need to work with an open active file and replace 1 variable there (in "Code" mode). But there are still certain technical shortcomings in roo cline:
Oh, I noticed that this is not a repo of roo cline, but of the creator of the model. |
deepseek-r1:32b or 14b almost cannot work in roo cline, because it never calls the tools leading to continuous errors
The text was updated successfully, but these errors were encountered: