-
Notifications
You must be signed in to change notification settings - Fork 882
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Running with TransformersModel
does not work
#414
Comments
Hi @aymeric-roucher 👋 |
TransformersModel
does not workTransformersModel
does not work
hi @aymeric-roucher @danielkorat 👋 I'm also facing this issue when trying the Code to reproduce the error I'm currently trying to build a LlamaCppModels class to allow users to work with llama.cpp models via llama-cpp-python. If I find a fix for |
I've seen such code parsing errors before while using As a side note, it'd be great to have a |
Thank you folks for reporting! |
FYI I tried that but it still did not solve the issue |
I encountered the same issue today. After a little digging, the problem seems to be that Explicitly passing Happy to submit a small pull request if needed. |
Describe the bug
When replacing
HfApiModel
withTransformersModel
inexamples/benchmark.ipynb
, the eval results formeta-llama/Llama-3.1-8B-Instruct
(and various other published models) are far worse than published (scores of less than 5).Code to reproduce the error
https://github.com/danielkorat/smolagents/blob/transformers/examples/benchmark-transformers.ipynb
Error logs (if any)
Seems like a big part of problem is the parsing of the LLM output (specifically the assistant role):
Also, the regex parsing error arises in nearly all examples.
Expected behavior
Trying to reproduce the results for
meta-llama/Llama-3.1-8B-Instruct
, as published in the original notebook:Packages version:
Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered: