You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
but I wanted to make sure if this was a bug. I did make a pr for it if the way the function is called is correct. For example, if we just directly put the keywords in to llamacpp instead of doing model_kwargs= it'll work but that'll be different from the transformers api.
Describe the issue as clearly as possible:
I found that when we do
similar to the documentation, the contents of model_kwargs get ignored. The reason I think is here
where model_kwargs becomes a dict with key model_kwargs. To fix this I just did
but I wanted to make sure if this was a bug. I did make a pr for it if the way the function is called is correct. For example, if we just directly put the keywords in to llamacpp instead of doing model_kwargs= it'll work but that'll be different from the transformers api.
Steps/code to reproduce the bug:
Expected result:
llm will be initialized with 2048 context length but instead it's 512 context length as the model_kwargs do not get set in model
Error message:
No response
Outlines/Python version information:
Version information
latest
Context for the issue:
No response
The text was updated successfully, but these errors were encountered: