-
Notifications
You must be signed in to change notification settings - Fork 10.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llava 1.6 different responses in CLI and Server #5514
Comments
I've focused on providing the required API and functionality into llava.cpp and clip.cpp, llava-cli.cpp was used as demo tool. From server.cpp
So it correctly processes the image, it will receive back multiple image embeddings but only uses the first one (like llava-1.5) That's what server.cpp should be using, otherwise it needs to have duplicate code for each architecture. I'm sure someone can implement that quickly, an example on using that function is in |
The llava-cli has the llama system prompt hardcoded. Try adding that as the system prompt to the server (with the API call, the command line parameter doesn't work right now) |
should be fixed in the pr which was merged |
I gave the same image to llava-cli and llava hosted on the server. At temperature 0 and no other parameters both of them are giving different results (with the same prompt and also without any prompts). Is this intentional? If yes, then how can I get the same output as the cli from the server? CLI seems to consistently give better outputs than the server.
The text was updated successfully, but these errors were encountered: