Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM inference - Openai Chat Api or Llama compatible chat params #5558

Open
raymon-io opened this issue Aug 4, 2024 · 0 comments
Open

LLM inference - Openai Chat Api or Llama compatible chat params #5558

raymon-io opened this issue Aug 4, 2024 · 0 comments
Assignees
Labels
stat:awaiting googler Waiting for Google Engineer's Response task:LLM inference Issues related to MediaPipe LLM Inference Gen AI setup type:feature Enhancement in the New Functionality or Request for a New Solution

Comments

@raymon-io
Copy link

raymon-io commented Aug 4, 2024

Describe the feature and the current behaviour/state

Does GenAi Llm inference support openai chat api or support taking parameters similar to Llama models? If not then I am requesting implementing the api similar to Openai Chat completion.

Please specify the use cases for this feature

For example having a stop param to stop the generation of the llm inference would be great. And Llama compatible chat format can be useful to integrate with other frameworks such as Langchain.

Any Other info

There is a mention of Llama in a comment in llm.h although I am not sure what this is about.

@raymon-io raymon-io added the type:feature Enhancement in the New Functionality or Request for a New Solution label Aug 4, 2024
@kuaashish kuaashish assigned kuaashish and unassigned ayushgdev Aug 5, 2024
@kuaashish kuaashish added the task:LLM inference Issues related to MediaPipe LLM Inference Gen AI setup label Aug 5, 2024
@kuaashish kuaashish added the stat:awaiting googler Waiting for Google Engineer's Response label Aug 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
stat:awaiting googler Waiting for Google Engineer's Response task:LLM inference Issues related to MediaPipe LLM Inference Gen AI setup type:feature Enhancement in the New Functionality or Request for a New Solution
Projects
None yet
Development

No branches or pull requests

4 participants