You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to preliminary investigate the deployment of our model with localai
So that we get familiar with localai and we can already identify possible problems.
Acceptance criteria
Key documentation of how to integrate our model with localai is identified and consulted
Identification of possible problems of the integration of the model into localai
Possible problems with the integration of our model into localai are identified
Possible problems are summarized
Possible problems are communicated to the POs so that they can be tackled with issues in the next sprint
A suggestion is made of how we should set up a frontend to work with localai (either own implementation or usage of an example chatbot with permissive license that may exist)
Further information
Our industry partner is currently setting up a cluster in which localai is deployed. So localai will run on the target infrastructure where we should deploy our model and we should make our model accessible using the API of localai
We will deploy our using localAI on a Kubermatic cluster.
Findings
It is entirely possible to deploy localAI on a Kubernetes cluster (see: https://localai.io/basics/kubernetes/). It is unclear however if this also works smoothly with Kubermatic, This will require further testing on the Hetzner machine.
It also supports GGUF models ("[...] Runs gguf, transformers, diffusers and many more models architectures", https://github.com/mudler/LocalAI).
Conclusion
It will now be important to implement localAI on a Kubermatic cluster. If this succeeds, the GUI for interacting with the LLM should be easily available.
User story
Acceptance criteria
Further information
Definition of done (DoD)
DoD general criteria
The text was updated successfully, but these errors were encountered: