Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Preliminary investigation of deployment with localai #103

Closed
dominic0df opened this issue Jun 26, 2024 · 1 comment
Closed

Preliminary investigation of deployment with localai #103

dominic0df opened this issue Jun 26, 2024 · 1 comment
Assignees
Labels

Comments

@dominic0df
Copy link
Contributor

dominic0df commented Jun 26, 2024

User story

  1. As a software architect
  2. I want to preliminary investigate the deployment of our model with localai
  3. So that we get familiar with localai and we can already identify possible problems.

Acceptance criteria

  • Key documentation of how to integrate our model with localai is identified and consulted
  • Identification of possible problems of the integration of the model into localai
    • Possible problems with the integration of our model into localai are identified
    • Possible problems are summarized
    • Possible problems are communicated to the POs so that they can be tackled with issues in the next sprint
  • A suggestion is made of how we should set up a frontend to work with localai (either own implementation or usage of an example chatbot with permissive license that may exist)

Further information

  • Our industry partner is currently setting up a cluster in which localai is deployed. So localai will run on the target infrastructure where we should deploy our model and we should make our model accessible using the API of localai
  • see https://localai.io/

Definition of done (DoD)

  • Bill of Materials in the planning document has been updated
  • All feature branches have been merged and closed
  • New feature code has been documented
  • Potential new licenses have been checked
  • All GitHub Actions are passing
  • The requirement.txt is updated

DoD general criteria

  • Feature has been fully implemented
  • Feature has been merged into the mainline
  • All acceptance criteria were met
  • Product owner approved features
  • All tests are passing
  • Developers agreed to release
@dominic0df dominic0df moved this to Product Backlog in amos2024ss08-feature-board Jun 26, 2024
@grayJiaaoLi grayJiaaoLi added User Story Label for User Stories SP 03 labels Jun 26, 2024
@grayJiaaoLi grayJiaaoLi moved this from Product Backlog to Sprint Backlog in amos2024ss08-feature-board Jun 26, 2024
@christianwielenberg christianwielenberg moved this from Sprint Backlog to In progress in amos2024ss08-feature-board Jul 3, 2024
@dnsch
Copy link
Contributor

dnsch commented Jul 3, 2024

Report on LocalAI

We will deploy our using localAI on a Kubermatic cluster.

Findings

  1. It is entirely possible to deploy localAI on a Kubernetes cluster (see: https://localai.io/basics/kubernetes/). It is unclear however if this also works smoothly with Kubermatic, This will require further testing on the Hetzner machine.
  2. It's also possible to use a GUI for interacting with the LLM (see: feat(ux): Add chat, tts, and image-gen pages to the WebUI mudler/LocalAI#2222).
  3. It also supports GGUF models ("[...] Runs gguf, transformers, diffusers and many more models architectures", https://github.com/mudler/LocalAI).

Conclusion

It will now be important to implement localAI on a Kubermatic cluster. If this succeeds, the GUI for interacting with the LLM should be easily available.

@dominic0df dominic0df moved this from In progress to Feature Archive in amos2024ss08-feature-board Jul 3, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
Status: Feature Archive
Development

No branches or pull requests

4 participants