Skip to content

A simple telegram chat bot for LLMs supported by llama.cpp

License

Notifications You must be signed in to change notification settings

apbodrov/llama-telegram-bot

 
 

Repository files navigation

Llama.cpp Telegram Bot

A simple Telegram chat bot for LLMs supported by llama.cpp

Easiest way to share your selfhosted ChatGPT style interface with friends and family! Even group chat with your AI friend!

Demo

Demo.mp4

Features

  1. Easy to setup and run
  2. Group chat
  3. Whitelist to restrict to a limited set of users
  4. Unlimited chats - Sliding window
  5. Typing indicator
  6. Streaming responses
  7. New chat command
  8. Supports LLMs supported by llama.cpp
  9. GPU Acceleration support
  10. Voice chat

Installation

What do you need?

  1. A free Telegram bot token from @BotFather
  2. A Llama.cpp supported model in your local system

Manual Installation

$ git clone https://github.com/aneeshjoy/llama-telegram-bot.git
$ cd llama-telegram-bot
$ sudo pip3 install -r requirements.txt
$ export BOT_TOKEN=<Your Telegram bot token>
$ export MODEL_PATH=/path/to/your/model/file
$ python3 bot.py

GPU Acceleration

For GPU acceleration specify additional environment variables:

LLAMA_CUBLAS=1 CMAKE_ARGS="-DLLAMA_CUBLAS=ON" FORCE_CMAKE=1 pip3 install -r requirements.txt

Using Prebuilt Docker image

You can use pre-built docker image to run the bot. Follow these steps:

  1. Clone the repo or download docker-compose.yml
$ git clone https://github.com/aneeshjoy/llama-telegram-bot.git
$ cd llama-telegram-bot
  1. Update BOT_TOKEN env variable in docker-compose.yml file
  2. Update MODEL_PATH env variable in docker-compose.yml file
  3. Update - /path/to/models in docker-compose.yml to point to the folder which contains the model
  4. Run the bot
$ docker-compose up

Using Self built Docker image

You can build your own docker image to run the bot. Follow these steps:

  1. Clone the repo
$ git clone https://github.com/aneeshjoy/llama-telegram-bot.git
$ cd llama-telegram-bot
  1. Update BOT_TOKEN env variable in docker-compose-build.yml file
  2. Update MODEL_PATH env variable in docker-compose-build.yml file
  3. Update - /path/to/models in docker-compose-build.yml to point to the folder which contains the model
  4. Run the bot
$ docker-compose -f docker-compose-build.yml up --build 

Usage

Once the bot is running, you can use the following commands to generate text:

  • /start - Get a welcome message from the bot.
  • /new_chat - To start a new conversation.

Contributing

To contribute, please:

  1. Fork the repository.
  2. Create a new branch for your changes.
  3. Make your changes and commit them.
  4. Push your changes to your fork.
  5. Submit a pull request.

License

This project is licensed under the MIT License. See the LICENSE file for more information.

Credits

About

A simple telegram chat bot for LLMs supported by llama.cpp

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 95.5%
  • Dockerfile 4.5%