Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

doc:llm deploy docs #692

Merged
merged 1 commit into from
Oct 19, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Binary file added docs/getting_started/application/chatdb/img.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
98 changes: 98 additions & 0 deletions docs/getting_started/install/cluster/vms/standalone.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,98 @@
Standalone Deployment
==================================
(standalone-index)=

### Install Prepare
```commandline
git clone https://github.com/eosphoros-ai/DB-GPT.git

cd DB-GPT
```

### Create conda environment
```commandline
conda create -n dbgpt_env python=3.10

conda activate dbgpt_env
```


### Install Default Requirements
```commandline
# Install Default Requirements
pip install -e ".[default]"
```

### Download and Prepare LLM Model and Embedding Model
```{tip}
If you don't have high performance hardware server
```
you can use openai api, tongyi api , bard api, etc.
```commandline
mkdir models && cd models

# download embedding model, eg: text2vec-large-chinese
git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese

```

set proxy api in .env
```commandline
#set LLM_MODEL TYPE
LLM_MODEL=proxyllm
#set your Proxy Api key and Proxy Server url
PROXY_API_KEY={your-openai-sk}
PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions
```
```{tip}
If you have high performance hardware server
```

```commandline
mkdir models && cd models

# # download embedding model, eg: vicuna-13b-v1.5 or
git clone https://huggingface.co/lmsys/vicuna-13b-v1.5

# download embedding model, eg: text2vec-large-chinese
git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese

popd
```
### Start all services with a single command.
```commandline
LLM_MODEL=vicuna-13b-v1.5
dbgpt start webserver --port 6006
```
By default, the "dbgpt start webserver" command will start the Webserver, Model Controller, and Model Worker in a single Python process. Here, we specify the service to be started on port 6006.

### View and validate the model service in the command line, you can use the following commands
##### 1.list the started model services and deployed Model Workers, you can use the following command
```commandline
dbgpt model list
```
output is:
```commandline
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
| Model Name | Model Type | Host | Port | Healthy | Enabled | Prompt Template | Last Heartbeat |
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
| vicuna-13b-v1.5 | llm | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.201313 |
| WorkerManager | service | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.246756 |
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
```
The WorkerManager is the management process for Model Workers

##### validate the deployed model in the command line, you can use the following command
```commandline
dbgpt model chat --model_name vicuna-13b-v1.5
```
Then an interactive page will be launched where you can have a conversation with the deployed LLM in the terminal.
```commandline
Chatbot started with model vicuna-13b-v1.5. Type 'exit' to leave the chat.


You: Hello
Bot: Hello! How can I assist you today?

You:
```
Original file line number Diff line number Diff line change
@@ -0,0 +1,123 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2023, csunny
# This file is distributed under the same license as the DB-GPT package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2023.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: DB-GPT 👏👏 0.3.9\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2023-10-17 17:24+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language: zh_CN\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.12.1\n"

#: ../../getting_started/application/chatagent/chatagent.md:1
#: ../../getting_started/application/chatagent/chatagent.md:3
#: 17aae18073594c7498c8b2434d50b218 77f0177e55e144a89e3bcf735b0d52de
msgid "ChatAgent"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:3
#: 41b42808938245dc946960571c09026a
msgid ""
"ChatAgent can automatically select the right tool to answer and solve "
"user problems through LLM. ![ChatAgent](https://github.com/eosphoros-ai"
"/DB-GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:6
#: bb76bb44442647a091d008c9e1427e4a
msgid "1.Update Plugins From Git"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:7
#: a7f188c3de7e4b85b2d3079128b535b2
msgid ""
"Update the latest plug-in information from the DB-GPT official plug-in "
"repository. ![Update From Github](https://github.com/eosphoros-ai/DB-"
"GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:7
#: 612cb75edb2f48a087fc99e6d0890b80
msgid "Update From Github"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:10
#: 1ad3f7680837410cb8018fff5787366f
msgid "2.Install Plugin"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:11
#: b32e9318a8e347339b6c4cad4b966880
msgid ""
"Install the plugin you want to use from the list of agent hub plugins. "
"![Install Plugin](https://github.com/eosphoros-ai/DB-"
"GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:11
#: bee96f7936564bc6be98655bfc6aae0d
msgid "Install Plugin"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:15
#: 20d562b9ed004336bb6df8d163554924
msgid "3.Upload My Local Plugin File"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:16
#: b43e485ff4f042dd80b5950444c0db08
msgid ""
"If you want to use a private plug-in developed by yourself, you can "
"upload the plug-in first and then use it. ![Upload Private "
"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:16
#: a05724784c0e45528039b8238fd9ed3c
msgid "Upload Private Plugin"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:19
#: 726b5c4cad7d4be9815e9fbeedaecc2b
msgid "4.Select installed plug-ins dialog"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:20
#: e3d9dc892d054cbca29267cd0fffe7be
msgid ""
"Select an installed plugin to have a conversation. ![Select "
"Plugins](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:20
#: 97f28bacc02444c580f6481b21e6c155
msgid "Select Plugins"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:23
#: d0272e7d94d743289fb2c1613ee85c36
msgid "5.Uninstall My Plugin"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:24
#: 2fee213a12da4674b19fe8891e4318f5
msgid ""
"Uninstall plugins you no longer need. ![Uninstall "
"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""

#: ../../getting_started/application/chatagent/chatagent.md:24
#: 32e5405249134863bcaf85b0ef91260f
msgid "Uninstall Plugin"
msgstr ""

Original file line number Diff line number Diff line change
@@ -0,0 +1,117 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2023, csunny
# This file is distributed under the same license as the DB-GPT package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2023.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: DB-GPT 👏👏 0.3.9\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2023-10-17 17:24+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language: zh_CN\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.12.1\n"

#: ../../getting_started/install/cluster/vms/standalone.md:1
#: 40af95d327574e5ea9a6c29f68a7709c
msgid "Standalone Deployment"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:5
#: 7ef23971a0d2415a857d5eb9c4d955ac
msgid "Install Prepare"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:12
#: f3cd2941fa2d49d686fc246237f82e68
msgid "Create conda environment"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:20
#: 589d80d5d97044ee93747091904583c5
msgid "Install Default Requirements"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:26
#: 1d151e72660644e7b43084ffccb99598
msgid "Download and Prepare LLM Model and Embedding Model"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:28
#: 4443c83f8dfa4fdba30929f1e5ecf619
msgid "If you don't have high performance hardware server"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:30
#: 33a6ee878a4a467d8bcfe5dfbdcbe184
msgid "you can use openai api, tongyi api , bard api, etc."
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:39
#: 7efdb99bc36d42819b1fcd1341b45eef
msgid "set proxy api in .env"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:48
#: 50154ece70bc421ebe7ac05966369bc8
msgid "If you have high performance hardware server"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:62
#: 4f3696ed069a4a84b2d794593df23765
msgid "Start all services with a single command."
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:67
#: f5f671962cb14a10aba34c4274d2fc2c
msgid ""
"By default, the \"dbgpt start webserver\" command will start the "
"Webserver, Model Controller, and Model Worker in a single Python process."
" Here, we specify the service to be started on port 6006."
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:69
#: 4883221a356843f6a6335c25847aecd2
msgid ""
"View and validate the model service in the command line, you can use the "
"following commands"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:70
#: 1346eb55d7c647be90a90a867ba04ec3
msgid ""
"1.list the started model services and deployed Model Workers, you can use"
" the following command"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:74
#: 8ab332ebce554758952dbe85c68330a9
msgid "output is:"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:83
#: efb460bbb70842218f5c056cf05ba1a0
msgid "The WorkerManager is the management process for Model Workers"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:85
#: 10e5b79f3d914ee3974b1a0d07962fe2
msgid ""
"validate the deployed model in the command line, you can use the "
"following command"
msgstr ""

#: ../../getting_started/install/cluster/vms/standalone.md:89
#: 39d062e7fea948949c1c5c099893f308
msgid ""
"Then an interactive page will be launched where you can have a "
"conversation with the deployed LLM in the terminal."
msgstr ""