Skip to content

Using LlamaChat Interface

ymcui edited this page Jun 8, 2023 · 7 revisions

⚠️ Due to breaking changes in llama.cpp, LlamaChat cannot load the latest ggml models.

LlamaChat: https://github.com/alexrozanski/LlamaChat

LlamaChat provides an interactive graphical interface for LLaMA-like models on macOS systems. The following instructions demonstrate the setup process using the Chinese Alpaca 7B model as an example.

⚠️ Warning: LlamaChat requires macOS 13 Ventura, and either an Intel or Apple Silicon processor.

Step 1: Download the latest version of LlamaChat

Choose the latest .dmg file.

Link: https://github.com/alexrozanski/LlamaChat/releases

Step 2: Install LlamaChat

Simply drag LlamaChat into the Applications folder (or any other folder).

image

Step 3: Configure the model

Follow the wizard to configure the model. In this example, choose the Alpaca model. image

Name the model, choose an avatar, and select the appropriate format from the Format dropdown list: PyTorch format (.pth extension) or GGML format (.bin extension). Specify the model path and model size. The GGML format is the model format obtained from converting with llama.cpp. For more information, refer to llama.cpp conversion.

⚠️ LlamaChat does not support the latest quantization format, such as Q5, Q8.

image

Step 4: Chat!

After successfully adding the model, you can interact with it. Click on the icon to the left of the chat box to end the current round of conversation (this clears the context cache but does not delete the corresponding chat history).

(Note: Loading the model is necessary for generating the first sentence, but the speed will return to normal afterward.)

image

Settings

LlamaChat v1.2 introduces customized settings for inference. Take a look and try.

image
Clone this wiki locally