KOZA 🐐 is an instruct model for Polish language with similar capabilities for LLMs like ChatGPT, but it can be run on a single machine 💻 with a strong GPU (RTX 4080/4090). This repo was originally forked from alpaca-lora.
Using this repo, you can reproduce the Stanford Alpaca for the Polish language. Original paper from Stanford: low-rank adaptation (LoRA).
In your terminal clone repo:
git clone git@github.com:bqpro1/alpaca-koza.git
From repo folder install requirements.txt
:
pip isntall -r requirements.txt
And run:
python generate.py
Go to http://127.0.0.1:7860 in your browser. Enjoy KOZA!
I used translation_instructions.ipynb
for auto-translation instructions from data/alpaca_data.json
to data/alpaca_data_pl_verified.json
. Translation is far from perfect, might be improved. Translation took 12h. You can also find translated data on huggingface🤗
Just run
python finetune.py --base_model='decapoda-research/llama-7b-hf'
Weights for Polsih language are also on huggingface🤗.