Skip to content

Latest commit

 

History

History
47 lines (36 loc) · 1.44 KB

README.md

File metadata and controls

47 lines (36 loc) · 1.44 KB

quick_cot

No-strings tiny Chain-of-Thought framework for your Large Language Model (LLM) that saves you time ⏰ and money 💰

The end goal of this framework is to serve chain of prompts (a.k.a. Chain-of-Thought) formed into schema towards LLM. It iterates through your data stored in CSV/JSONL/sqlite.

Features

  • Provides iterator over infinite amount of input contexts served in CSV/JSONL.
  • Caching progress: withstanding exception during LLM calls by using sqlite3 engine for caching LLM answers;
  • Support schemas descriptions for Chain-of-Thought concept.

Installation

TODO: Work in progress. Use dependencies installation instead.

pip install git+https://github.com/nicolay-r/quick_cot

Usage

Just two simple steps:

  1. Define your sequence of prompts with their dependencies
  2. Launch inference:
python infer.py \
    --model "dynamic:ext/flan_t5.py:FlanT5" \
    --schema "data/default.json" \
    --device "cpu" \
    --temp 0.1 \
    --output "data/output.csv" \
    --max-length 512 \
    --api-token "<API_TOKEN>" \
    --limit 10000 \
    --limit-prompt 10000 \
    --bf16 \
    --l4b

Embed your model

TODO. To be updated.