Skip to content

Latest commit

 

History

History
19 lines (12 loc) · 532 Bytes

README.md

File metadata and controls

19 lines (12 loc) · 532 Bytes

minilm

A tiny transformer language model that learns to babble its own code.

Demo animation

requirements

The program is written in Python using jax. It requires regex and tqdm for training.

usage

The application consists of two trainable components: a tokenizer and the language model. The following code first trains a tokenizer, then trains a model, and then uses both to generate text.

python3 tokenizer.py
python3 train.py
python3 generate.py