Skip to content

Using open source LLMs to build synthetic datasets for direct preference optimization

Notifications You must be signed in to change notification settings

davanstrien/haiku-dpo

Repository files navigation

🌸 Synthetic Haiku DPO 🌸

Using open source LLMs to build synthetic datasets for direct preference optimization

Banner for a dataset card featuring a fusion of digital and traditional Japanese elements. The design includes stylized digital prompts and haiku within text bubbles and on digital screens, set against a backdrop of delicate cherry blossoms and a serene Japanese landscape. The color scheme is dominated by soft pastel pink tones, creating a harmonious blend of modern technology and classical poetry aesthetics.

Lines in code retreat,
Synthetic haiku compete,
Nature's rhythm, neat.

📖 About

Built with Distilabel

Haiku (俳句) is a type of short form poetry that originated in Japan. Traditional Japanese haiku consist of three phrases composed of 17 phonetic units (called on in Japanese, which are similar to syllables) in a 5, 7, 5 pattern - https://en.wikipedia.org/wiki/Haiku

A current request for a haiku looks like when using the TinyLlama/TinyLlama-1.1B-Chat-v1.0 model looks like this:

>>> Write a haiku about moss.
Mosses soft, green and shining,
Gracefully drifting in the breeze,
A symphony of light and sound.

This has some of the properties of a traditional haiku (as expressed in the English language), it has three lines, and the themes of nature and the seasons are present. However, the syllable count is off.

What happens if we do a little prompt engineering and ask the model to write a haiku with a 5, 7, 5 syllable pattern?

>>> Write a haiku about moss. Follow the traditional 5-7-5 syllable pattern.
Mosses, tiny green eyes, inchworm life,
Growing tall, they're hard to see,
Syllables flow, as the world moves by: haiku of life.

Prompt engineering didn't help use here. What if we use a slightly larger model? The teknium/OpenHermes-2.5-Mistral-7B model is a 7B parameter model, which I've found to perform very well for many tasks.

>>> Write a haiku about moss. Follow the traditional 5-7-5 syllable pattern.
Moss, soft emerald hue,
Creeps on stone, whispers secrets -
Nature's silent truth.

This repository contains the code used to generate the Synthetic Haiku DPO dataset, a dataset of synthetic haiku generated using Distilabel.

The goal of this repo is to help the author explore the process of using synthetic data to train a model for direct preference optimization (DPO).

📝 The Plan

  1. Generate synthetic haiku prompt/response pairs from teknium/OpenHermes-2.5-Mistral-7B
  2. Rank the generated haiku using a syllable counter
  3. Train a model on the generated haiku
  4. Evaluate the model's performance in generating "technically" correct haiku

The next step will involve 5. annotating haiku generated by the model by aesthetic quality i.e. which haiku do I like the most 6. seeing the extent to which a DPO dataset can be used to optimize for aesthetic quality 7. Profit?

📚 Notebooks

These notebooks work through the first few stages of this project; the end result is a DPO dataset. The plan is to also add notebooks for the other stages over time.

  1. 01_generate_haiku_prompts.ipynb Open In Colab - This notebook contains code for generating haiku prompts using Distilabel
  2. 02_generate_haiku.ipynb Open In Colab - This notebook generates the actual haiku
  3. 03_rule_based_ranking.ipynb Open In Colab - This notebook carries out the rule-based ranking of the generated haiku which can then be used to create a preference dataset.

About

Using open source LLMs to build synthetic datasets for direct preference optimization

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published