Skip to content

Commit

Permalink
doc(v6e): mention initial v6e support (#131)
Browse files Browse the repository at this point in the history
  • Loading branch information
tengomucho authored Dec 18, 2024
1 parent eb20d37 commit e302950
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 4 deletions.
6 changes: 3 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ We currently support a few LLM models targeting text generation scenarios:
## Inference

`optimum-tpu` provides a set of dedicated tools and integrations in order to leverage Cloud TPUs for inference, especially
on the latest TPU version `v5e`.
on the latest TPU version `v5e` and `v6e`.

Other TPU versions will be supported along the way.

Expand All @@ -64,8 +64,8 @@ To enable the support, export the environment variable `JETSTREAM_PT=1`.

Fine-tuning is supported and tested on the TPU `v5e`. We have tested so far:

- πŸ¦™ Llama-2 7B and Llama-3 8B
- πŸ’Ž Gemma 2B and 7B
- πŸ¦™ Llama-2 7B, Llama-3 8B and newer;
- πŸ’Ž Gemma 2B and 7B.

You can check the examples:

Expand Down
5 changes: 4 additions & 1 deletion docs/source/howto/training.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -4,15 +4,18 @@ Welcome to the πŸ€— Optimum-TPU training guide! This section covers how to fine-

## Currently Supported Models

The following models have been tested and validated for fine-tuning on TPU v5e:
The following models have been tested and validated for fine-tuning on TPU `v5e` and `v6e`:

- πŸ¦™ LLaMA Family
- LLaMA-2 7B
- LLaMA-3 8B
- LLaMA-3.2 1B
- πŸ’Ž Gemma Family
- Gemma 2B
- Gemma 7B

Bigger models are supported, but not yet tested.

## Getting Started

### Prerequisites
Expand Down

0 comments on commit e302950

Please sign in to comment.