Skip to content

Latest commit

 

History

History
19 lines (12 loc) · 760 Bytes

README.md

File metadata and controls

19 lines (12 loc) · 760 Bytes

TT-Inference-Server

Tenstorrent Inference Server (tt-inference-server) is the repo of available model APIs for deploying on Tenstorrent hardware.

Official Repository

https://github.com/tenstorrent/tt-inference-server

Getting Started

Please follow setup instructions found in each model folder's README.md doc


Model Implementations

Model Hardware
LLaMa 3.1 70B TT-QuietBox & TT-LoudBox
Mistral 7B n150 and n300