From 117586b816b5db6883d3ad7c0384c589a05960ed Mon Sep 17 00:00:00 2001 From: Guoqing Bao Date: Wed, 24 Jul 2024 11:32:57 +0800 Subject: [PATCH] Fix typo & update ReadMe --- README.md | 4 ++-- src/openai/streaming.rs | 4 ++-- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index 8f0b309..f9e1f0a 100644 --- a/README.md +++ b/README.md @@ -23,8 +23,8 @@ Currently, candle-vllm supports chat serving for the following models. | #2 | **Mistral** |✅|70 tks/s (7B)| | #3 | **Phi (v1, v1.5, v2)** |✅|97 tks/s (2.7B, F32+BF16)| | #4 | **Phi-3 (3.8B, 7B)** |✅|107 tks/s (3.8B)| -| #5 | **Yi** |✅|TBD| -| #6 | **StableLM** |✅|TBD| +| #5 | **Yi** |✅|75 tks/s (6B)| +| #6 | **StableLM** |✅|99 tks/s (3B)| | #7 | BigCode/StarCode |TBD|TBD| | #8 | ChatGLM |TBD|TBD| | #9 | **QWen2 (1.8B, 7B)** |✅|148 tks/s (1.8B)| diff --git a/src/openai/streaming.rs b/src/openai/streaming.rs index e972b28..455dab1 100644 --- a/src/openai/streaming.rs +++ b/src/openai/streaming.rs @@ -11,7 +11,7 @@ use std::{ pub enum StreamingStatus { Uninitilized, Started, - Interupted, + Interrupted, Stopped, } pub enum ChatResponse { @@ -55,7 +55,7 @@ impl Stream for Streamer { if self.status == StreamingStatus::Started && e == flume::TryRecvError::Disconnected { //no TryRecvError::Disconnected returned even if the client closed the stream or disconnected - self.status = StreamingStatus::Interupted; + self.status = StreamingStatus::Interrupted; Poll::Ready(None) } else { Poll::Pending