From 6658dd5e7013cb2884be22157558b0cab55aecc0 Mon Sep 17 00:00:00 2001 From: Ying Sheng Date: Sat, 3 Feb 2024 10:47:26 +0000 Subject: [PATCH 1/3] bump version to v0.1.11 --- README.md | 4 ++++ python/pyproject.toml | 2 +- python/sglang/__init__.py | 2 +- 3 files changed, 6 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 2667000cf1..7b14acb4b7 100644 --- a/README.md +++ b/README.md @@ -8,6 +8,10 @@ The core features of SGLang include: - **A Flexible Front-End Language**: This allows for easy programming of LLM applications with multiple chained generation calls, advanced prompting techniques, control flow, multiple modalities, parallelism, and external interaction. - **A High-Performance Runtime with RadixAttention**: This feature significantly accelerates the execution of complex LLM programs by automatic KV cache reuse across multiple calls. It also supports other common techniques like continuous batching and tensor parallelism. +## News +- [2024/01] 🔥 SGLang powers the serving of offical LLaVA v1.6 release demo. [blog](https://llava-vl.github.io/blog/2024-01-30-llava-1-6/) +- [2024/01] SGLang provides up to 5x faster inference with RadixAttention. [blog](https://lmsys.org/blog/2024-01-17-sglang/). + ## Contents - [Install](#install) - [Quick Start](#quick-start) diff --git a/python/pyproject.toml b/python/pyproject.toml index 6fb94df409..2cd9869240 100644 --- a/python/pyproject.toml +++ b/python/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta" [project] name = "sglang" -version = "0.1.10" +version = "0.1.11" description = "A structured generation langauge for LLMs." readme = "README.md" requires-python = ">=3.8" diff --git a/python/sglang/__init__.py b/python/sglang/__init__.py index 2d01bcb034..e39817ab23 100644 --- a/python/sglang/__init__.py +++ b/python/sglang/__init__.py @@ -1,4 +1,4 @@ -__version__ = "0.1.10" +__version__ = "0.1.11" from sglang.api import * from sglang.global_config import global_config From 833b16f702824bb05c6d58d8697b82217f67652e Mon Sep 17 00:00:00 2001 From: Ying Sheng Date: Sat, 3 Feb 2024 10:48:33 +0000 Subject: [PATCH 2/3] update --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 7b14acb4b7..d4941c76f2 100644 --- a/README.md +++ b/README.md @@ -9,8 +9,8 @@ The core features of SGLang include: - **A High-Performance Runtime with RadixAttention**: This feature significantly accelerates the execution of complex LLM programs by automatic KV cache reuse across multiple calls. It also supports other common techniques like continuous batching and tensor parallelism. ## News -- [2024/01] 🔥 SGLang powers the serving of offical LLaVA v1.6 release demo. [blog](https://llava-vl.github.io/blog/2024-01-30-llava-1-6/) -- [2024/01] SGLang provides up to 5x faster inference with RadixAttention. [blog](https://lmsys.org/blog/2024-01-17-sglang/). +- [2024/01] 🔥 SGLang powers the serving of offical LLaVA v1.6 release demo ([blog](https://llava-vl.github.io/blog/2024-01-30-llava-1-6/)). +- [2024/01] SGLang provides up to 5x faster inference with RadixAttention ([blog](https://lmsys.org/blog/2024-01-17-sglang/)). ## Contents - [Install](#install) From 68cae4ea42b349e92bf5d5ed7dc15bb0e38e2794 Mon Sep 17 00:00:00 2001 From: Ying Sheng Date: Sat, 3 Feb 2024 10:49:38 +0000 Subject: [PATCH 3/3] update --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index d4941c76f2..6dcb0bb03e 100644 --- a/README.md +++ b/README.md @@ -9,7 +9,7 @@ The core features of SGLang include: - **A High-Performance Runtime with RadixAttention**: This feature significantly accelerates the execution of complex LLM programs by automatic KV cache reuse across multiple calls. It also supports other common techniques like continuous batching and tensor parallelism. ## News -- [2024/01] 🔥 SGLang powers the serving of offical LLaVA v1.6 release demo ([blog](https://llava-vl.github.io/blog/2024-01-30-llava-1-6/)). +- [2024/01] 🔥 SGLang powers the serving of the offical LLaVA v1.6 release demo ([blog](https://llava-vl.github.io/blog/2024-01-30-llava-1-6/)). - [2024/01] SGLang provides up to 5x faster inference with RadixAttention ([blog](https://lmsys.org/blog/2024-01-17-sglang/)). ## Contents