Skip to content

Actions: triple-Mu/vllm_official

clang-format

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
231 workflow runs
231 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

[Bugfix] Set SamplingParams.max_tokens for OpenAI requests if not pro…
clang-format #106: Commit 3c10591 pushed by triple-Mu
August 1, 2024 05:12 15s main
August 1, 2024 05:12 15s
PP comm optimization: replace send with partial send + allgather (#6695)
clang-format #105: Commit 0437492 pushed by triple-Mu
August 1, 2024 04:06 19s main
August 1, 2024 04:06 19s
[Bugfix][Model] Skip loading lm_head weights if using tie_word_embedd…
clang-format #104: Commit 630dd9e pushed by triple-Mu
August 1, 2024 03:04 15s main
August 1, 2024 03:04 15s
[Bugfix][TPU] Do not use torch.Generator for TPUs (#6981)
clang-format #103: Commit 23993a7 pushed by triple-Mu
August 1, 2024 02:23 17s main
August 1, 2024 02:23 17s
Support W4A8 quantization for vllm (#5218)
clang-format #102: Commit 6512937 pushed by triple-Mu
July 31, 2024 14:30 18s main
July 31, 2024 14:30 18s
[Bugfix] fix logit processor excceed vocab size issue (#6927)
clang-format #101: Commit c0644cf pushed by triple-Mu
July 31, 2024 08:39 18s main
July 31, 2024 08:39 18s
[Bugfix][TPU] Set readonly=True for non-root devices (#6980)
clang-format #100: Commit 533d193 pushed by triple-Mu
July 31, 2024 07:44 16s main
July 31, 2024 07:44 16s
[CI/Build] Fix mypy errors (#6968)
clang-format #99: Commit 9f0e69b pushed by triple-Mu
July 31, 2024 03:05 19s main
July 31, 2024 03:05 19s
[Bugfix] Fix broadcasting logic for multi_modal_kwargs (#6836)
clang-format #98: Commit f230cc2 pushed by triple-Mu
July 31, 2024 02:46 18s main
July 31, 2024 02:46 18s
July 31, 2024 02:23 14s
[Bugfix] Fix PaliGemma MMP (#6930)
clang-format #96: Commit c66c7f8 pushed by triple-Mu
July 30, 2024 09:25 17s main
July 30, 2024 09:25 17s
[Kernel] Tuned int8 kernels for Ada Lovelace (#6848)
clang-format #95: Commit af647fb pushed by triple-Mu
July 30, 2024 03:03 15s main
July 30, 2024 03:03 15s
[Kernel] Tuned FP8 Kernels for Ada Lovelace (#6677)
clang-format #94: Commit 766435e pushed by triple-Mu
July 29, 2024 16:05 20s main
July 29, 2024 16:05 20s
[Model] Initialize support for InternVL2 series models (#6514)
clang-format #93: Commit 7cbd9ec pushed by triple-Mu
July 29, 2024 15:40 16s main
July 29, 2024 15:40 16s
[Misc] Pass cutlass_fp8_supported correctly in fbgemm_fp8 (#6871)
clang-format #92: Commit 3eeb148 pushed by triple-Mu
July 29, 2024 01:51 17s main
July 29, 2024 01:51 17s
Add Nemotron to PP_SUPPORTED_MODELS (#6863)
clang-format #91: Commit b1366a9 pushed by triple-Mu
July 28, 2024 13:20 16s main
July 28, 2024 13:20 16s
[CI/Build][Doc] Update CI and Doc for VLM example changes (#6860)
clang-format #90: Commit ecb33a2 pushed by triple-Mu
July 27, 2024 10:05 17s main
July 27, 2024 10:05 17s
[doc][debugging] add known issues for hangs (#6816)
clang-format #89: Commit 85ad7e2 pushed by triple-Mu
July 26, 2024 08:30 20s main
July 26, 2024 08:30 20s
July 25, 2024 06:02 15s
[Model] Adding support for MiniCPM-V (#4087)
clang-format #87: Commit 9e169a4 pushed by triple-Mu
July 25, 2024 04:00 21s main
July 25, 2024 04:00 21s
[Frontend] Represent tokens with identifiable strings (#6626)
clang-format #86: Commit 5689e25 pushed by triple-Mu
July 25, 2024 02:02 12s main
July 25, 2024 02:02 12s
[Bugfix]fix modelscope compatible issue (#6730)
clang-format #85: Commit f4f8a9d pushed by triple-Mu
July 24, 2024 13:45 17s main
July 24, 2024 13:45 17s
[Docs][ROCm] Detailed instructions to build from source (#6680)
clang-format #84: Commit ccc4a73 pushed by triple-Mu
July 24, 2024 08:21 13s main
July 24, 2024 08:21 13s
[Bugfix] fix flashinfer cudagraph capture for PP (#6708)
clang-format #83: Commit 5e8ca97 pushed by triple-Mu
July 24, 2024 05:09 16s main
July 24, 2024 05:09 16s
[Misc] Add a wrapper for torch.inference_mode (#6618)
clang-format #82: Commit 42de2ce pushed by triple-Mu
July 22, 2024 01:49 34s main
July 22, 2024 01:49 34s