commit_hash string | pr_url string | pr_date string | timeline_extracted_at string | analysis_extracted_at string | models list | perf_command string | has_serving bool | has_latency bool | has_throughput bool | uses_lm_eval bool | commit_subject string | commit_message string | commit_date string | files_changed list | stats dict | diff_text string | apis list | affected_paths list | repo string | hardware string | lm_eval_command string |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
fc542144c4477ffec1d3de6fa43e54f8fb5351e8 | https://github.com/vllm-project/vllm/pull/12563 | 2025-01-31 | 2025-09-07 17:46:50 | 2025-09-07 17:46:50 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --num-prompts 100 | true | false | false | true | [Feature] Fix guided decoding blocking bitmask memcpy (#12563) | [Feature] Fix guided decoding blocking bitmask memcpy (#12563)
**[Guided decoding performance optimization]** Sending the guided
decoding bitmask in xgrammar to the GPU
(`self.token_bitmask.to(scores.device)`) is a blocking operation that
prevents the CPU from pre-launching the sampler kernels. The CPU waits
until dec... | 2025-01-31T15:37:30-08:00 | [
"vllm/model_executor/guided_decoding/xgrammar_decoding.py"
] | {
"commit_year": 2025,
"num_edited_lines": 4,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 4,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/guided_decoding/xgrammar_decoding.py b/vllm/model_executor/guided_decoding/xgrammar_decoding.py
index 2d8594cb8..ee30ce96f 100644
--- a/vllm/model_executor/guided_decoding/xgrammar_decoding.py
+++ b/vllm/model_executor/guided_decoding/xgrammar_decoding.py
@@ -307,8 +307,8 @@ class XGram... | [
"None"
] | [
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/llm.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
fa63e710c7fbaae3a445f669d3b5ba6b9a4ef412 | https://github.com/vllm-project/vllm/pull/12094 | 2025-01-26 | 2025-09-07 17:46:54 | 2025-09-07 17:46:54 | [
"meta-llama/Meta-Llama-3-8B"
] | VLLM_USE_V1=1 python benchmarks/benchmark_latency.py --model meta-llama/Meta-Llama-3-8B --tensor-parallel-size 1 --input-len 1000 --batch-size 32 | false | true | false | true | [V1][Perf] Reduce scheduling overhead in model runner after cuda sync (#12094) | [V1][Perf] Reduce scheduling overhead in model runner after cuda sync (#12094) | 2025-01-26T00:42:37-08:00 | [
"vllm/v1/outputs.py",
"vllm/v1/sample/sampler.py",
"vllm/v1/worker/gpu_model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 34,
"num_files": 3,
"num_hunks": 6,
"num_non_test_edited_lines": 34,
"num_non_test_files": 3,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/v1/outputs.py b/vllm/v1/outputs.py
index acc3a944e..32aee44e3 100644
--- a/vllm/v1/outputs.py
+++ b/vllm/v1/outputs.py
@@ -8,7 +8,7 @@ import torch
class SamplerOutput:
# [num_reqs]
- sampled_token_ids: List[int]
+ sampled_token_ids: torch.Tensor
# [num_reqs, max_num_logprobs + 1]... | [
"vllm.v1.outputs.SamplerOutput",
"vllm.v1.sample.sampler.Sampler.forward",
"vllm.v1.worker.GPUModelRunner.execute_model"
] | [
"vllm/v1/worker/gpu_model_runner.py",
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py",
"vllm/outputs.py",
"vllm/v1/outputs.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
6dd94dbe94c1820a1e224cba65efcf0befa97995 | https://github.com/vllm-project/vllm/pull/12380 | 2025-01-24 | 2025-09-07 17:46:57 | 2025-09-07 17:46:57 | [
"meta-llama/Meta-Llama-3-8B"
] | python benchmarks/benchmark_latency.py --model meta-llama/Meta-Llama-3-8B --load-format dummy | false | true | false | true | [perf] fix perf regression from #12253 (#12380) | [perf] fix perf regression from #12253 (#12380) | 2025-01-24T11:34:27+08:00 | [
"vllm/worker/model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 5,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 5,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/worker/model_runner.py b/vllm/worker/model_runner.py
index cf2f1c6b3..bf1a40d48 100644
--- a/vllm/worker/model_runner.py
+++ b/vllm/worker/model_runner.py
@@ -455,7 +455,6 @@ class ModelInputForGPUBuilder(ModelRunnerInputBuilderBase[ModelInputForGPU]):
self.enable_prompt_adapter = (self.runne... | [
"vllm.worker.model_runner.ModelInputForGPUBuilder.__init__"
] | [
"vllm/worker/model_runner.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/openai/serving_completion.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
310aca88c984983189a57f1b72e3b1dde89fb92f | https://github.com/vllm-project/vllm/pull/11870 | 2025-01-09 | 2025-09-07 17:47:12 | 2025-09-07 17:47:12 | [
"meta-llama/Meta-Llama-3-70B"
] | python benchmarks/benchmark_latency.py --model meta-llama/Meta-Llama-3-70B --load-format dummy --enforce-eager -tp 4 | false | true | false | true | [perf]fix current stream (#11870) | [perf]fix current stream (#11870) | 2025-01-09T07:18:21Z | [
"vllm/distributed/device_communicators/pynccl.py",
"vllm/distributed/parallel_state.py",
"vllm/utils.py",
"vllm/worker/multi_step_model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 61,
"num_files": 4,
"num_hunks": 14,
"num_non_test_edited_lines": 61,
"num_non_test_files": 4,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/distributed/device_communicators/pynccl.py b/vllm/distributed/device_communicators/pynccl.py
index fda4d007c..efc599871 100644
--- a/vllm/distributed/device_communicators/pynccl.py
+++ b/vllm/distributed/device_communicators/pynccl.py
@@ -10,6 +10,7 @@ from vllm.distributed.device_communicators.pynccl... | [
"vllm.distributed.device_communicators.pynccl.PyNcclCommunicator.all_reduce",
"vllm.utils.current_stream",
"vllm.worker.multi_step_model_runner.MultiStepModelRunner.execute_model"
] | [
"vllm/distributed/device_communicators/pynccl.py",
"vllm/distributed/parallel_state.py",
"vllm/worker/multi_step_model_runner.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Meta-Llama-3-70B,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
b55ed6ef8ab0dce7fb0f79ff292dafdb4d22610c | https://github.com/vllm-project/vllm/pull/11692 | 2025-01-02 | 2025-09-07 17:47:18 | 2025-09-07 17:47:18 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm | true | false | false | true | [V1][Minor] Optimize token_ids_cpu copy (#11692) | [V1][Minor] Optimize token_ids_cpu copy (#11692) | 2025-01-02T12:04:58-07:00 | [
"vllm/v1/worker/gpu_input_batch.py",
"vllm/v1/worker/gpu_model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 14,
"num_files": 2,
"num_hunks": 4,
"num_non_test_edited_lines": 14,
"num_non_test_files": 2,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/v1/worker/gpu_input_batch.py b/vllm/v1/worker/gpu_input_batch.py
index e79145300..f8a1427c6 100644
--- a/vllm/v1/worker/gpu_input_batch.py
+++ b/vllm/v1/worker/gpu_input_batch.py
@@ -66,8 +66,9 @@ class InputBatch:
pin_memory=False,
)
self.token_ids_cpu = self.token_ids_... | [
"InputBatch.add_request",
"InputBatch.condense",
"GPUModelRunner._update_states"
] | [
"vllm/v1/worker/gpu_input_batch.py",
"vllm/v1/worker/gpu_model_runner.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
3b61cb450d899dc423feb264c297d4d18d701678 | https://github.com/vllm-project/vllm/pull/10989 | 2024-12-09 | 2025-09-07 17:47:34 | 2025-09-07 17:47:34 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_latency.py --model meta-llama/Llama-3.1-8B-Instruct --batch-size 32 --input-len 512 --output-len 128 | false | true | false | true | [V1] Further reduce CPU overheads in flash-attn (#10989) | [V1] Further reduce CPU overheads in flash-attn (#10989) | 2024-12-09T12:38:46-08:00 | [
"csrc/cache_kernels.cu",
"vllm/v1/attention/backends/flash_attn.py"
] | {
"commit_year": 2024,
"num_edited_lines": 35,
"num_files": 2,
"num_hunks": 2,
"num_non_test_edited_lines": 35,
"num_non_test_files": 2,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/csrc/cache_kernels.cu b/csrc/cache_kernels.cu
index 1be806bbf..8a95279f9 100644
--- a/csrc/cache_kernels.cu
+++ b/csrc/cache_kernels.cu
@@ -307,10 +307,20 @@ void reshape_and_cache_flash(
torch::Tensor& key_cache, // [num_blocks, block_size, num_heads, head_size]
torch::Tensor&
value_ca... | [
"vllm.v1.attention.backends.flash_attn.FlashAttentionImpl.forward",
"torch.ops._C_cache_ops.reshape_and_cache_flash"
] | [
"vllm/attention/backends/flash_attn.py",
"vllm/v1/attention/backends/flash_attn.py",
"vllm/_custom_ops.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
98f47f2a4032f8c395268de80858c64ffcfc60fa | https://github.com/vllm-project/vllm/pull/10733 | 2024-11-28 | 2025-09-07 17:47:41 | 2025-09-07 17:47:41 | [
"facebook/opt-125m"
] | python benchmarks/benchmark_latency.py --model facebook/opt-125m | false | true | false | true | [V1] Optimize the CPU overheads in FlashAttention custom op (#10733) | [V1] Optimize the CPU overheads in FlashAttention custom op (#10733) | 2024-11-28T09:01:02-08:00 | [
"vllm/v1/attention/backends/flash_attn.py"
] | {
"commit_year": 2024,
"num_edited_lines": 17,
"num_files": 1,
"num_hunks": 4,
"num_non_test_edited_lines": 17,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/v1/attention/backends/flash_attn.py b/vllm/v1/attention/backends/flash_attn.py
index 5f8535eaa..e618edf7d 100644
--- a/vllm/v1/attention/backends/flash_attn.py
+++ b/vllm/v1/attention/backends/flash_attn.py
@@ -135,6 +135,13 @@ class FlashAttentionImpl(AttentionImpl):
assert k_scale == 1.0 an... | [
"vllm.v1.attention.backends.flash_attn.FlashAttentionImpl.forward",
"vllm.v1.attention.backends.flash_attn.unified_v1_flash_attention"
] | [
"vllm/attention/backends/flash_attn.py",
"vllm/v1/attention/backends/flash_attn.py",
"vllm/_custom_ops.py",
"csrc/torch_bindings.cpp",
"csrc/cpu/torch_bindings.cpp",
"csrc/rocm/torch_bindings.cpp",
"csrc/moe/torch_bindings.cpp"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=facebook/opt-125m,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
8c1e77fb585c4f42783a3d88c1efc7c9e15fd89f | https://github.com/vllm-project/vllm/pull/10742 | 2024-11-28 | 2025-09-07 17:47:44 | 2025-09-07 17:47:44 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_latency.py --model meta-llama/Llama-3.1-8B-Instruct --batch-size 32 --input-len 512 --output-len 128 | false | true | false | true | [Kernel] Update vllm-flash-attn version to reduce CPU overheads (#10742) | [Kernel] Update vllm-flash-attn version to reduce CPU overheads (#10742) | 2024-11-28T08:31:28-08:00 | [
"CMakeLists.txt"
] | {
"commit_year": 2024,
"num_edited_lines": 2,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 2,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/CMakeLists.txt b/CMakeLists.txt
index 45a3b484e..f43bf8143 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -522,7 +522,7 @@ else()
FetchContent_Declare(
vllm-flash-attn
GIT_REPOSITORY https://github.com/vllm-project/flash-attention.git
- GIT_TAG d886f88165702b3c7e77445... | [
"None"
] | [
"vllm/attention/backends/flash_attn.py",
"vllm/v1/attention/backends/flash_attn.py",
"vllm/attention/ops/triton_flash_attention.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
6e36f4fa6ce64619b9ea94c88a157f5783a63a65 | https://github.com/vllm-project/vllm/pull/7874 | 2024-09-02 | 2025-09-07 17:48:01 | 2025-09-07 17:48:01 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | improve chunked prefill performance | improve chunked prefill performance
[Bugfix] Fix #7592 vllm 0.5.4 enable_chunked_prefill throughput is slightly lower than 0.5.3~0.5.0. (#7874) | 2024-09-02T14:20:12-07:00 | [
"tests/basic_correctness/test_chunked_prefill.py",
"vllm/core/scheduler.py"
] | {
"commit_year": 2024,
"num_edited_lines": 18,
"num_files": 2,
"num_hunks": 2,
"num_non_test_edited_lines": 15,
"num_non_test_files": 1,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/basic_correctness/test_chunked_prefill.py b/tests/basic_correctness/test_chunked_prefill.py
index fc6f829c3..a63ac380e 100644
--- a/tests/basic_correctness/test_chunked_prefill.py
+++ b/tests/basic_correctness/test_chunked_prefill.py
@@ -116,6 +116,9 @@ def test_models_with_fp8_kv_cache(
pyt... | [
"vllm.core.scheduler.Scheduler.schedule",
"vllm.core.scheduler.SchedulerOutputs"
] | [
"vllm/core/scheduler.py",
"vllm/v1/core/sched/scheduler.py",
"vllm/attention/ops/chunked_prefill_paged_decode.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
e3580537a41a46b0f3cd750b86b633c1857a8c90 | https://github.com/vllm-project/vllm/pull/7753 | 2024-08-28 | 2025-09-07 17:48:09 | 2025-09-07 17:48:09 | [
"RedHatAI/Meta-Llama-3-8B-Instruct-FP8"
] | python benchmarks/benchmark_serving.py --model RedHatAI/Meta-Llama-3-8B-Instruct-FP8 --enable-prefix-caching --enable-chunked-prefill --max-num-batched-tokens 2048 | true | false | false | true | [Performance] Enable chunked prefill and prefix caching together (#7753) | [Performance] Enable chunked prefill and prefix caching together (#7753) | 2024-08-28T00:36:31-07:00 | [
"tests/basic_correctness/test_chunked_prefill.py",
"tests/core/test_block_manager.py",
"tests/core/test_chunked_prefill_scheduler.py",
"vllm/core/block_manager_v1.py",
"vllm/core/block_manager_v2.py",
"vllm/core/embedding_model_block_manager.py",
"vllm/core/interfaces.py",
"vllm/core/scheduler.py",
... | {
"commit_year": 2024,
"num_edited_lines": 252,
"num_files": 9,
"num_hunks": 12,
"num_non_test_edited_lines": 107,
"num_non_test_files": 6,
"num_test_files": 3,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/basic_correctness/test_chunked_prefill.py b/tests/basic_correctness/test_chunked_prefill.py
index 1211e6ba5..fc6f829c3 100644
--- a/tests/basic_correctness/test_chunked_prefill.py
+++ b/tests/basic_correctness/test_chunked_prefill.py
@@ -6,6 +6,7 @@ prefill requests are chunked.
Run `pytest tests/... | [
"ModelRunner.generate_greedy",
"Scheduler.schedule",
"BlockSpaceManager.mark_blocks_as_computed"
] | [
"vllm/worker/model_runner.py",
"vllm/core/scheduler.py",
"vllm/v1/core/sched/scheduler.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=RedHatAI/Meta-Llama-3-8B-Instruct-FP8,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
2deb029d115dadd012ce5ea70487a207cb025493 | https://github.com/vllm-project/vllm/pull/7822 | 2024-08-26 | 2025-09-07 17:48:12 | 2025-09-07 17:48:12 | [
"RedHatAI/Meta-Llama-3-8B-Instruct-FP8"
] | python benchmarks/benchmark_prefix_caching.py --model RedHatAI/Meta-Llama-3-8B-Instruct-FP8 --output-len 200 --enable-prefix-caching | true | false | false | true | [Performance][BlockManagerV2] Mark prefix cache block as computed after schedule (#7822) | [Performance][BlockManagerV2] Mark prefix cache block as computed after schedule (#7822) | 2024-08-26T11:24:53-07:00 | [
"tests/core/block/test_prefix_caching_block.py",
"vllm/core/block/prefix_caching_block.py",
"vllm/core/block_manager_v2.py"
] | {
"commit_year": 2024,
"num_edited_lines": 63,
"num_files": 3,
"num_hunks": 6,
"num_non_test_edited_lines": 32,
"num_non_test_files": 2,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/core/block/test_prefix_caching_block.py b/tests/core/block/test_prefix_caching_block.py
index c2226870c..25be2dd13 100644
--- a/tests/core/block/test_prefix_caching_block.py
+++ b/tests/core/block/test_prefix_caching_block.py
@@ -708,6 +708,37 @@ class TestPrefixCachingBlockAllocator:
... | [
"PrefixCachingBlockAllocator.mark_blocks_as_computed",
"BlockSpaceManagerV2.mark_blocks_as_computed"
] | [
"vllm/core/block/prefix_caching_block.py",
"vllm/core/block_manager.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=RedHatAI/Meta-Llama-3-8B-Instruct-FP8,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
fc7b8d1eefcbe837a56b7c080509417fe5167e6c | https://github.com/vllm-project/vllm/pull/7364 | 2024-08-09 | 2025-09-07 17:48:14 | 2025-09-07 17:48:14 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | [Performance] e2e overheads reduction: Small followup diff (#7364) | [Performance] e2e overheads reduction: Small followup diff (#7364) | 2024-08-09T15:49:36Z | [
"vllm/core/block_manager_v1.py",
"vllm/sequence.py"
] | {
"commit_year": 2024,
"num_edited_lines": 7,
"num_files": 2,
"num_hunks": 2,
"num_non_test_edited_lines": 7,
"num_non_test_files": 2,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/core/block_manager_v1.py b/vllm/core/block_manager_v1.py
index 622aca66a..ad26d3c51 100644
--- a/vllm/core/block_manager_v1.py
+++ b/vllm/core/block_manager_v1.py
@@ -336,9 +336,9 @@ class BlockSpaceManagerV1(BlockSpaceManager):
# Assign the self-attention block tables for each sequence.
... | [
"BlockSpaceManagerV1",
"SequenceGroup.get_finished_seqs"
] | [
"vllm/sequence.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
660470e5a36b8e52083615ad7c85e9b4fd4c72ce | https://github.com/vllm-project/vllm/pull/7193 | 2024-08-06 | 2025-09-07 17:48:19 | 2025-09-07 17:48:19 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --tensor-parallel-size 1 --enable-prefix-caching --use-v2-block-manager | true | false | false | true | [Core] Optimize evictor-v2 performance (#7193) | [Core] Optimize evictor-v2 performance (#7193) | 2024-08-06T12:34:25-07:00 | [
"vllm/core/evictor_v2.py"
] | {
"commit_year": 2024,
"num_edited_lines": 6,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 6,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/core/evictor_v2.py b/vllm/core/evictor_v2.py
index 3dd12e2e2..5b1a208b7 100644
--- a/vllm/core/evictor_v2.py
+++ b/vllm/core/evictor_v2.py
@@ -91,8 +91,9 @@ class LRUEvictor(Evictor):
# at the start of OrderedDict. Loop through all these blocks to
# find the one with maximum number o... | [
"None"
] | [
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/api_server.py",
"vllm/entrypoints/openai/api_server.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
89a84b0bb7b30706a02836234a94493ea8f780bf | https://github.com/vllm-project/vllm/pull/6779 | 2024-07-26 | 2025-09-07 17:48:26 | 2025-09-07 17:48:26 | [
"Qwen/Qwen1.5-0.5B"
] | python benchmarks/benchmark_serving.py --model Qwen/Qwen1.5-0.5B --backend vllm --num-prompts 2048 --input-len 1024 | true | false | false | true | [Core] Use array to speedup padding (#6779) | [Core] Use array to speedup padding (#6779) | 2024-07-25T21:31:31-07:00 | [
"vllm/model_executor/layers/sampler.py",
"vllm/model_executor/sampling_metadata.py",
"vllm/sequence.py"
] | {
"commit_year": 2024,
"num_edited_lines": 46,
"num_files": 3,
"num_hunks": 9,
"num_non_test_edited_lines": 46,
"num_non_test_files": 3,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/sampler.py b/vllm/model_executor/layers/sampler.py
index 5c376797a..121458f81 100644
--- a/vllm/model_executor/layers/sampler.py
+++ b/vllm/model_executor/layers/sampler.py
@@ -220,7 +220,7 @@ def _apply_min_tokens_penalty(
seqs_to_penalize: List[int] = []
... | [
"vllm.sequence.SequenceData.prompt_token_ids_array",
"vllm.sequence.SequenceData.output_token_ids_array",
"vllm.model_executor.sampling_metadata.SamplingTensors.from_lists"
] | [
"vllm/sequence.py",
"vllm/model_executor/sampling_metadata.py",
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=Qwen/Qwen1.5-0.5B,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
9ed82e7074a18e25680ab106fc846364ad97bc00 | https://github.com/vllm-project/vllm/pull/6520 | 2024-07-19 | 2025-09-07 17:48:29 | 2025-09-07 17:48:29 | [
"meta-llama/Llama-2-7b-hf"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-2-7b-hf --backend vllm --num-prompts 100 | true | false | false | true | [Misc] Small perf improvements (#6520) | [Misc] Small perf improvements (#6520) | 2024-07-19T12:10:56-07:00 | [
"tests/core/block/test_block_manager_v2.py",
"tests/core/block/test_cpu_gpu_block_allocator.py",
"vllm/core/block/block_table.py",
"vllm/core/block/prefix_caching_block.py",
"vllm/model_executor/models/__init__.py",
"vllm/sequence.py",
"vllm/utils.py"
] | {
"commit_year": 2024,
"num_edited_lines": 69,
"num_files": 7,
"num_hunks": 11,
"num_non_test_edited_lines": 50,
"num_non_test_files": 5,
"num_test_files": 2,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/core/block/test_block_manager_v2.py b/tests/core/block/test_block_manager_v2.py
index d0ca09c4b..d7863a9ae 100644
--- a/tests/core/block/test_block_manager_v2.py
+++ b/tests/core/block/test_block_manager_v2.py
@@ -249,10 +249,13 @@ def test_append_slots(block_size, prompt_len, num_slots_to_append,
... | [
"BlockTable.get_num_token_blocks",
"ModelRegistry.load_model_cls",
"SequenceGroup.prompt",
"SequenceGroup.prompt_token_ids",
"SequenceGroup.multi_modal_data"
] | [] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-2-7b-hf,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
3476ed0809ec91a3457da0cb90543133a4f4b519 | https://github.com/vllm-project/vllm/pull/5602 | 2024-07-02 | 2025-09-07 17:48:40 | 2025-09-07 17:48:40 | [
"meta-llama/Llama-2-7b-hf"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-2-7b-hf --dtype float16 --num-prompts 300 --seed 0 | true | false | false | true | [Core] Optimize block_manager_v2 vs block_manager_v1 (to make V2 default) (#5602) | [Core] Optimize block_manager_v2 vs block_manager_v1 (to make V2 default) (#5602) | 2024-07-01T20:10:37-07:00 | [
"benchmarks/benchmark_latency.py",
"tests/conftest.py",
"tests/core/block/test_block_table.py",
"tests/core/block/test_cpu_gpu_block_allocator.py",
"tests/core/block/test_naive_block.py",
"tests/core/block/test_prefix_caching_block.py",
"tests/spec_decode/test_batch_expansion.py",
"vllm/core/block/blo... | {
"commit_year": 2024,
"num_edited_lines": 1721,
"num_files": 19,
"num_hunks": 107,
"num_non_test_edited_lines": 1570,
"num_non_test_files": 13,
"num_test_files": 6,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/benchmarks/benchmark_latency.py b/benchmarks/benchmark_latency.py
index a46ee1581..8d0554b0f 100644
--- a/benchmarks/benchmark_latency.py
+++ b/benchmarks/benchmark_latency.py
@@ -46,6 +46,7 @@ def main(args: argparse.Namespace):
load_format=args.load_format,
distributed_executor_backend=... | [
"CpuGpuBlockAllocator.allocate_mutable_block",
"CpuGpuBlockAllocator.allocate_immutable_block",
"PrefixCachingBlockAllocator.allocate_mutable_block",
"PrefixCachingBlockAllocator.allocate_immutable_block",
"PrefixCachingBlock.__init__"
] | [
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/openai/serving_completion.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-2-7b-hf,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
d7740ea4dcee4ab75d7d6eef723f33cae957b288 | https://github.com/vllm-project/vllm/pull/4594 | 2024-05-08 | 2025-09-07 17:49:00 | 2025-09-07 17:49:00 | [
"meta-llama/Llama-2-7b-hf"
] | python benchmarks/benchmark_throughput.py --model meta-llama/Llama-2-7b-hf --input-len 256 --output-len 256 | false | false | true | true | [Core] Optimize sampler get_logprobs (#4594) | [Core] Optimize sampler get_logprobs (#4594) | 2024-05-08T08:42:28-07:00 | [
"vllm/model_executor/layers/sampler.py"
] | {
"commit_year": 2024,
"num_edited_lines": 117,
"num_files": 1,
"num_hunks": 3,
"num_non_test_edited_lines": 117,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/sampler.py b/vllm/model_executor/layers/sampler.py
index 1f19d2053..e52e350d2 100644
--- a/vllm/model_executor/layers/sampler.py
+++ b/vllm/model_executor/layers/sampler.py
@@ -782,13 +782,14 @@ def _get_logprobs(
top_logprobs, top_token_ids = torch.topk(logprobs,
... | [
"None"
] | [
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/llm.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-2-7b-hf,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
ad8d696a99ca1eee19f1404e16e8e82df592ff85 | https://github.com/vllm-project/vllm/pull/4270 | 2024-04-22 | 2025-09-07 17:49:06 | 2025-09-07 17:49:06 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | [Core] Scheduler perf fix (#4270) | [Core] Scheduler perf fix (#4270) | 2024-04-22T21:11:06Z | [
"tests/core/test_scheduler.py",
"vllm/core/scheduler.py"
] | {
"commit_year": 2024,
"num_edited_lines": 25,
"num_files": 2,
"num_hunks": 13,
"num_non_test_edited_lines": 7,
"num_non_test_files": 1,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/core/test_scheduler.py b/tests/core/test_scheduler.py
index 9588a1bea..a25112385 100644
--- a/tests/core/test_scheduler.py
+++ b/tests/core/test_scheduler.py
@@ -540,7 +540,7 @@ def test_decode_schedule_preempted():
curr_loras = None
for i in range(3):
_, seq_group = create_dummy_p... | [
"vllm.core.scheduler.Scheduler._allocate_and_set_running"
] | [
"vllm/core/scheduler.py",
"vllm/v1/core/sched/scheduler.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/llm.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
9474e89ba4ecae253b585eb6b3e1d85f4e108f01 | https://github.com/vllm-project/vllm/pull/3357 | 2024-03-20 | 2025-09-07 17:49:25 | 2025-09-07 17:49:25 | [
"huggyllama/llama-7b"
] | python benchmarks/benchmark_throughput.py --model huggyllama/llama-7b --dataset-name sharegpt --num-prompts 2000 | false | false | true | true | [PREFIX CACHING FOLLOW UP] A bunch of fixes to block allocator performance when automatic prefix caching is disabled (#3357) | [PREFIX CACHING FOLLOW UP] A bunch of fixes to block allocator performance when automatic prefix caching is disabled (#3357) | 2024-03-20T00:11:11-07:00 | [
"tests/core/test_block_manager.py",
"tests/prefix_caching/test_prefix_caching.py",
"vllm/core/block_manager.py",
"vllm/core/evictor.py"
] | {
"commit_year": 2024,
"num_edited_lines": 286,
"num_files": 4,
"num_hunks": 25,
"num_non_test_edited_lines": 260,
"num_non_test_files": 2,
"num_test_files": 2,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/core/test_block_manager.py b/tests/core/test_block_manager.py
index 44ac05a14..9473a33f0 100644
--- a/tests/core/test_block_manager.py
+++ b/tests/core/test_block_manager.py
@@ -4,7 +4,7 @@ from typing import List
from vllm import SamplingParams
from vllm.block import PhysicalTokenBlock
-from vll... | [
"vllm.core.block_manager.UncachedBlockAllocator",
"vllm.core.block_manager.CachedBlockAllocator",
"vllm.core.block_manager.BlockSpaceManager"
] | [
"vllm/core/block_manager.py",
"vllm/core/block/prefix_caching_block.py",
"examples/offline_inference/automatic_prefix_caching.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=huggyllama/llama-7b,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
b690e34824fd5a5c4054a0c0468ebfb6aa1dd215 | https://github.com/vllm-project/vllm/pull/21075 | 2025-08-02 | 2025-09-07 17:49:40 | 2025-09-07 17:49:40 | [
"ibm-ai-platform/Bamba-9B-v2"
] | python benchmarks/benchmark_serving.py --model ibm-ai-platform/Bamba-9B-v2 --dataset-name sharegpt | true | false | false | true | [Model] Mamba2 preallocate SSM output tensor to avoid d2d copy overhead (#21075) | [Model] Mamba2 preallocate SSM output tensor to avoid d2d copy overhead (#21075) | 2025-08-02T01:59:34-07:00 | [
"tests/kernels/mamba/test_mamba_ssm.py",
"tests/kernels/mamba/test_mamba_ssm_ssd.py",
"vllm/model_executor/layers/mamba/mamba_mixer.py",
"vllm/model_executor/layers/mamba/mamba_mixer2.py",
"vllm/model_executor/layers/mamba/ops/mamba_ssm.py",
"vllm/model_executor/layers/mamba/ops/ssd_chunk_scan.py",
"vll... | {
"commit_year": 2025,
"num_edited_lines": 262,
"num_files": 9,
"num_hunks": 39,
"num_non_test_edited_lines": 165,
"num_non_test_files": 7,
"num_test_files": 2,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/kernels/mamba/test_mamba_ssm.py b/tests/kernels/mamba/test_mamba_ssm.py
index 8dece26dd..4c32ae81b 100644
--- a/tests/kernels/mamba/test_mamba_ssm.py
+++ b/tests/kernels/mamba/test_mamba_ssm.py
@@ -365,6 +365,7 @@ def test_selective_state_update(dim, dstate, has_z, itype):
batch_size = 1
st... | [
"vllm.model_executor.layers.mamba.ops.mamba_ssm.selective_state_update",
"vllm.model_executor.layers.mamba.ops.ssd_combined.mamba_chunk_scan_combined"
] | [
"vllm/model_executor/layers/mamba/mamba_mixer2.py",
"vllm/model_executor/models/phi4flash.py",
"vllm/model_executor/models/plamo2.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=ibm-ai-platform/Bamba-9B-v2,dtype=float16 --tasks gsm8k --batch_size auto --limit 100 |
58eee5f2e05b74eb2cb1a3bbda9c04df4805e4cc | https://github.com/vllm-project/vllm/pull/20000 | 2025-08-02 | 2025-09-07 17:49:44 | 2025-09-07 17:49:44 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | [PERF] Use faster way of decode in tokenizer: avoid useless list-to-list conversion (#20000) | [PERF] Use faster way of decode in tokenizer: avoid useless list-to-list conversion (#20000) | 2025-08-02T01:43:52-07:00 | [
"vllm/transformers_utils/tokenizer.py"
] | {
"commit_year": 2025,
"num_edited_lines": 7,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 7,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/transformers_utils/tokenizer.py b/vllm/transformers_utils/tokenizer.py
index 24ddd35ab..6a31a4198 100644
--- a/vllm/transformers_utils/tokenizer.py
+++ b/vllm/transformers_utils/tokenizer.py
@@ -50,11 +50,12 @@ def decode_tokens(
`skip_special_tokens=None` means to use the backend's default
... | [
"vllm.transformers_utils.tokenizer.decode_tokens"
] | [
"vllm/transformers_utils/tokenizer.py",
"vllm/transformers_utils/tokenizer_base.py",
"vllm/transformers_utils/tokenizer_group.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
a32237665df876fcb51196dc209e8aff9fd89d29 | https://github.com/vllm-project/vllm/pull/21245 | 2025-07-22 | 2025-09-07 17:50:16 | 2025-09-07 17:50:16 | [
"facebook/opt-125m"
] | vllm bench serve --dataset-name random --model facebook/opt-125m --served-model-name facebook/opt-125m --random-input-len 700 --random-output-len 1 --endpoint /v1/completions --ignore-eos --host localhost --port 8000 --request-rate 200 --num-prompts 100 | true | false | false | true | [Core] Optimize update checks in LogitsProcessor (#21245) | [Core] Optimize update checks in LogitsProcessor (#21245) | 2025-07-22T05:27:18-07:00 | [
"vllm/v1/sample/logits_processor.py"
] | {
"commit_year": 2025,
"num_edited_lines": 18,
"num_files": 1,
"num_hunks": 3,
"num_non_test_edited_lines": 18,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/v1/sample/logits_processor.py b/vllm/v1/sample/logits_processor.py
index 3a4c25964..3a06e7105 100644
--- a/vllm/v1/sample/logits_processor.py
+++ b/vllm/v1/sample/logits_processor.py
@@ -335,14 +335,19 @@ class LogitBiasLogitsProcessor(LogitsProcessor):
if not batch_update:
retur... | [
"LogitBiasLogitsProcessor.update_state",
"MinTokensLogitsProcessor.update_state"
] | [
"vllm/v1/sample/logits_processor.py",
"vllm/model_executor/layers/logits_processor.py",
"vllm/entrypoints/openai/logits_processors.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=facebook/opt-125m,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
e7b204268132cb775c139574c1ff4ad7e15c8f66 | https://github.com/vllm-project/vllm/pull/21334 | 2025-07-22 | 2025-09-07 17:50:20 | 2025-09-07 17:50:20 | [
"meta-llama/Llama-4-Maverick-17B-128E-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-4-Maverick-17B-128E-Instruct --dataset-name sharegpt --num-prompts 100 | true | false | false | true | Revert "[Performance] Performance improvements in non-blockwise fp8 CUTLASS MoE (#20762) (#21334) | Revert "[Performance] Performance improvements in non-blockwise fp8 CUTLASS MoE (#20762) (#21334) | 2025-07-21T21:49:01-07:00 | [
"benchmarks/kernels/benchmark_grouped_gemm_cutlass.py",
"csrc/moe/moe_permute_unpermute_op.cu",
"tests/kernels/moe/test_cutlass_moe.py",
"tests/kernels/moe/test_pplx_cutlass_moe.py",
"vllm/model_executor/layers/fused_moe/cutlass_moe.py",
"vllm/model_executor/layers/quantization/compressed_tensors/compress... | {
"commit_year": 2025,
"num_edited_lines": 212,
"num_files": 6,
"num_hunks": 30,
"num_non_test_edited_lines": 176,
"num_non_test_files": 4,
"num_test_files": 2,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/benchmarks/kernels/benchmark_grouped_gemm_cutlass.py b/benchmarks/kernels/benchmark_grouped_gemm_cutlass.py
index a6b42406b..1d4e730f9 100644
--- a/benchmarks/kernels/benchmark_grouped_gemm_cutlass.py
+++ b/benchmarks/kernels/benchmark_grouped_gemm_cutlass.py
@@ -80,11 +80,6 @@ def bench_run(
a, s... | [
"vllm.cutlass_moe_fp8",
"CutlassExpertsFp8.apply",
"CompressedTensorsW8A8Fp8MoECutlassMethod.select_gemm_impl"
] | [
"vllm/model_executor/layers/fused_moe/cutlass_moe.py",
"vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors_moe.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-4-Maverick-17B-128E-Instruct,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
015069b01741e9ecb9e604c7fe87fbdfc306ebe5 | https://github.com/vllm-project/vllm/pull/17515 | 2025-05-01 | 2025-09-07 17:51:10 | 2025-09-07 17:51:10 | [
"Qwen/Qwen3-7B-Instruct"
] | python benchmarks/benchmark_serving.py --model Qwen/Qwen3-7B-Instruct --dataset-name sharegpt --request-rate 1 | true | false | false | true | [Misc] Optimize the Qwen3_ReasoningParser extract_reasoning_content (#17515) | [Misc] Optimize the Qwen3_ReasoningParser extract_reasoning_content (#17515) | 2025-05-01T03:29:01-07:00 | [
"vllm/reasoning/qwen3_reasoning_parser.py"
] | {
"commit_year": 2025,
"num_edited_lines": 53,
"num_files": 1,
"num_hunks": 3,
"num_non_test_edited_lines": 53,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/reasoning/qwen3_reasoning_parser.py b/vllm/reasoning/qwen3_reasoning_parser.py
index f588f4016..7095034b1 100644
--- a/vllm/reasoning/qwen3_reasoning_parser.py
+++ b/vllm/reasoning/qwen3_reasoning_parser.py
@@ -1,6 +1,5 @@
# SPDX-License-Identifier: Apache-2.0
-import re
from collections.abc impor... | [
"Qwen3ReasoningParser.extract_reasoning_content"
] | [
"vllm/reasoning/qwen3_reasoning_parser.py",
"examples/online_serving/openai_chat_completion_with_reasoning.py",
"examples/online_serving/openai_chat_completion_with_reasoning_streaming.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=Qwen/Qwen3-7B-Instruct,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
bc7c4d206bbfb56b06d218b6c2971e8ca191db36 | https://github.com/vllm-project/vllm/pull/13305 | 2025-04-23 | 2025-09-07 17:51:14 | 2025-09-07 17:51:14 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --dtype float16 --num-prompts 300 --seed 0 | true | false | false | true | [Kernel][ROCM] Upstream prefix prefill speed up for vLLM V1 (#13305) | [Kernel][ROCM] Upstream prefix prefill speed up for vLLM V1 (#13305) | 2025-04-22T19:11:56-07:00 | [
"tests/core/block/e2e/test_correctness.py",
"vllm/attention/ops/prefix_prefill.py"
] | {
"commit_year": 2025,
"num_edited_lines": 1640,
"num_files": 2,
"num_hunks": 4,
"num_non_test_edited_lines": 1634,
"num_non_test_files": 1,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/core/block/e2e/test_correctness.py b/tests/core/block/e2e/test_correctness.py
index e9b537ed5..9e8e315d8 100644
--- a/tests/core/block/e2e/test_correctness.py
+++ b/tests/core/block/e2e/test_correctness.py
@@ -195,15 +195,15 @@ def test_lookahead_greedy_equality_with_preemption(baseline_llm_generator... | [
"vllm.attention.ops.prefix_prefill.context_attention_fwd"
] | [
"vllm/attention/ops/prefix_prefill.py",
"vllm/attention/layer.py",
"vllm/model_executor/layers/fused_moe/layer.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
299ebb62b269ce167eb1c71b5e39a1dc1f65ce1c | https://github.com/vllm-project/vllm/pull/16436 | 2025-04-21 | 2025-09-07 17:51:17 | 2025-09-07 17:51:17 | [
"Qwen/Qwen2.5-1.5B-Instruct"
] | vllm bench serve --model Qwen/Qwen2.5-1.5B-Instruct --request-rate 1 --num-prompts 100 --random-input-len 1000 --random-output-len 100 --tokenizer Qwen/Qwen2.5-1.5B-Instruct --ignore-eos | true | false | false | true | [Core] Speed up decode by remove synchronizing operation in sampler (#16436) | [Core] Speed up decode by remove synchronizing operation in sampler (#16436) | 2025-04-21T18:18:22Z | [
"vllm/model_executor/layers/utils.py"
] | {
"commit_year": 2025,
"num_edited_lines": 13,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 13,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/utils.py b/vllm/model_executor/layers/utils.py
index a9ef97391..5e56be061 100644
--- a/vllm/model_executor/layers/utils.py
+++ b/vllm/model_executor/layers/utils.py
@@ -47,10 +47,15 @@ def apply_penalties(logits: torch.Tensor, prompt_tokens_tensor: torch.Tensor,
output_t... | [
"vllm.model_executor.layers.utils.apply_penalties"
] | [
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/openai/serving_completion.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=Qwen/Qwen2.5-1.5B-Instruct,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
35fad35a485eac9195c510731ba4a9d297dfd963 | https://github.com/vllm-project/vllm/pull/15478 | 2025-03-26 | 2025-09-07 17:51:35 | 2025-09-07 17:51:35 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm | true | false | false | true | [V1][Sampler] Faster top-k only implementation (#15478) | [V1][Sampler] Faster top-k only implementation (#15478) | 2025-03-26T10:56:47-07:00 | [
"tests/v1/sample/test_topk_topp_sampler.py",
"vllm/v1/sample/ops/topk_topp_sampler.py",
"vllm/v1/sample/sampler.py"
] | {
"commit_year": 2025,
"num_edited_lines": 96,
"num_files": 3,
"num_hunks": 7,
"num_non_test_edited_lines": 59,
"num_non_test_files": 2,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/v1/sample/test_topk_topp_sampler.py b/tests/v1/sample/test_topk_topp_sampler.py
new file mode 100644
index 000000000..8a5076412
--- /dev/null
+++ b/tests/v1/sample/test_topk_topp_sampler.py
@@ -0,0 +1,37 @@
+# SPDX-License-Identifier: Apache-2.0
+import torch
+from torch import Generator
+
+from vllm... | [
"vllm.v1.sample.ops.topk_topp_sampler.apply_top_k_top_p",
"vllm.v1.sample.ops.topk_topp_sampler.apply_top_k_only",
"vllm.v1.sample.ops.topk_topp_sampler.TopKTopPSampler.forward_native",
"vllm.v1.sample.sampler.Sampler.forward"
] | [
"vllm/v1/sample/ops/topk_topp_sampler.py",
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
296f927f2493908984707354e3cc5d7b2e41650b | https://github.com/vllm-project/vllm/pull/14857 | 2025-03-21 | 2025-09-07 17:51:42 | 2025-09-07 17:51:42 | [
"ibm-ai-platform/Bamba-9B"
] | python benchmarks/benchmark_serving.py --model ibm-ai-platform/Bamba-9B --dataset-name sharegpt | true | false | false | true | [Model] RE: Mamba2 Prefill Performance Tweaks: Fixing Flurry of Unnecessary Memory Copies (#14857) | [Model] RE: Mamba2 Prefill Performance Tweaks: Fixing Flurry of Unnecessary Memory Copies (#14857) | 2025-03-20T19:21:08-07:00 | [
"vllm/model_executor/layers/mamba/mamba_mixer2.py"
] | {
"commit_year": 2025,
"num_edited_lines": 13,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 13,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/mamba/mamba_mixer2.py b/vllm/model_executor/layers/mamba/mamba_mixer2.py
index fec6d6112..d7a45bc51 100644
--- a/vllm/model_executor/layers/mamba/mamba_mixer2.py
+++ b/vllm/model_executor/layers/mamba/mamba_mixer2.py
@@ -470,10 +470,11 @@ class MambaMixer2(CustomOp):
if ... | [
"MambaMixer2.forward_cuda",
"mamba_mixer2"
] | [
"vllm/model_executor/layers/mamba/mamba_mixer2.py",
"vllm/model_executor/models/mamba2.py",
"vllm/model_executor/models/mamba_cache.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=ibm-ai-platform/Bamba-9B,dtype=float16 --tasks gsm8k --batch_size auto --limit 100 |
22d33baca2c0c639cfd45c48e99803e56c3efa74 | https://github.com/vllm-project/vllm/pull/15150 | 2025-03-19 | 2025-09-07 17:51:45 | 2025-09-07 17:51:45 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | [FrontEnd][Perf] `merge_async_iterators` fast-path for single-prompt requests (#15150) | [FrontEnd][Perf] `merge_async_iterators` fast-path for single-prompt requests (#15150) | 2025-03-19T21:04:41Z | [
"vllm/utils.py"
] | {
"commit_year": 2025,
"num_edited_lines": 5,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 5,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/utils.py b/vllm/utils.py
index 79787303a..9bc081890 100644
--- a/vllm/utils.py
+++ b/vllm/utils.py
@@ -411,6 +411,11 @@ async def merge_async_iterators(
When it yields, it yields a tuple (i, item) where i is the index of the
iterator that yields the item.
"""
+ if len(iterators) == 1... | [
"vllm.utils.merge_async_iterators"
] | [
"vllm/entrypoints/api_server.py",
"vllm/entrypoints/openai/api_server.py",
"vllm/entrypoints/openai/serving_completion.py",
"vllm/engine/async_llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
99abb8b650c66664cdc84d815b7f306f33bd9881 | https://github.com/vllm-project/vllm/pull/14930 | 2025-03-18 | 2025-09-07 17:51:49 | 2025-09-07 17:51:49 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --speculative-model '[ngram]' --num-speculative-tokens 3 --dataset-name sharegpt | true | false | false | true | [V1][Spec Decode] Optimize Rejection Sampler with Triton Kernels (#14930) | [V1][Spec Decode] Optimize Rejection Sampler with Triton Kernels (#14930) | 2025-03-18T14:31:54-07:00 | [
"tests/v1/sample/test_rejection_sampler.py",
"vllm/envs.py",
"vllm/v1/outputs.py",
"vllm/v1/sample/ops/utils.py",
"vllm/v1/sample/rejection_sampler.py",
"vllm/v1/spec_decode/metadata.py",
"vllm/v1/spec_decode/utils.py",
"vllm/v1/worker/gpu_model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 1329,
"num_files": 8,
"num_hunks": 34,
"num_non_test_edited_lines": 1098,
"num_non_test_files": 7,
"num_test_files": 1,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/v1/sample/test_rejection_sampler.py b/tests/v1/sample/test_rejection_sampler.py
index 84139a40b..8c423e367 100644
--- a/tests/v1/sample/test_rejection_sampler.py
+++ b/tests/v1/sample/test_rejection_sampler.py
@@ -6,20 +6,23 @@ import torch
import torch.nn.functional as F
from vllm.v1.sample.meta... | [
"RejectionSampler.forward",
"RejectionSampler.parse_output",
"SpecDecodeMetadata.make_dummy",
"compiled_softmax"
] | [
"vllm/v1/sample/rejection_sampler.py",
"vllm/v1/spec_decode/metadata.py",
"vllm/v1/sample/metadata.py",
"vllm/v1/pool/metadata.py",
"vllm/v1/sample/tpu/metadata.py",
"vllm/v1/worker/gpu_model_runner.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct --tasks gsm8k --num_fewshot 5 --limit 100 |
fe66b34728e5d383e3d19aefc544eeee808c99fb | https://github.com/vllm-project/vllm/pull/14778 | 2025-03-14 | 2025-09-07 17:51:55 | 2025-09-07 17:51:55 | [
"ibm-ai-platform/Bamba-9B"
] | python benchmarks/benchmark_serving.py --model ibm-ai-platform/Bamba-9B --dtype float16 --num-prompts 300 --seed 0 | true | false | false | true | [Model] Mamba2 Prefill Performance Tweaks: Fixing Flurry of Unnecessary Memory Copies (#14778) | [Model] Mamba2 Prefill Performance Tweaks: Fixing Flurry of Unnecessary Memory Copies (#14778) | 2025-03-14T16:36:18-04:00 | [
"vllm/model_executor/layers/mamba/mamba_mixer2.py"
] | {
"commit_year": 2025,
"num_edited_lines": 30,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 30,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/mamba/mamba_mixer2.py b/vllm/model_executor/layers/mamba/mamba_mixer2.py
index b53a540ed..5b19e3f35 100644
--- a/vllm/model_executor/layers/mamba/mamba_mixer2.py
+++ b/vllm/model_executor/layers/mamba/mamba_mixer2.py
@@ -466,10 +466,17 @@ class MambaMixer2(CustomOp):
if ... | [
"MambaMixer2.forward_cuda",
"mamba_mixer2"
] | [
"vllm/model_executor/layers/mamba/mamba_mixer2.py",
"vllm/model_executor/models/mamba2.py",
"vllm/model_executor/models/mamba_cache.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=ibm-ai-platform/Bamba-9B,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
70b808fe1a63322bc6bf5f46a91981a8f6b8af00 | https://github.com/vllm-project/vllm/pull/14377 | 2025-03-11 | 2025-09-07 17:51:59 | 2025-09-07 17:51:59 | [
"Qwen/Qwen2-VL-7B"
] | python benchmarks/benchmark_serving.py --model Qwen/Qwen2-VL-7B --dataset-name random --request-rate 1 | true | false | false | true | [Perf]:Optimize qwen2-vl to reduce cudaMemcpyAsync (#14377) | [Perf]:Optimize qwen2-vl to reduce cudaMemcpyAsync (#14377) | 2025-03-11T07:39:56Z | [
"vllm/model_executor/models/qwen2_5_vl.py",
"vllm/model_executor/models/qwen2_vl.py"
] | {
"commit_year": 2025,
"num_edited_lines": 94,
"num_files": 2,
"num_hunks": 12,
"num_non_test_edited_lines": 94,
"num_non_test_files": 2,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/models/qwen2_5_vl.py b/vllm/model_executor/models/qwen2_5_vl.py
index ef3d28c80..ae48c7794 100644
--- a/vllm/model_executor/models/qwen2_5_vl.py
+++ b/vllm/model_executor/models/qwen2_5_vl.py
@@ -255,10 +255,12 @@ class Qwen2_5_VisionAttention(nn.Module):
return q, k, v
... | [
"Qwen2VisionAttention.forward",
"Qwen2VisionBlock.forward",
"Qwen2VisionTransformer.forward",
"Qwen2_5_VisionAttention.forward",
"Qwen2_5_VisionTransformer.forward"
] | [
"vllm/model_executor/models/qwen2_vl.py",
"vllm/model_executor/models/qwen2_5_vl.py",
"vllm/multimodal/registry.py",
"vllm/inputs/registry.py",
"vllm/model_executor/models/registry.py",
"vllm/transformers_utils/chat_templates/registry.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=Qwen/Qwen2-VL-7B,dtype=auto --trust_remote_code --tasks gsm8k --batch_size auto --limit 100 |
9f1710f1ace3535920c0bb6d4cc329c36289080e | https://github.com/vllm-project/vllm/pull/13897 | 2025-03-06 | 2025-09-07 17:52:14 | 2025-09-07 17:52:14 | [
"deepseek-ai/DeepSeek-V2-Lite-Chat"
] | python benchmarks/benchmark_serving.py --model deepseek-ai/DeepSeek-V2-Lite-Chat --input-len 28000 --output-len 64 | true | false | false | true | Fix mla prefill context performance (#13897) | Fix mla prefill context performance (#13897) | 2025-03-06T09:35:49-08:00 | [
"vllm/attention/backends/mla/common.py",
"vllm/v1/attention/backends/mla/common.py"
] | {
"commit_year": 2025,
"num_edited_lines": 4,
"num_files": 2,
"num_hunks": 2,
"num_non_test_edited_lines": 4,
"num_non_test_files": 2,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/attention/backends/mla/common.py b/vllm/attention/backends/mla/common.py
index 8184b0732..109e8496f 100644
--- a/vllm/attention/backends/mla/common.py
+++ b/vllm/attention/backends/mla/common.py
@@ -1308,7 +1308,7 @@ class MLACommonImpl(MLAAttentionImpl[T], Generic[T]):
)
k... | [
"MLACommonImpl.forward"
] | [
"vllm/core/block/common.py",
"vllm/attention/backends/mla/common.py",
"vllm/v1/attention/backends/mla/common.py",
"vllm/attention/ops/prefix_prefill.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=deepseek-ai/DeepSeek-V2-Lite-Chat,dtype=float16 --tasks gsm8k --batch_size auto --limit 100 |
9badee53decb3d432dc805336abfb0eb81dfb48f | https://github.com/vllm-project/vllm/pull/14223 | 2025-03-04 | 2025-09-07 17:52:18 | 2025-09-07 17:52:18 | [
"meta-llama/Llama-3.2-1B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.2-1B-Instruct --dataset-path ShareGPT_V3_unfiltered_cleaned_split.json | true | false | false | true | Fix performance when `--generation-config` is not `None` (#14223) | Fix performance when `--generation-config` is not `None` (#14223) | 2025-03-04T20:59:22+01:00 | [
"vllm/entrypoints/llm.py",
"vllm/entrypoints/openai/serving_chat.py",
"vllm/entrypoints/openai/serving_completion.py",
"vllm/entrypoints/openai/serving_transcription.py"
] | {
"commit_year": 2025,
"num_edited_lines": 48,
"num_files": 4,
"num_hunks": 8,
"num_non_test_edited_lines": 48,
"num_non_test_files": 4,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/entrypoints/llm.py b/vllm/entrypoints/llm.py
index 122e2ed86..fc585ee9e 100644
--- a/vllm/entrypoints/llm.py
+++ b/vllm/entrypoints/llm.py
@@ -244,6 +244,7 @@ class LLM:
engine_args, usage_context=UsageContext.LLM_CLASS)
self.request_counter = Counter()
+ self.default_sa... | [
"vLLM.LLM.get_default_sampling_params",
"OpenAIServingChat.create_chat_completion",
"OpenAIServingCompletion.create_completion",
"OpenAIServingTranscription._preprocess_transcription"
] | [
"vllm/entrypoints/llm.py",
"vllm/entrypoints/openai/serving_chat.py",
"vllm/entrypoints/openai/serving_completion.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.2-1B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
19d98e0c7db96713f0e2201649159431177a56e2 | https://github.com/vllm-project/vllm/pull/13625 | 2025-03-03 | 2025-09-07 17:52:21 | 2025-09-07 17:52:21 | [
"deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct"
] | python benchmarks/benchmark_serving.py --model deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct --dataset-name sharegpt --num-prompts 100 | true | false | false | true | [Kernel] Optimize moe intermediate_cache usage (#13625) | [Kernel] Optimize moe intermediate_cache usage (#13625) | 2025-03-03T16:29:53-05:00 | [
"vllm/model_executor/layers/fused_moe/fused_moe.py"
] | {
"commit_year": 2025,
"num_edited_lines": 17,
"num_files": 1,
"num_hunks": 1,
"num_non_test_edited_lines": 17,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/layers/fused_moe/fused_moe.py b/vllm/model_executor/layers/fused_moe/fused_moe.py
index 00260313e..5336b3c10 100644
--- a/vllm/model_executor/layers/fused_moe/fused_moe.py
+++ b/vllm/model_executor/layers/fused_moe/fused_moe.py
@@ -1240,15 +1240,20 @@ def fused_experts_impl(hidden_state... | [
"torch.ops.vllm.inplace_fused_experts",
"torch.ops.vllm.outplace_fused_experts"
] | [
"vllm/model_executor/layers/fused_moe/fused_moe.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/llm.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct,dtype=float16 --tasks gsm8k --batch_size auto |
e206b5433109d298e53451015465b2bf8f03ef0a | https://github.com/vllm-project/vllm/pull/13837 | 2025-02-26 | 2025-09-07 17:52:25 | 2025-09-07 17:52:25 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 --guided-decoding-backend xgrammar | true | false | false | true | [v0][Core] Use xgrammar shared context to avoid copy overhead for offline engine (#13837) | [v0][Core] Use xgrammar shared context to avoid copy overhead for offline engine (#13837) | 2025-02-26T14:58:24+08:00 | [
"vllm/model_executor/guided_decoding/xgrammar_decoding.py"
] | {
"commit_year": 2025,
"num_edited_lines": 26,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 26,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/model_executor/guided_decoding/xgrammar_decoding.py b/vllm/model_executor/guided_decoding/xgrammar_decoding.py
index 329b03a57..e6ba7f5ec 100644
--- a/vllm/model_executor/guided_decoding/xgrammar_decoding.py
+++ b/vllm/model_executor/guided_decoding/xgrammar_decoding.py
@@ -3,7 +3,6 @@
# noqa: UP007
... | [
"vllm.model_executor.guided_decoding.xgrammar_decoding.XGrammarLogitsProcessor.clone"
] | [
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py",
"vllm/entrypoints/llm.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
6a417b8600d4d1e57698a91b71a38446e8fc5c45 | https://github.com/vllm-project/vllm/pull/13589 | 2025-02-20 | 2025-09-07 17:52:28 | 2025-09-07 17:52:28 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | fix neuron performance issue (#13589) | fix neuron performance issue (#13589) | 2025-02-20T10:59:36-08:00 | [
"vllm/worker/neuron_worker.py"
] | {
"commit_year": 2025,
"num_edited_lines": 4,
"num_files": 1,
"num_hunks": 2,
"num_non_test_edited_lines": 4,
"num_non_test_files": 1,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/vllm/worker/neuron_worker.py b/vllm/worker/neuron_worker.py
index 5f0eb0019..95e7acd02 100644
--- a/vllm/worker/neuron_worker.py
+++ b/vllm/worker/neuron_worker.py
@@ -76,7 +76,7 @@ class NeuronWorker(LoraNotSupportedWorkerBase, LocalOrDistributedWorkerBase):
# Set the number of GPU blocks to be t... | [
"NeuronWorker.determine_num_available_blocks",
"NeuronWorker.initialize_cache"
] | [
"vllm/worker/neuron_worker.py",
"vllm/worker/neuron_model_runner.py",
"vllm/platforms/neuron.py",
"examples/offline_inference/neuron.py",
"vllm/model_executor/model_loader/neuron.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
4c822298981a8f7521492075ff72659985fc4c3f | https://github.com/vllm-project/vllm/pull/13365 | 2025-02-18 | 2025-09-07 17:52:34 | 2025-09-07 17:52:34 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_latency.py --model meta-llama/Llama-3.1-8B-Instruct --speculative-model meta-llama/Llama-3.2-1B-Instruct --num-speculative-tokens 5 | false | true | false | true | [V1][Spec Decode] Optimize N-gram matching with Numba (#13365) | [V1][Spec Decode] Optimize N-gram matching with Numba (#13365) | 2025-02-18T13:19:58-08:00 | [
"requirements-common.txt",
"vllm/v1/spec_decode/ngram_proposer.py",
"vllm/v1/worker/gpu_model_runner.py"
] | {
"commit_year": 2025,
"num_edited_lines": 127,
"num_files": 3,
"num_hunks": 5,
"num_non_test_edited_lines": 127,
"num_non_test_files": 3,
"num_test_files": 0,
"only_non_test_files": 1,
"only_test_files": 0
} | diff --git a/requirements-common.txt b/requirements-common.txt
index b7c94cbdb..c52980bc7 100644
--- a/requirements-common.txt
+++ b/requirements-common.txt
@@ -1,6 +1,7 @@
psutil
sentencepiece # Required for LLaMA tokenizer.
numpy < 2.0.0
+numba == 0.60.0 # v0.61 doesn't support Python 3.9. Required for N-gram spe... | [
"vllm.v1.spec_decode.ngram_proposer.NgramProposer.propose",
"vllm.v1.worker.GPUModelRunner.__init__"
] | [
"vllm/v1/spec_decode/ngram_proposer.py",
"vllm/v1/worker/gpu_model_runner.py",
"vllm/engine/llm_engine.py",
"vllm/v1/engine/llm_engine.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
30172b4947c52890b808c6da3a6c7580f55cbb74 | https://github.com/vllm-project/vllm/pull/13244 | 2025-02-18 | 2025-09-07 17:52:38 | 2025-09-07 17:52:38 | [
"meta-llama/Llama-3.1-8B-Instruct"
] | python benchmarks/benchmark_serving.py --model meta-llama/Llama-3.1-8B-Instruct --backend vllm --num-prompts 100 | true | false | false | true | [V1] Optimize handling of sampling metadata and req_ids list (#13244) | [V1] Optimize handling of sampling metadata and req_ids list (#13244) | 2025-02-18T12:15:33-08:00 | [
"tests/v1/sample/test_rejection_sampler.py",
"tests/v1/sample/test_sampler.py",
"tests/v1/worker/test_gpu_input_batch.py",
"tests/v1/worker/test_gpu_model_runner.py",
"vllm/model_executor/layers/utils.py",
"vllm/v1/core/scheduler.py",
"vllm/v1/sample/metadata.py",
"vllm/v1/sample/ops/penalties.py",
... | {
"commit_year": 2025,
"num_edited_lines": 553,
"num_files": 15,
"num_hunks": 74,
"num_non_test_edited_lines": 420,
"num_non_test_files": 11,
"num_test_files": 4,
"only_non_test_files": 0,
"only_test_files": 0
} | diff --git a/tests/v1/sample/test_rejection_sampler.py b/tests/v1/sample/test_rejection_sampler.py
index 8bc33e841..3e810e525 100644
--- a/tests/v1/sample/test_rejection_sampler.py
+++ b/tests/v1/sample/test_rejection_sampler.py
@@ -26,17 +26,13 @@ def create_logits_tensor(token_ids: List[int],
def create_sampling_met... | [
"vllm.v1.sample.metadata.SamplingMetadata",
"vllm.v1.sample.sampler.Sampler",
"vllm.v1.worker.gpu_input_batch.InputBatch",
"vllm.v1.worker.gpu_model_runner.GPUModelRunner"
] | [
"vllm/v1/spec_decode/metadata.py",
"vllm/v1/sample/metadata.py",
"vllm/v1/pool/metadata.py",
"vllm/v1/sample/tpu/metadata.py",
"vllm/v1/sample/sampler.py",
"vllm/model_executor/layers/sampler.py",
"vllm/v1/sample/tpu/sampler.py",
"vllm/v1/worker/gpu_input_batch.py"
] | vllm | H100 | lm_eval --model vllm --model_args pretrained=meta-llama/Llama-3.1-8B-Instruct,dtype=auto --tasks gsm8k --batch_size auto --limit 100 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.