File tree Expand file tree Collapse file tree 2 files changed +6
-3
lines changed
tests/integration/test_lists/test-db Expand file tree Collapse file tree 2 files changed +6
-3
lines changed Original file line number Diff line number Diff line change @@ -35,7 +35,8 @@ l0_dgx_h100:
35
35
- accuracy/test_disaggregated_serving.py::TestGemma3_1BInstruct::test_auto_dtype[False]
36
36
- accuracy/test_disaggregated_serving.py::TestGemma3_1BInstruct::test_auto_dtype[True]
37
37
- accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_ngram
38
- - accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_eagle3
38
+ - accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_eagle3[overlap_scheduler=False-eagle3_one_model=True]
39
+ - accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_eagle3[overlap_scheduler=False-eagle3_one_model=False]
39
40
- test_e2e.py::test_ptp_quickstart_advanced_bs1
40
41
- condition :
41
42
ranges :
Original file line number Diff line number Diff line change @@ -34,14 +34,16 @@ l0_h100:
34
34
- accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_fp8[fp8kv=False-attn_backend=TRTLLM-torch_compile=True]
35
35
- accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_fp8[fp8kv=True-attn_backend=TRTLLM-torch_compile=False]
36
36
- accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_fp8[fp8kv=True-attn_backend=TRTLLM-torch_compile=True]
37
- - accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_eagle3
37
+ - accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_eagle3[overlap_scheduler=False-eagle3_one_model=False]
38
+ - accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_eagle3[overlap_scheduler=False-eagle3_one_model=True]
38
39
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[mtp=disable-fp8kv=True-attention_dp=False-cuda_graph=True-overlap_scheduler=True-torch_compile=True]
39
40
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[mtp=eagle-fp8kv=True-attention_dp=True-cuda_graph=True-overlap_scheduler=True-torch_compile=False]
40
41
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[mtp=vanilla-fp8kv=True-attention_dp=False-cuda_graph=True-overlap_scheduler=True-torch_compile=True]
41
42
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_no_kv_cache_reuse[quant_dtype=fp8-mtp_nextn=2-fp8kv=True-attention_dp=True-cuda_graph=True-overlap_scheduler=True]
42
43
- accuracy/test_llm_api_pytorch.py::TestQwen3_8B::test_fp8_block_scales[latency]
43
44
- accuracy/test_llm_api_pytorch.py::TestQwen3_30B_A3B::test_fp8[latency]
44
- - accuracy/test_llm_api_pytorch.py::TestQwen3_8B::test_eagle3
45
+ - accuracy/test_llm_api_pytorch.py::TestQwen3_8B::test_eagle3[overlap_scheduler=False-eagle3_one_model=False]
46
+ - accuracy/test_llm_api_pytorch.py::TestQwen3_8B::test_eagle3[overlap_scheduler=False-eagle3_one_model=True]
45
47
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales_cuda_graph_padding[mtp_nextn=0]
46
48
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales_cuda_graph_padding[mtp_nextn=2]
47
49
- test_e2e.py::test_trtllm_bench_pytorch_backend_sanity[meta-llama/Llama-3.1-8B-llama-3.1-8b-False-False]
You can’t perform that action at this time.
0 commit comments