Skip to content

Commit 4055b76

Browse files
authored
[None][fix] disagg ctx pp4 + gen pp4 integ test (#6489)
Signed-off-by: raayandhar <[email protected]> Signed-off-by: Raayan Dhar <[email protected]>
1 parent 0223de0 commit 4055b76

File tree

3 files changed

+4
-7
lines changed

3 files changed

+4
-7
lines changed

tests/integration/defs/disaggregated/test_disaggregated.py

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -648,7 +648,6 @@ def test_disaggregated_ctxpp2_gentp2(disaggregated_test_root, llm_venv,
648648
def test_disaggregated_ctxtp2pp2_gentp2pp2(disaggregated_test_root, llm_venv,
649649
disaggregated_example_root,
650650
llama_model_root):
651-
pytest.skip(f"8 GPU test times out currently, skipping")
652651
src_dst_dict = {
653652
llama_model_root:
654653
f"{llm_venv.get_working_directory()}/TinyLlama/TinyLlama-1.1B-Chat-v1.0",
@@ -664,15 +663,13 @@ def test_disaggregated_ctxtp2pp2_gentp2pp2(disaggregated_test_root, llm_venv,
664663

665664

666665
@pytest.mark.skip_less_device(8)
667-
@pytest.mark.parametrize("llama_model_root", ['TinyLlama-1.1B-Chat-v1.0'],
668-
indirect=True)
666+
@pytest.mark.parametrize("llama_model_root", ['llama-3.1-8b'], indirect=True)
669667
def test_disaggregated_ctxpp4_genpp4(disaggregated_test_root, llm_venv,
670668
disaggregated_example_root,
671669
llama_model_root):
672-
pytest.skip(f"8 GPU test times out currently, skipping")
673670
src_dst_dict = {
674671
llama_model_root:
675-
f"{llm_venv.get_working_directory()}/TinyLlama/TinyLlama-1.1B-Chat-v1.0",
672+
f"{llm_venv.get_working_directory()}/llama-3.1-models/Meta-Llama-3.1-8B",
676673
}
677674
for src, dst in src_dst_dict.items():
678675
if not os.path.islink(dst):

tests/integration/test_lists/qa/llm_function_full.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -670,7 +670,7 @@ disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_qwen3[
670670
disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_qwen3[True-False-Qwen3-8B-FP8]
671671
disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_qwen3[True-True-Qwen3-8B-FP8]
672672
disaggregated/test_disaggregated.py::test_disaggregated_ctxtp2pp2_gentp2pp2[TinyLlama-1.1B-Chat-v1.0]
673-
disaggregated/test_disaggregated.py::test_disaggregated_ctxpp4_genpp4[TinyLlama-1.1B-Chat-v1.0]
673+
disaggregated/test_disaggregated.py::test_disaggregated_ctxpp4_genpp4[llama-3.1-8b]
674674
disaggregated/test_workers.py::test_workers_conditional_disaggregation[TinyLlama-1.1B-Chat-v1.0]
675675
disaggregated/test_workers.py::test_workers_kv_cache_events[TinyLlama-1.1B-Chat-v1.0]
676676
disaggregated/test_workers.py::test_workers_kv_cache_aware_router[TinyLlama-1.1B-Chat-v1.0]

tests/integration/test_lists/test-db/l0_dgx_h200.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ l0_dgx_h200:
3030
- accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_ctx_pp_gen_tp_asymmetric[MMLU-gen_tp=1-ctx_pp=4]
3131
- accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_ctx_pp_gen_tp_asymmetric[MMLU-gen_tp=2-ctx_pp=4]
3232
- disaggregated/test_disaggregated.py::test_disaggregated_ctxtp2pp2_gentp2pp2[TinyLlama-1.1B-Chat-v1.0]
33-
- disaggregated/test_disaggregated.py::test_disaggregated_ctxpp4_genpp4[TinyLlama-1.1B-Chat-v1.0]
33+
- disaggregated/test_disaggregated.py::test_disaggregated_ctxpp4_genpp4[llama-3.1-8b]
3434
- unittest/_torch/multi_gpu_modeling/test_llama4.py::test_llama4[pp1-ep1-disable_adp-enable_graph-tp8-trtllm-scout]
3535
- unittest/_torch/multi_gpu_modeling/test_llama4.py::test_llama4[pp1-ep4-enable_adp-enable_graph-tp8-trtllm-scout]
3636
- unittest/llmapi/test_llm_pytorch.py::test_nemotron_nas_lora

0 commit comments

Comments
 (0)