File tree Expand file tree Collapse file tree 3 files changed +3
-0
lines changed
auto_deploy/unit/singlegpu Expand file tree Collapse file tree 3 files changed +3
-0
lines changed Original file line number Diff line number Diff line change @@ -319,3 +319,4 @@ disaggregated/test_disaggregated.py::test_disaggregated_diff_max_tokens[TinyLlam
319319disaggregated/test_disaggregated.py::test_disaggregated_deepseek_v3_lite_fp8_tp1_single_gpu_mtp[DeepSeek-V3-Lite-fp8] SKIP (https://nvbugs/5465642)
320320examples/test_multimodal.py::test_llm_multimodal_general[Mistral-Small-3.1-24B-Instruct-2503-pp:1-tp:1-bfloat16-bs:1-cpp_e2e:False-nb:1] SKIP (https://nvbugs/5431146)
321321accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_fp8_blockscale[latency] SKIP (https://nvbugs/5464461)
322+ disaggregated/test_disaggregated.py::test_disaggregated_benchmark_on_diff_backends[DeepSeek-V3-Lite-fp8] SKIP (https://nvbugs/5448449)
Original file line number Diff line number Diff line change @@ -600,6 +600,7 @@ def test_trtllm_bench(llm_root): # noqa: F811
600600 run_benchmark (model_name , dataset_path , temp_dir )
601601
602602
603+ @pytest .mark .skip (reason = "https://nvbugs/5458798" )
603604@pytest .mark .no_xdist
604605def test_trtllm_bench_backend_comparison (llm_root ): # noqa: F811
605606 """Test that compares autodeploy backend performance against pytorch backend
Original file line number Diff line number Diff line change @@ -289,6 +289,7 @@ def per_rank_test_fused_moe_alltoall(job_id):
289289 assert r is None
290290
291291
292+ @pytest .mark .skip (reason = "https://nvbugs/5467531" )
292293@pytest .mark .skipif (torch .cuda .device_count () < 4 ,
293294 reason = "needs 4 GPUs to run this test" )
294295@pytest .mark .parametrize ("alltoall_method_type" , [
You can’t perform that action at this time.
0 commit comments