We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b0558c7 commit efaefcaCopy full SHA for efaefca
tests/integration/test_lists/qa/llm_perf_sanity.yml
@@ -191,8 +191,8 @@ llm_perf_sanity:
191
192
tests:
193
#llama_v3.1_70b
194
- #pytorch backend
195
- - perf/test_perf.py::test_perf[llama_v3.1_70b-bench-pytorch-bfloat16-maxbs:1-input_output_len:512,32-quant:fp8-gpus:8]
+ #trt backend
+ - perf/test_perf.py::test_perf[llama_v3.1_70b-bench-bfloat16-maxbs:1-maxnt:544-input_output_len:512,32-quant:fp8-gpus:8]
196
#llama_v3.3_70b_instruct_fp8
197
#pytorch backend
198
- perf/test_perf.py::test_perf[llama_v3.3_70b_instruct_fp8-bench-pytorch-float8-input_output_len:128,128-gpus:8]
0 commit comments