Skip to content

Commit efaefca

Browse files
authored
[None][test] Update case that not support passing quantization fp8 for pytorch backend (#7302)
Signed-off-by: nvamyt <[email protected]>
1 parent b0558c7 commit efaefca

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

tests/integration/test_lists/qa/llm_perf_sanity.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -191,8 +191,8 @@ llm_perf_sanity:
191191

192192
tests:
193193
#llama_v3.1_70b
194-
#pytorch backend
195-
- perf/test_perf.py::test_perf[llama_v3.1_70b-bench-pytorch-bfloat16-maxbs:1-input_output_len:512,32-quant:fp8-gpus:8]
194+
#trt backend
195+
- perf/test_perf.py::test_perf[llama_v3.1_70b-bench-bfloat16-maxbs:1-maxnt:544-input_output_len:512,32-quant:fp8-gpus:8]
196196
#llama_v3.3_70b_instruct_fp8
197197
#pytorch backend
198198
- perf/test_perf.py::test_perf[llama_v3.3_70b_instruct_fp8-bench-pytorch-float8-input_output_len:128,128-gpus:8]

0 commit comments

Comments
 (0)