We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4e66251 commit 08b98a9Copy full SHA for 08b98a9
src/llmcompressor/observers/base.py
@@ -214,13 +214,13 @@ def get_qparams(
214
self._scale = torch.empty(
215
(num_br, num_bc), dtype=observed.dtype, device=observed.device
216
)
217
-
+
218
# Use same dtype logic as GROUP strategy for zero_point
219
if is_fp4(quantization_args=self.quantization_args):
220
zp_dtype = FP8_E4M3_DATA.dtype
221
else:
222
zp_dtype = self.quantization_args.pytorch_dtype()
223
224
self._zero_point = torch.empty(
225
(num_br, num_bc), dtype=zp_dtype, device=observed.device
226
0 commit comments