Skip to content

Commit 4c8f036

Browse files
committed
Enable AWQ on Intel GPU.
1 parent a8d1a46 commit 4c8f036

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

torchao/dtypes/uintx/int4_xpu_layout.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -105,9 +105,9 @@ def _linear_fp_act_uint4_weight_int8_zero_check(input_tensor, weight_tensor, bia
105105

106106

107107
def _linear_fp_act_uint4_weight_int8_zero_impl(input_tensor, weight_tensor, bias):
108-
assert weight_tensor.block_size[0] == 1, (
109-
f"Requires groupwise quantization, got block_size: {weight_tensor.block_size}"
110-
)
108+
assert (
109+
weight_tensor.block_size[0] == 1
110+
), f"Requires groupwise quantization, got block_size: {weight_tensor.block_size}"
111111
assert input_tensor.shape[-1] == weight_tensor.shape[1], (
112112
f"need input_tensor shape: {input_tensor.shape} final"
113113
f"dim to match weight_tensor shape: {weight_tensor.shape} second dim "

0 commit comments

Comments
 (0)