Skip to content

Commit 6861f5f

Browse files
rand-flyhuachenheli
authored andcommitted
[Bugfix] fix adding bias twice in ipex GPTQ quantization (vllm-project#18363)
Signed-off-by: rand-fly <[email protected]> Signed-off-by: Chenheli Hua <[email protected]>
1 parent 2f882df commit 6861f5f

File tree

1 file changed

+0
-2
lines changed

1 file changed

+0
-2
lines changed

vllm/model_executor/layers/quantization/ipex_quant.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -181,8 +181,6 @@ def apply(self,
181181
bias: Optional[torch.Tensor] = None) -> torch.Tensor:
182182
reshaped_x = x.reshape(-1, x.shape[-1])
183183
out = layer.ipex_qlinear(reshaped_x)
184-
if bias is not None:
185-
out.add_(bias)
186184
return out.reshape(x.shape[:-1] + (layer.ipex_output_size, ))
187185

188186

0 commit comments

Comments
 (0)