Skip to content

Commit 3892e58

Browse files
authored
[Misc] Upgrade BNB version (#15183)
1 parent d20e261 commit 3892e58

File tree

4 files changed

+10
-10
lines changed

4 files changed

+10
-10
lines changed

Diff for: Dockerfile

+1-1
Original file line numberDiff line numberDiff line change
@@ -286,7 +286,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
286286
if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
287287
uv pip install accelerate hf_transfer 'modelscope!=1.15.0' 'bitsandbytes>=0.42.0' 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3]; \
288288
else \
289-
uv pip install accelerate hf_transfer 'modelscope!=1.15.0' 'bitsandbytes>=0.45.0' 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3]; \
289+
uv pip install accelerate hf_transfer 'modelscope!=1.15.0' 'bitsandbytes>=0.45.3' 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3]; \
290290
fi
291291

292292
ENV VLLM_USAGE_SOURCE production-docker-image

Diff for: docs/source/features/quantization/bnb.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Compared to other quantization methods, BitsAndBytes eliminates the need for cal
99
Below are the steps to utilize BitsAndBytes with vLLM.
1010

1111
```console
12-
pip install bitsandbytes>=0.45.0
12+
pip install bitsandbytes>=0.45.3
1313
```
1414

1515
vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.

Diff for: vllm/model_executor/layers/quantization/bitsandbytes.py

+4-4
Original file line numberDiff line numberDiff line change
@@ -155,12 +155,12 @@ class BitsAndBytesLinearMethod(LinearMethodBase):
155155
def __init__(self, quant_config: BitsAndBytesConfig):
156156
try:
157157
import bitsandbytes
158-
if bitsandbytes.__version__ < "0.45.0":
158+
if bitsandbytes.__version__ < "0.45.3":
159159
raise ImportError("bitsandbytes version is wrong. Please "
160-
"install bitsandbytes>=0.45.0.")
160+
"install bitsandbytes>=0.45.3.")
161161
except ImportError as err:
162-
raise ImportError("Please install bitsandbytes>=0.45.0 via "
163-
"`pip install bitsandbytes>=0.45.0` to use "
162+
raise ImportError("Please install bitsandbytes>=0.45.3 via "
163+
"`pip install bitsandbytes>=0.45.3` to use "
164164
"bitsandbytes quantizer.") from err
165165

166166
self.quant_config = quant_config

Diff for: vllm/model_executor/model_loader/loader.py

+4-4
Original file line numberDiff line numberDiff line change
@@ -862,12 +862,12 @@ def _get_quantized_weights_iterator(
862862
try:
863863
import bitsandbytes
864864

865-
if bitsandbytes.__version__ < "0.45.0":
865+
if bitsandbytes.__version__ < "0.45.3":
866866
raise ImportError("bitsandbytes version is wrong. Please "
867-
"install bitsandbytes>=0.45.0.")
867+
"install bitsandbytes>=0.45.3.")
868868
except ImportError as err:
869-
raise ImportError("Please install bitsandbytes>=0.45.0 via "
870-
"`pip install bitsandbytes>=0.45.0` to use "
869+
raise ImportError("Please install bitsandbytes>=0.45.3 via "
870+
"`pip install bitsandbytes>=0.45.3` to use "
871871
"bitsandbytes quantizer.") from err
872872

873873
hf_weights_files, use_safetensors = self._prepare_weights(

0 commit comments

Comments
 (0)