Skip to content

Commit 2e03d34

Browse files
release 0.43.3
1 parent b64cbe3 commit 2e03d34

File tree

3 files changed

+11
-2
lines changed

3 files changed

+11
-2
lines changed

CHANGELOG.md

+9
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,12 @@
1+
### 0.43.3
2+
3+
#### Improvements:
4+
5+
- FSDP: Enable loading prequantized weights with bf16/fp16/fp32 quant_storage
6+
- Background: This update, linked to [Transformer PR #32276](https://github.com/huggingface/transformers/pull/32276), allows loading prequantized weights with alternative storage formats. Metadata is tracked similarly to `Params4bit.__new__` post PR #970. It supports models exported with non-default `quant_storage`, such as [this NF4 model with BF16 storage](https://huggingface.co/hugging-quants/Meta-Llama-3.1-405B-BNB-NF4-BF16).
7+
- Special thanks to @winglian and @matthewdouglas for enabling FSDP+QLoRA finetuning of Llama 3.1 405B on a single 8xH100 or 8xA100 node with as little as 256GB system RAM.
8+
9+
110
### 0.43.2
211

312
This release is quite significant as the QLoRA bug fix big implications for higher `seqlen` and batch sizes.

bitsandbytes/__init__.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -21,4 +21,4 @@
2121
"optim.optimizer.MockArgs": False,
2222
}
2323

24-
__version__ = "0.43.3.dev"
24+
__version__ = "0.43.3"

setup.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ def has_ext_modules(self):
2525

2626
setup(
2727
name="bitsandbytes",
28-
version="0.43.3.dev",
28+
version="0.43.3",
2929
author="Tim Dettmers",
3030
author_email="[email protected]",
3131
description="k-bit optimizers and matrix multiplication routines.",

0 commit comments

Comments
 (0)