Skip to content

Commit 2f7c620

Browse files
authored
Remove missing weights silencers in favor of HFQuantizer solution (#1017)
## Purpose ## * Remove warning silencing code that was previously needed for loading quantized models but is now handled by huggingface/transformers#36152 --------- Signed-off-by: Kyle Sayers <[email protected]>
1 parent 23bde49 commit 2f7c620

File tree

2 files changed

+0
-19
lines changed

2 files changed

+0
-19
lines changed

src/llmcompressor/transformers/sparsification/compressed_tensors_utils.py

-9
Original file line numberDiff line numberDiff line change
@@ -86,15 +86,6 @@ def save_pretrained_wrapper(
8686
# https://github.com/huggingface/transformers/pull/30488
8787
transformers.modeling_utils.dtype_byte_size = new_dtype_byte_size
8888

89-
def skip(*args, **kwargs):
90-
pass
91-
92-
# Skip the initializer step. This accelerates the loading
93-
# of the models, especially for the quantized models
94-
torch.nn.init.kaiming_uniform_ = skip
95-
torch.nn.init.uniform_ = skip
96-
torch.nn.init.normal_ = skip
97-
9889
# state_dict gets passed in as a kwarg for FSDP models
9990
state_dict = kwargs.pop("state_dict", None)
10091
if state_dict is None:

tests/llmcompressor/transformers/sparsification/test_compress_tensor_utils.py

-10
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
import logging
21
import math
32
import shutil
43

@@ -70,19 +69,10 @@ def test_sparse_model_reload(compressed, config, dtype, tmp_path):
7069
clear_sparse_session=False,
7170
)
7271

73-
# temporarily set the log level to error, to ignore printing out long missing
74-
# and unexpected key error messages (these are EXPECTED for quantized models)
75-
transformers_logger = logging.getLogger("transformers.modeling_utils")
76-
restore_log_level = transformers_logger.getEffectiveLevel()
77-
transformers_logger.setLevel(level=logging.ERROR)
78-
7972
model = AutoModelForCausalLM.from_pretrained(
8073
tmp_path / "oneshot_out", torch_dtype=dtype
8174
)
8275

83-
# restore transformers logging level now that model shell is loaded
84-
transformers_logger.setLevel(level=restore_log_level)
85-
8676
# assert that sample layer has the intended sparsity
8777
assert math.isclose(
8878
tensor_sparsity(model.state_dict()[one_of_sparse_weights]),

0 commit comments

Comments
 (0)