Skip to content

Conversation

@yiliu30
Copy link
Owner

@yiliu30 yiliu30 commented Sep 5, 2025

SUMMARY:
"please provide a brief summary"

TEST PLAN:
"please outline how the changes were tested"

yiliu30 and others added 5 commits September 4, 2025 08:24
Signed-off-by: yiliu30 <[email protected]>
SUMMARY:
With the newest transformer change,
`test_kv_cache_gptq_model_state_dict_attr` is failing because it's
initializing empty weights on meta device and attempting to decompress
on meta device. I don't think this is the expected usage. When
model_decompress is called, the weights should finish being loaded
already.


TEST PLAN:
tested locally with the following command and passed:
pytest test
tests/llmcompressor/transformers/kv_cache/test_kv_cache.py::test_kv_cache_gptq_model_state_dict_attr

---------

Signed-off-by: shanjiaz <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants