-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] Fix saving offloaded state dict #172
Conversation
👋 Hi! Thank you for contributing to llm-compressor. Please add the ready label when the PR is ready for review. |
I'm pretty sure this should be merged, but I want to fully test checkpoint loading before this is merged in |
…llm-project/llm-compressor into kylesayrs/fix-offloaded-saving
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. Just one confirmation question
src/llmcompressor/transformers/sparsification/compressed_tensors_utils.py
Show resolved
Hide resolved
@kylesayrs can you resolve the conflict |
@kylesayrs Seems like some of the errors for the sparsification tests are new? |
I addressed the cuda errors by separating out GPU-dependent tests |
tests/llmcompressor/transformers/sparsification/test_compress_tensor_utils.py
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Purpose
Changes
save_safetensors
kwarg, since the originalsave_pretrained
function already defaults its value toTrue
Testing
Main
This branch