-
Notifications
You must be signed in to change notification settings - Fork 61
Issues: NVIDIA/TensorRT-Model-Optimizer
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Which instructions should I follow to quant my model from bf16 to nvfp4?
#164
opened Mar 28, 2025 by
ghostplant
Large model offloaded to huggingface accelerate is not able to export the weights using unified export.
bug
Something isn't working
#157
opened Mar 18, 2025 by
michaelfeil
int4 quantization output onnx does not load
bug
Something isn't working
#156
opened Mar 13, 2025 by
thejaswi01
PyTorch Quantization Failed to Quantize Scaled Dot Product
#149
opened Mar 7, 2025 by
YixuanSeanZhou
sm_100 not defined for option gpu-name when running calibration in DeepSeek
#144
opened Mar 4, 2025 by
imenselmi
Restore functionality: lm_head option to disable quantization
#138
opened Feb 20, 2025 by
michaelfeil
More modes for model opt quantization than halving the batch size
#133
opened Feb 18, 2025 by
michaelfeil
Previous Next
ProTip!
Updated in the last three days: updated:>2025-03-27.