v0.1.0
What's Changed
- Address Test Failures by @Satrat in #1
- Remove SparseZoo Usage by @Satrat in #2
- SparseML Cleanup by @markurtz in #6
- Remove all references to Neural Magic copyright within LLM Compressor by @markurtz in #7
- Add FP8 Support by @Satrat in #4
- Fix Weekly Test Failure by @Satrat in #8
- Add Scheme UX for QuantizationModifier by @Satrat in #9
- Add Group Quantization Test Case by @Satrat in #10
- Loguru logging standardization for LLM Compressor by @markurtz in #11
- Clarify Function Names for Logging by @Satrat in #12
- [ Examples ] E2E Examples by @robertgshaw2-neuralmagic in #5
- Update setup.py by @robertgshaw2-neuralmagic in #15
- SmoothQuant Mapping Defaults by @Satrat in #13
- Initial README by @bfineran in #3
- [Bug] Fix validation errors for smoothquant modifier + update examples by @rahul-tuli in #19
- [MOE Quantization] Warn against "undercalibrated" modules by @dbogunowicz in #20
- Port SparseML Remote Code Fix by @Satrat in #21
- Update Quantization Save Defaults by @Satrat in #22
- [Bugfix] Add fix to preserve modifier order when passed as a list by @rahul-tuli in #26
- GPTQ - move calibration of quantiztion params to after hessian calibration by @bfineran in #25
- Fix typos by @eldarkurtic in #31
- Remove ceiling from
datasets
dep by @mgoin in #27 - Revert naive compression format by @Satrat in #32
- Fix layerwise targets by @Satrat in #36
- Move Weight Update Out Of Loop by @Satrat in #40
- Fix End Epoch Default by @Satrat in #39
- Fix typos in example for w8a8 quant by @eldarkurtic in #38
- Model Offloading Support Pt 2 by @Satrat in #34
- set version to 1.0.0 for release by @bfineran in #44
- Update version for first release by @markurtz in #50
- BugFix: Update TRL example scripts to point to the right SFTTrainer by @rahul-tuli in #51
- Update examples/quantization_24_sparse_w4a16 README by @dbarbuzzi in #52
- Fix Failing Transformers Tests by @Satrat in #53
- Offloading Bug Fix by @Satrat in #58
New Contributors
- @markurtz made their first contribution in #6
- @bfineran made their first contribution in #3
- @dbogunowicz made their first contribution in #20
- @eldarkurtic made their first contribution in #31
- @mgoin made their first contribution in #27
- @dbarbuzzi made their first contribution in #52
Full Changelog: https://github.com/vllm-project/llm-compressor/commits/0.1.0