auto-round
feat: handle per-tensor FP8 dequantization for Devstral models
#1356
Merged

feat: handle per-tensor FP8 dequantization for Devstral models #1356

SwekeR-463
SwekeR-463 feat: handle per-tensor FP8 dequantization for Devstral models
8ad08868
pre-commit-ci[bot] [pre-commit.ci] auto fixes from pre-commit.com hooks
d43b6f19
yiliu30
yiliu30
yiliu30 Merge branch 'main' into feat/fp8-per-tensor-dequant
982e8866
SwekeR-463
SwekeR-463 SwekeR-463 closed this 11 days ago
SwekeR-463 SwekeR-463 reopened this 11 days ago
SwekeR-463
yiliu30 Merge branch 'main' into feat/fp8-per-tensor-dequant
06f55413
yiliu30
SwekeR-463 Merge origin/main and resolve conflicts in model.py
9c0121ea
SwekeR-463 refactor for signature and tests with new transformer version
29c12dad
SwekeR-463
yiliu30
yiliu30 commented on 2026-01-31
SwekeR-463 refactor: rename `dequant_block_fp8_weight` to `_dequant_fp8_linear_w…
2c502915
pre-commit-ci[bot] [pre-commit.ci] auto fixes from pre-commit.com hooks
4d258d27
yiliu30
yiliu30 approved these changes on 2026-02-01
yiliu30 yiliu30 merged eddb6de2 into main 7 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone