convert : handle compressed-tensors quant method #17069
convert : handle compressed-tensors quant method
33dba6ce
convert : handle int-quantized models
d23bdd57
convert : handle naive-quantized models
33dcb44a
gguf-py : __pos__ is also unary
987862ad
convert : fix flake8 lint
3770d941
convert : use F32 for dequant of pack-quantized tensors
128118fd
CISC
approved these changes
on 2025-11-07
ngxson
approved these changes
on 2025-11-07
compilade
merged
1c07c0c6
into master 99 days ago
Assignees
No one assigned
Labels
enhancement
python
Login to write a write a comment.
Login via GitHub