Enable qwen3 vl moe quant and load #1182
refine update_fused_layer_global_scales to fix device mismatch for nv…
09b3a1ce
enable qwen3_vl_moe quantization & quantized model loading
91280321
[pre-commit.ci] auto fixes from pre-commit.com hooks
5d816459
fixtypo
9f61c9e4
Merge branch 'enable_qwen3_vl_moe_quant' of https://github.com/intel/…
8769f04d
yiliu30
approved these changes
on 2025-12-23
Merge branch 'main' into enable_qwen3_vl_moe_quant
a39c5513
Update auto_round/modelling/qwen3_vl_moe.py
2afa2698
set calib_all_experts to false
c268e3dc
fix typo
b9b8914a
yiliu30
merged
d1935065
into main 4 days ago
yiliu30
deleted the enable_qwen3_vl_moe_quant branch 4 days ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub