[Model] Add AWQ quantization support for InternVL2 model #7187
add awq quantization support for internvl2
fe8ba998
code format
cadb2548
Add internvl awq test
f5c94f30
code format
53537381
Merge branch 'vllm-project:main' into awq-vl
bd9a65c3
add awq 26B test
e6ed33ac
Isotr0py
marked this pull request as ready for review 1 year ago
Merge branch 'vllm-project:main' into awq-vl
26f6d06b
revert internvl test
068f3804
code format
8daae6dd
fix failed test and add cpu offload
ffc1a077
26B test use tp_size=2
7726c98b
fix a typo
d2c76658
remove 26B AWQ test
f4c097e8
Merge branch 'vllm-project:main' into awq-vl
cceca324
fix fuse linear loading
177b49b5
Merge branch 'vllm-project:main' into awq-vl
7cba85a2
simon-mo
merged
12e1c65b
into main 1 year ago
Isotr0py
deleted the awq-vl branch 1 year ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub