vllm
[Model] Add AWQ quantization support for InternVL2 model
#7187
Merged

[Model] Add AWQ quantization support for InternVL2 model #7187

simon-mo merged 16 commits into vllm-project:main from Isotr0py:awq-vl
Isotr0py
Isotr0py add awq quantization support for internvl2
fe8ba998
github-actions
Isotr0py code format
cadb2548
Isotr0py Add internvl awq test
f5c94f30
Isotr0py code format
53537381
Isotr0py Merge branch 'vllm-project:main' into awq-vl
bd9a65c3
Isotr0py add awq 26B test
e6ed33ac
Isotr0py
github-actions github-actions added ready
Isotr0py
Isotr0py Isotr0py marked this pull request as ready for review 1 year ago
Isotr0py Merge branch 'vllm-project:main' into awq-vl
26f6d06b
Isotr0py revert internvl test
068f3804
Isotr0py code format
8daae6dd
Isotr0py fix failed test and add cpu offload
ffc1a077
Isotr0py
Isotr0py
DarkLight1337
Isotr0py
Isotr0py 26B test use tp_size=2
7726c98b
Isotr0py fix a typo
d2c76658
Isotr0py remove 26B AWQ test
f4c097e8
Isotr0py Merge branch 'vllm-project:main' into awq-vl
cceca324
Isotr0py fix fuse linear loading
177b49b5
Isotr0py Merge branch 'vllm-project:main' into awq-vl
7cba85a2
Isotr0py
DarkLight1337
DarkLight1337 approved these changes on 2024-08-21
simon-mo simon-mo merged 12e1c65b into main 1 year ago
Isotr0py Isotr0py deleted the awq-vl branch 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone