vllm
27b78c73
- [Kernel] add triton fused moe kernel for gptq/awq (#12185)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
166 days ago
[Kernel] add triton fused moe kernel for gptq/awq (#12185)
References
#12185 - [Kernel] add triton fused moe kernel for gptq/awq
Author
jinzhen-lin
Parents
b02fd288
Files
4
tests/kernels
test_moe.py
vllm/model_executor/layers
fused_moe
fused_moe.py
quantization
__init__.py
moe_wna16.py
Loading