llama.cpp
Fix ffn_down quantization mix for MoE models
#4927
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
3
Changes
View On
GitHub
Fix ffn_down quantization mix for MoE models
#4927
ikawrakow
merged 3 commits into
master
from
ik/fix_qxm_moe
Fix ffn_down quantization mix for MoE models
998b635a
Fix the fix
121eb066
ggerganov
approved these changes on 2024-01-14
Review suggestion
00cc67e2
ikawrakow
merged
a128c38d
into master
1 year ago
ikawrakow
deleted the ik/fix_qxm_moe branch
1 year ago
Login to write a write a comment.
Login via GitHub
Reviewers
ggerganov
Assignees
No one assigned
Labels
None yet
Milestone
No milestone
Login to write a write a comment.
Login via GitHub