llama.cpp
IBM Granite MoE Architecture
#9438
Merged

IBM Granite MoE Architecture #9438

ggerganov merged 10 commits into ggml-org:master from gabe-l-hart:GraniteMoE
gabe-l-hart
gabe-l-hart gabe-l-hart force pushed 1 year ago
gabe-l-hart gabe-l-hart force pushed 1 year ago
github-actions github-actions added python
compilade
compilade commented on 2024-09-14
gabe-l-hart gabe-l-hart force pushed 1 year ago
gabe-l-hart gabe-l-hart force pushed 1 year ago
gabe-l-hart gabe-l-hart force pushed 1 year ago
gabe-l-hart gabe-l-hart marked this pull request as ready for review 1 year ago
gabe-l-hart gabe-l-hart force pushed 1 year ago
gabe-l-hart
ggerganov
ggerganov approved these changes on 2024-09-23
gabe-l-hart
compilade
compilade commented on 2024-09-23
ggerganov
gabe-l-hart gabe-l-hart force pushed 1 year ago
compilade
compilade commented on 2024-09-23
gabe-l-hart
compilade
compilade approved these changes on 2024-09-23
gabe-l-hart feat(gguf-py): Add granitemoe architecture
8a4ca231
gabe-l-hart feat(convert_hf_to_gguf): Add GraniteMoeModel
e0b72290
gabe-l-hart fix(granitemoe convert): Split the double-sized input layer into gate…
014e59d3
gabe-l-hart feat(granitemoe): Implement granitemoe
eca37cd4
gabe-l-hart Typo fix in docstring
71bc4c1f
gabe-l-hart fix(conversion): Simplify tensor name mapping in conversion
5eb28c47
gabe-l-hart fix(convert): Remove unused tensor name mappings
f2360996
gabe-l-hart fix(convert): Sanity check on merged FFN tensor sizes
317b15bb
gabe-l-hart fix: Allow "output" layer in granite moe architecture (convert and cpp)
1c8b3e4c
gabe-l-hart gabe-l-hart force pushed to 1c8b3e4c 1 year ago
gabe-l-hart fix(granite): Add missing 'output' tensor for Granite
a843f1fa
gabe-l-hart
compilade compilade added merge ready
ggerganov ggerganov merged 3d6bf691 into master 1 year ago
gabe-l-hart gabe-l-hart deleted the GraniteMoE branch 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone