llama.cpp
Fix kq_scale for the attention layers of PLaMo2
#14892
Merged

Fix kq_scale for the attention layers of PLaMo2 #14892

mitmul
mitmul Fix dimensions for expand
7baf4fd1
mitmul Change dimensions to copy states to cache
e39bc092
mitmul Fix the default value for plamo2 conversion
bd4d2e1c
mitmul Fix scale given to build_attn
c475203c
github-actions github-actions added python
CISC
CISC approved these changes on 2025-07-26
CISC
CISC commented on 2025-07-26
mitmul Update src/llama-model.cpp
75f0a0d7
mitmul Update src/llama-model.cpp
429639d0
mitmul Update src/llama-model.cpp
60a705de
CISC CISC merged 1dc9614e into master 139 days ago
mitmul mitmul deleted the mitmul/fix-build-attn-scale-plamo2 branch 139 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone