llama.cpp
llama : store non-RoPEd K cache
#3234
Open
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
llama : store non-RoPEd K cache
#3234
ggerganov
wants to merge 1 commit into
custom-attention-mask
from
custom-attention-mask-no-roped-cache
llama : store non-RoPEd K cache (WIP)
784d14ed
ggerganov
added
demo
ggerganov
force-pushed the
custom-attention-mask
branch
from
5bda9e27
to
0161372b
2 years ago
Login to write a write a comment.
Login via GitHub
Reviewers
No reviews
Assignees
No one assigned
Labels
demo
Milestone
No milestone
Login to write a write a comment.
Login via GitHub