vllm
99ef59cf
- [Llama4] Enable attention temperature tuning by default for long context (>32k) (#16439)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
288 days ago
[Llama4] Enable attention temperature tuning by default for long context (>32k) (#16439) Signed-off-by: Ye (Charlotte) Qi <yeq@meta.com> Co-authored-by: Ye (Charlotte) Qi <yeq@meta.com>
References
#16439 - [Llama4] Enable attention temperature tuning by default for long context (>32k)
Author
sarckk
Parents
d544d141
Loading