vllm
99ef59cf - [Llama4] Enable attention temperature tuning by default for long context (>32k) (#16439)

Commit
288 days ago
[Llama4] Enable attention temperature tuning by default for long context (>32k) (#16439) Signed-off-by: Ye (Charlotte) Qi <yeq@meta.com> Co-authored-by: Ye (Charlotte) Qi <yeq@meta.com>
Author
Parents
Loading