vllm
[Llama4] Enable attention temperature tuning by default for long context (>32k)
#16439
Merged

[Llama4] Enable attention temperature tuning by default for long context (>32k) #16439

sarckk
yeqcharlotte turn on attn temp tuning by default
7fe8f463
yeqcharlotte default attn temp tuning for >32k only
7575935b
sarckk sarckk requested a review from houseroad houseroad 288 days ago
github-actions
yeqcharlotte
yeqcharlotte
yeqcharlotte approved these changes on 2025-04-11
houseroad
houseroad approved these changes on 2025-04-11
houseroad houseroad added ready
mgoin
mgoin approved these changes on 2025-04-11
vllm-bot vllm-bot merged 99ef59cf into main 288 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone