[Llama4] Enable attention temperature tuning by default for long context (>32k) #16439
turn on attn temp tuning by default
7fe8f463
default attn temp tuning for >32k only
7575935b
houseroad
approved these changes
on 2025-04-11
mgoin
approved these changes
on 2025-04-11
vllm-bot
merged
99ef59cf
into main 288 days ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub