vllm
[Kernel][Performance] Fuse float cast and renormalize to topk softmax kernel
#26717
Merged

[Kernel][Performance] Fuse float cast and renormalize to topk softmax kernel #26717

izhuhaoran
izhuhaoran perf: fuse to64 and renormalize to topk softmax kernel
f0abf853
izhuhaoran fix lint error
a0ca4e7b
izhuhaoran izhuhaoran requested a review from mgoin mgoin 62 days ago
gemini-code-assist
gemini-code-assist commented on 2025-10-13
chatgpt-codex-connector
chatgpt-codex-connector commented on 2025-10-13
izhuhaoran Merge branch 'main' into fuse-topk-softmax
2d64ff17
youkaichao
izhuhaoran
izhuhaoran Merge branch 'main' into fuse-topk-softmax
28a11a52
mgoin
mgoin
mgoin commented on 2025-10-15
izhuhaoran Merge branch 'main' into fuse-topk-softmax
969c425b
izhuhaoran fix: use TORCH_CHECK for int64 dtype of topk_indices
34cfff69
izhuhaoran feat: support float16 for fused_topk_softmax
b20c66b3
izhuhaoran refactor: use template dispatch_topk_softmax_launch for diff gating_o…
02eb929c
izhuhaoran
izhuhaoran refactor: remove unused assert for toFloat in topk_softmax
dac7677e
izhuhaoran izhuhaoran requested a review from mgoin mgoin 59 days ago
mgoin
mgoin approved these changes on 2025-10-16
mgoin mgoin enabled auto-merge (squash) 59 days ago
github-actions github-actions added ready
izhuhaoran Merge branch 'main' into fuse-topk-softmax
0bd79b64
izhuhaoran
mgoin mgoin merged 75c7ad99 into main 58 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone