onnxruntime
[JS/WebGU] Support fp16 in Attention by performing the computation in fp32.
#20486
Merged

[JS/WebGU] Support fp16 in Attention by performing the computation in fp32. #20486

satyajandhyala
satyajandhyala Take case fp16 support
4191156a
satyajandhyala satyajandhyala changed the title [JS/WebGU] Take case fp16 support [JS/WebGU] Take case fp16 support in Attention by performing the computation in fp32. 1 year ago
satyajandhyala Clean-up the code related to RelativePositionBias.
7c61867c
guschmue guschmue added ep:WebGPU
satyajandhyala satyajandhyala changed the title [JS/WebGU] Take case fp16 support in Attention by performing the computation in fp32. [JS/WebGU] Support fp16 in Attention by performing the computation in fp32. 1 year ago
satyajandhyala satyajandhyala requested a review from guschmue guschmue 1 year ago
satyajandhyala satyajandhyala requested a review from fs-eire fs-eire 1 year ago
satyajandhyala satyajandhyala added release:1.18.0
satyajandhyala satyajandhyala marked this pull request as ready for review 1 year ago
guschmue
guschmue approved these changes on 2024-04-27
satyajandhyala satyajandhyala merged 736cbb39 into main 1 year ago
satyajandhyala satyajandhyala deleted the sajandhy/webgpu_attention_fp16_support_fit branch 1 year ago
sophies927 sophies927 added triage:approved
yihonglyu yihonglyu added cherry-picked
yihonglyu yihonglyu added rel-merged

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone