onnxruntime
330b8e74
- Fix attention parity for GPT-2 (#8549)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
4 years ago
Fix attention parity for GPT-2 (#8549) * Use persistent softmax to parity with huggingface * fix undirectional mask logic * add test
References
#8549 - Fix attention parity for GPT-2
Author
tianleiwu
Parents
816ad86d
Loading