pytorch
fcec27f7 - [MPS] Numerical stability and reduction fixes (#95317)

Commit
1 year ago
[MPS] Numerical stability and reduction fixes (#95317) * [MPS] Fixes for LSTM. (#94889) - Backward pass has to give explicit bias tensor of zeros if none is passed to the op or the bias gradient will not be calculated. - Fixed bias tensor mistakenly getting overwritten to zeros - Fixes crash when lstm op called with has_biases set to false. Change takes into account the changed shape of the input params TensorList depending on the bias flag. Fixes #ISSUE_NUMBER Pull Request resolved: https://github.com/pytorch/pytorch/pull/94889 Approved by: https://github.com/DenisVieriu97 * [MPS] LogSoftmax numerical stability (#95091) Fixes #94043 Calculations are now consistent with numericaly stable formula and CPU: $LogSoftmax(X, \dim) = X - \max(X, \dim) - \log(sum(X - \max(X, \dim), \dim))$ @malfet Pull Request resolved: https://github.com/pytorch/pytorch/pull/95091 Approved by: https://github.com/malfet, https://github.com/kulinseth * [MPS] Cast int64 to int32 for reduction ops (#95231) - give warnings of converting int64 for reduction ops - use cast tensor for reduction sum on trace - unblock trace from running Pull Request resolved: https://github.com/pytorch/pytorch/pull/95231 Approved by: https://github.com/razarmehr * [MPS] Fix Float16 issue with Reduction ops for macOS 12 (#94952) This would fix the issue with `__rdiv__` with float16 Pull Request resolved: https://github.com/pytorch/pytorch/pull/94952 Approved by: https://github.com/kulinseth --------- Co-authored-by: alexdremov <dremov.me@gmail.com> Co-authored-by: Denis Vieriu <dvieriu@apple.com> Co-authored-by: Ramin Azarmehr <razarmehr@apple.com>
Author
Parents
Loading