pytorch
553eaaba - Disable tf32 in functorch transform tests (#86799)

Commit
2 years ago
Disable tf32 in functorch transform tests (#86799) This PR applies a large hammer and disables TF32 in specific functorch transform tests. TF32 isn't precise enough to test correctness. We could have applied a smaller hammer by disabling TF32 per-OpInfo, but that doesn't seem to have too much additional benefit (e.g. if a convolution batching rule is correct on fp32 then I would expect it to be correct under TF32 modulo precision issues because the actual sequence of PyTorch operators we invoke has not changed, only the backend did). Test Plan: - I tested this locally on a machine with A100 GPUs. Fixes #ISSUE_NUMBER Pull Request resolved: https://github.com/pytorch/pytorch/pull/86799 Approved by: https://github.com/malfet
Author
Committer
Parents
Loading