[PyTorch] Hook CUDA LayerNormKernel up for dispatch (#74259)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/74259
The CPU one is hooked up but the CUDA one isn't. I want to use the CUDA one too.
Test Plan: builds
Reviewed By: ngimel
Differential Revision: D34697478
fbshipit-source-id: 2a7e7adc1b7f90d3a9ee911083b2e286deb3b547
(cherry picked from commit 60ec5d19c485d24ca55967593f3b3cc5a485cec6)