add support for mutated params (#106098)
Previously, this didn't work because of the warmup run. Now that we do not run warmup, and then execution on one inductor invocation this works. llama inference 1.6->4.4
Pull Request resolved: https://github.com/pytorch/pytorch/pull/106098
Approved by: https://github.com/ezyang