[inductor] Make AOT CPU Inductor work in fbcode (#106225)
Summary:
This diff has a couple of hacks to make inductor-CPU work for AOT codegen in fbcode:
- We need to add the CUDA link flags; AOT-Inductor is specialized for CUDA
right now and uses a lot of `at::cuda` stuff. We should do a proper AOT CPU
at some point but this unblocks perf measurement.
- Add an include path to the cpp_prefix. It's kind of hilarious; we remove the
include path for remote execution, but then for AOT we need it back. :shrug:
Test Plan: internal test
Differential Revision: D47882848
Pull Request resolved: https://github.com/pytorch/pytorch/pull/106225
Approved by: https://github.com/mikekgfb, https://github.com/bdhirsh, https://github.com/jansel