torch-mlir/python
Stella Laurenzo e2343cf4ce
[fx] Implement auto_functionalized higher order op. (#3063)
* Also adds the basic scaffolding for handling more of these, which will
be needed for cond, while, etc.
* Refactors some of the support in the generic OpOverload emitter so it
can be shared with these other special forms.

This has been on my list for a while, but it just so happens that as
part of upgrading to PyTorch 2.3 and a pure upstream flow in Turbine, we
were using a feature that required integration with auto_functionalized.
This is perhaps the "weirdest" of the higher-order ops and a poor place
to start, but needs must. We have testing for this in Turbine.

Full support in Turbine has an entire custom ops facility. I've reduced
this down to a unit test in torch-mlir.
2024-03-26 17:06:05 -07:00
..
torch_mlir [fx] Implement auto_functionalized higher order op. (#3063) 2024-03-26 17:06:05 -07:00
CMakeLists.txt add support for decomposition (#2879) 2024-02-14 21:00:52 -08:00
TorchMLIRModule.cpp Upstream the ONNX importer. (#2636) 2023-12-12 19:02:51 -08:00