torch-mlir/python/test
Sean Silva af9e8a5e63 [torchdynamo] Move to aot_autograd instead of raw make_fx
As [@ezyang suggested](https://github.com/pytorch/pytorch/issues/90276#issuecomment-1339791275),
use `torch._dynamo.optimizations.training.aot_autograd` instead of raw
`make_fx`. This is more future proof and gives us the backward pass and
functionalization. We don't currently get functionalization because of
https://github.com/pytorch/pytorch/issues/90759

This also incidentally fixes the source location handling, which makes
`lockstep_basic.py` give an accurate source location!
2022-12-15 01:55:50 -08:00
..
compile_api Automatically strip overloads for FX-based models. 2022-11-29 22:19:09 -05:00
debug [torchdynamo] Move to aot_autograd instead of raw make_fx 2022-12-15 01:55:50 -08:00
lazy_backend Fix LTC lib_torch_mlir_ltc.so import error (#1283) 2022-08-25 18:25:01 -04:00
torchscript_e2e_test Remove "torchscript" association from the e2e framework. 2022-08-29 14:10:03 -07:00
CMakeLists.txt Move external/torch-mlir to the root of the repo. 2021-09-27 17:11:08 -07:00
annotations-sugar.py Remove "torchscript" association from the e2e framework. 2022-08-29 14:10:03 -07:00
lit.cfg.py Miscellaneous fixes for Windows builds (#1376) 2022-09-29 12:07:43 -05:00
lit.site.cfg.py.in Dual license the torch-mlir project. 2021-10-01 10:46:08 -07:00