torch-mlir/python
Aart Bik be8375d350
[torch-mlir][sparse] implement first sparse_jit end-to-end path (#2894)
This PR introduces a sparse_jit wrapper that can run simple models with
sparse tensor inputs end-to-end. The implementation shows all required
components on modifying sparse tensor types with a 1:N relation on the
call sites. Two tests shows that the JIT runs end-to-end while computing
the correct results.

More details to follow (generalizing to COO and different ranks, as well
as support for *output* sparse tensors), but the general concepts are
all here now.

**_Update: Thanks to Rob, bump to proper LLVM/MLIR hash is done!_**

_**NOTE that all parameter passing changes are nicely done "downstream"
in MLIR, so very little changes are required in torch-mlir code
proper**_

---------

Co-authored-by: Franz Haniel <77495327+frafranz@users.noreply.github.com>
Co-authored-by: Franz Haniel <franz.haniel@amd.com>
2024-02-12 10:04:54 -08:00
..
torch_mlir [torch-mlir][sparse] implement first sparse_jit end-to-end path (#2894) 2024-02-12 10:04:54 -08:00
CMakeLists.txt Rename torch_mlir.compile APIs and introduce FX based analogs (#2842) 2024-02-06 19:07:59 -08:00
TorchMLIRModule.cpp Upstream the ONNX importer. (#2636) 2023-12-12 19:02:51 -08:00