torch-mlir/projects/pt1
Aart Bik be8375d350
[torch-mlir][sparse] implement first sparse_jit end-to-end path (#2894)
This PR introduces a sparse_jit wrapper that can run simple models with
sparse tensor inputs end-to-end. The implementation shows all required
components on modifying sparse tensor types with a 1:N relation on the
call sites. Two tests shows that the JIT runs end-to-end while computing
the correct results.

More details to follow (generalizing to COO and different ranks, as well
as support for *output* sparse tensors), but the general concepts are
all here now.

**_Update: Thanks to Rob, bump to proper LLVM/MLIR hash is done!_**

_**NOTE that all parameter passing changes are nicely done "downstream"
in MLIR, so very little changes are required in torch-mlir code
proper**_

---------

Co-authored-by: Franz Haniel <77495327+frafranz@users.noreply.github.com>
Co-authored-by: Franz Haniel <franz.haniel@amd.com>
2024-02-12 10:04:54 -08:00
..
e2e_testing Add lowering support for math::AbsIOp (#2875) 2024-02-08 14:53:40 -08:00
examples Rename torch_mlir.compile APIs and introduce FX based analogs (#2842) 2024-02-06 19:07:59 -08:00
python [torch-mlir][sparse] implement first sparse_jit end-to-end path (#2894) 2024-02-12 10:04:54 -08:00
test Rename torch_mlir.compile APIs and introduce FX based analogs (#2842) 2024-02-06 19:07:59 -08:00
tools Re-organize project structure to separate PyTorch dependencies from core project. (#2542) 2023-11-02 19:45:55 -07:00
CMakeLists.txt Re-organize project structure to separate PyTorch dependencies from core project. (#2542) 2023-11-02 19:45:55 -07:00