torch-mlir/lib/Conversion/TorchOnnxToTorch
Rob Suderman 60bf6c25af
[onnx] Lower `onnx.QLinearMatMul` lowering to `torch` operators (#2776)
We can plumb the linear matmul into pytorch using its quantized types
with side channel information. To handle the final int8 operation we
dequantize and requantize.
2024-01-24 12:28:48 -08:00
..
CMakeLists.txt [MLIR][ONNX] Add OnnxToTorch support for Maxpool Op (#2695) 2024-01-12 14:54:38 -08:00
DefaultDomainAtoF.cpp [ONNX][MLIR] Add support for onnx.Exp op (#2792) 2024-01-23 13:45:00 -08:00
DefaultDomainGtoP.cpp [MLIR][TORCH] Support for `onnx.LayerNormalization` (#2789) 2024-01-24 11:08:20 -08:00
DefaultDomainQtoZ.cpp [onnx] Lower `onnx.QLinearMatMul` lowering to `torch` operators (#2776) 2024-01-24 12:28:48 -08:00
PassDetail.h Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Passes.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Patterns.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
TorchOnnxToTorch.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Utils.cpp [MLIR][ONNX] Add OnnxToTorch support for Maxpool Op (#2695) 2024-01-12 14:54:38 -08:00