torch-mlir/lib/Conversion/TorchOnnxToTorch
zjgarvey de28c8540b
[ONNX] add int16 quantization support (#3446)
There is currently no int16 quantization support in torch. This patch
adds a new mlir type to correspond to the missing "torch.qint16" type,
and enables lowering of quantization-related onnx ops using int16 types.

In follow-up patches, custom quantization logic for ops like
aten.matmul/aten.mm/aten.convolution may need to be revisited to allow
support for qint16. The passes in FuseQuantizedOps.cpp may also need
slight modifications.
2024-06-12 10:37:22 +05:30
..
CMakeLists.txt [ONNX] Support onnx.LSTM (#2969) 2024-04-08 12:23:33 -07:00
DefaultDomainAtoF.cpp [ONNX] add int16 quantization support (#3446) 2024-06-12 10:37:22 +05:30
DefaultDomainGtoP.cpp [ONNX] add int16 quantization support (#3446) 2024-06-12 10:37:22 +05:30
DefaultDomainQtoZ.cpp [ONNX] add int16 quantization support (#3446) 2024-06-12 10:37:22 +05:30
OnnxLstmExpander.cpp [ONNX] Support onnx.LSTM (#2969) 2024-04-08 12:23:33 -07:00
PassDetail.h Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Passes.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Patterns.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
TorchOnnxToTorch.cpp Initial TorchOnnxToTorch conversion pipeline. (#2585) 2023-11-21 21:02:55 -08:00
Utils.cpp [ONNX] add int16 quantization support (#3446) 2024-06-12 10:37:22 +05:30