torch-mlir/lib/Conversion/TorchToLinalg
zjgarvey de28c8540b
[ONNX] add int16 quantization support (#3446)
There is currently no int16 quantization support in torch. This patch
adds a new mlir type to correspond to the missing "torch.qint16" type,
and enables lowering of quantization-related onnx ops using int16 types.

In follow-up patches, custom quantization logic for ops like
aten.matmul/aten.mm/aten.convolution may need to be revisited to allow
support for qint16. The passes in FuseQuantizedOps.cpp may also need
slight modifications.
2024-06-12 10:37:22 +05:30
..
CMakeLists.txt Re-organize project structure to separate PyTorch dependencies from core project. (#2542) 2023-11-02 19:45:55 -07:00
DataMovement.cpp [torch-mlir][sparse] re-enable all sparse tests (#3444) 2024-06-10 11:19:32 -07:00
IndirectDataMovement.cpp [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
Linear.cpp [Torch] fix toBuiltinTensor() (#3415) 2024-06-08 09:36:32 +08:00
Pooling.cpp [MLIR][Torch] Add TorchToLinalg lowering for AtenAvgPool3dOp (#3030) 2024-06-04 22:12:34 +05:30
PopulatePatterns.h Re-enable custom op support 2022-08-16 22:49:08 +05:30
Random.cpp [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
Reduction.cpp [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
TensorConstructors.cpp [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
TensorScalarInterop.cpp [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
TorchToLinalg.cpp [NFC] Remove unused header files (#3386) 2024-05-30 14:30:36 +08:00
Uncategorized.cpp [ONNX] Fix resize ceil numerics and add half_pixel_symmetric support (#3443) 2024-06-11 22:35:50 -05:00
Utils.cpp [ONNX] add int16 quantization support (#3446) 2024-06-12 10:37:22 +05:30