torch-mlir/lib
zjgarvey 6ff71b40c8
[ONNX] onnx.DynamicQuantizeLinear to Torch (#3009)
This adds support for converting DynamicQuantizeLinear from torch-onnx
to torch.

I could not get an e2e test to pass, since there seems to be some issues
with uint8 casting somewhere lower in the pipeline. For example
compiling with IREE for llvm-cpu, I would get either the correct zero
point (if zp < 128) or the correct zero-point minus 256 (if zp >= 128).
The output tensor seems to always return a tensor of zeros, which also
occurs when running uint8 examples through QuantizeLinear.

Edit: the first problem can be resolved by casting the output back to
uint8 on output, the second problem is resolved with PR #3018
2024-03-20 10:58:25 -07:00
..
CAPI Clang format refresh (#2812) 2024-01-29 12:59:33 -05:00
Conversion [ONNX] onnx.DynamicQuantizeLinear to Torch (#3009) 2024-03-20 10:58:25 -07:00
Dialect [Stablehlo] lowering chlo to stablehlo in torch-to-stablehlo pipeline (#3037) 2024-03-19 21:18:54 +08:00
RefBackend [TorchToLinalg] Lower aten.cat to tensor.concat (#2650) 2023-12-15 15:45:32 -05:00
CMakeLists.txt [Stablehlo] lowering chlo to stablehlo in torch-to-stablehlo pipeline (#3037) 2024-03-19 21:18:54 +08:00
InitAll.cpp [Stablehlo] lowering chlo to stablehlo in torch-to-stablehlo pipeline (#3037) 2024-03-19 21:18:54 +08:00