torch-mlir/lib/Dialect/Torch
zjgarvey 0fb8b017d8
Adds misc fixes for some padding related issues (#3528)
This patch adds a few misc pad op related changes:

1. Addresses issue <https://github.com/llvm/torch-mlir/issues/3457>
2. Addresses issue <https://github.com/llvm/torch-mlir/issues/3442>
3. Fixes the padding order for asymmetrically padded onnx.Conv ops
4. Enables passing quantization through those onnx.Conv op pre-paddings
5. Modifies the torch-to-linalg lowering of AtenReplicationPad2d op to
enable support for input rank != 4

Unfortunately, even with all of these changes, the e2e tests for the
ReplicationPad2d still fail the onnx config, since the torch export
procedure for rearranging the pad order is complicated enough that the
padding ints end up not being able to fold back to constants.
2024-07-11 20:01:45 -05:00
..
IR [ONNX] Fix bug in ONNXToTorch PadOp's pads tensor rearrangement (#3485) 2024-07-03 15:02:49 -05:00
Transforms Adds misc fixes for some padding related issues (#3528) 2024-07-11 20:01:45 -05:00
Utils [TorchToLinalg] Fix Quantized Convolution Accumulator Type (#3459) 2024-06-20 13:54:20 -07:00
CMakeLists.txt E2e support for aten.softmax.int and aten.embedding 2021-10-18 17:57:45 -04:00