torch-mlir/lib
zjgarvey 8995c90879
[TorchToLinalg] add support for quantized group conv (#3341)
This addresses 7 of the model failures I'm seeing in the test suite. See
[Shark-Turbine issue
#566](https://github.com/nod-ai/SHARK-Turbine/issues/566).

Need the op ```linalg.conv_2d_ngchw_gfchw_q``` to be added upstream
before merging this. See [llvm-project PR #92136
](https://github.com/llvm/llvm-project/pull/92136).

A small additional expansion to operand quantization is included in this
patch to address a model failure that occurs when unblocking the
quantized group convolutions in one of these onnx models.
2024-06-03 21:57:44 +05:30
..
CAPI [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
Conversion [TorchToLinalg] add support for quantized group conv (#3341) 2024-06-03 21:57:44 +05:30
Dialect [TorchToLinalg] add support for quantized group conv (#3341) 2024-06-03 21:57:44 +05:30
RefBackend [NFC] Change to *cast instead of .*cast variants (#3405) 2024-05-30 23:45:13 -07:00
CMakeLists.txt [Stablehlo] add stablehlo-aggressive-simplification in e2e test (#3109) 2024-04-07 10:48:11 +08:00
InitAll.cpp [Stablehlo] add stablehlo-aggressive-simplification in e2e test (#3109) 2024-04-07 10:48:11 +08:00