mirror of https://github.com/llvm/torch-mlir
0314188dbe
This patch adds basic support for lowering graphs with per-channel quantization. Per-channel quantized ops have to be excluded from `FuseQuantizedOps` for now but can be used in QDQ quantized form. Using this patch, we're able to import and execute (on the linalg backend) graphs with per-channel quantization applied using the "new" PyTorch 2.0 Export Quantization. |
||
---|---|---|
.. | ||
TMTensor | ||
Torch | ||
TorchConversion |