2023-11-22 13:02:55 +08:00
|
|
|
//===------------------------------------------------------------*- C++ -*-===//
|
|
|
|
//
|
|
|
|
// This file is licensed under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
|
|
// Also available under a BSD-style license. See LICENSE.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2024-02-01 03:40:53 +08:00
|
|
|
#include "mlir/IR/DialectResourceBlobManager.h"
|
2023-11-22 13:02:55 +08:00
|
|
|
#include "torch-mlir/Conversion/TorchOnnxToTorch/Patterns.h"
|
2024-04-23 00:58:07 +08:00
|
|
|
#include "torch-mlir/Conversion/TorchOnnxToTorch/Utils.h"
|
2023-11-27 21:44:16 +08:00
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchOps.h"
|
2023-12-19 08:17:11 +08:00
|
|
|
#include "torch-mlir/Dialect/Torch/Utils/Utils.h"
|
2024-01-23 23:36:25 +08:00
|
|
|
#include "llvm/Support/FormatVariadic.h"
|
2023-11-22 13:02:55 +08:00
|
|
|
|
|
|
|
using namespace mlir;
|
|
|
|
using namespace mlir::torch;
|
|
|
|
using namespace mlir::torch::onnx_c;
|
|
|
|
|
2024-05-04 00:04:57 +08:00
|
|
|
namespace {
|
|
|
|
LogicalResult windowFunctionImpl(OpBinder binder,
|
|
|
|
ConversionPatternRewriter &rewriter,
|
|
|
|
Value size, Value a0, Value a1, Value a2,
|
|
|
|
Torch::ValueTensorType resultType,
|
|
|
|
int64_t output_datatype, int64_t periodic) {
|
|
|
|
|
|
|
|
Location loc = binder.getLoc();
|
|
|
|
ImplicitLocOpBuilder b(loc, rewriter);
|
|
|
|
|
|
|
|
double isPeriodicFp = static_cast<double>(periodic);
|
|
|
|
|
|
|
|
Value zero = b.create<Torch::ConstantFloatOp>(rewriter.getF64FloatAttr(0.0));
|
|
|
|
Value one = b.create<Torch::ConstantFloatOp>(rewriter.getF64FloatAttr(1.0));
|
|
|
|
Value two = b.create<Torch::ConstantFloatOp>(rewriter.getF64FloatAttr(2.0));
|
|
|
|
|
|
|
|
constexpr double pi = llvm::numbers::pi;
|
|
|
|
Value tau = b.create<Torch::ConstantFloatOp>(
|
|
|
|
rewriter.getFloatAttr(rewriter.getF64Type(), 2.0 * pi));
|
|
|
|
|
|
|
|
Value noneVal = b.create<Torch::ConstantNoneOp>();
|
|
|
|
Value cstFalse = b.create<Torch::ConstantBoolOp>(false);
|
|
|
|
Value float32Type = b.create<Torch::ConstantIntOp>(
|
|
|
|
rewriter.getI64IntegerAttr(/*float32Type*/ 6));
|
|
|
|
|
|
|
|
// Create an f32 ValueTensorType with thse same size as size, the
|
|
|
|
// operand
|
|
|
|
auto shapeOfOperand =
|
2024-05-24 00:01:47 +08:00
|
|
|
dyn_cast<Torch::ValueTensorType>(size.getType()).getOptionalSizes();
|
2024-05-04 00:04:57 +08:00
|
|
|
auto f32ResultType = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
shapeOfOperand, rewriter.getF32Type());
|
|
|
|
Value periodicSizeFloat = b.create<Torch::AtenToDtypeOp>(
|
|
|
|
f32ResultType, size, float32Type, cstFalse, cstFalse, noneVal);
|
|
|
|
Value symmetricSizeFloat = b.create<Torch::AtenSubScalarOp>(
|
|
|
|
periodicSizeFloat.getType(), periodicSizeFloat, one, one);
|
|
|
|
|
|
|
|
Value isPeriodic =
|
|
|
|
b.create<Torch::ConstantFloatOp>(rewriter.getF64FloatAttr(isPeriodicFp));
|
|
|
|
Value isSymmetricFloat = b.create<Torch::ConstantFloatOp>(
|
|
|
|
rewriter.getF64FloatAttr(1.0 - isPeriodicFp));
|
|
|
|
|
|
|
|
Value periodicComponent = b.create<Torch::AtenMulScalarOp>(
|
|
|
|
periodicSizeFloat.getType(), periodicSizeFloat, isPeriodic);
|
|
|
|
Value symmetricComponent = b.create<Torch::AtenMulScalarOp>(
|
|
|
|
symmetricSizeFloat.getType(), symmetricSizeFloat, isSymmetricFloat);
|
|
|
|
Value sizeFloat = b.create<Torch::AtenAddTensorOp>(
|
|
|
|
symmetricComponent.getType(), symmetricComponent, periodicComponent, one);
|
|
|
|
|
|
|
|
// Here, size can be used in the place of periodicSizeFloat, as the
|
|
|
|
// latter is just a float representation of the former.
|
|
|
|
Value scalarLimit = getItemOp<Torch::IntType>(binder, rewriter, size);
|
|
|
|
|
|
|
|
Value rangeArr = b.create<Torch::AtenArangeStartStepOp>(
|
|
|
|
resultType, zero, scalarLimit, one, noneVal, noneVal, noneVal, noneVal);
|
|
|
|
|
|
|
|
Value rangeTimesTau =
|
|
|
|
b.create<Torch::AtenMulScalarOp>(resultType, rangeArr, tau);
|
|
|
|
Value rangeAngular =
|
|
|
|
b.create<Torch::AtenDivTensorOp>(resultType, rangeTimesTau, sizeFloat);
|
|
|
|
Value twoRangeAngular =
|
|
|
|
b.create<Torch::AtenMulScalarOp>(resultType, rangeAngular, two);
|
|
|
|
|
|
|
|
Value cosRangeAngular = b.create<Torch::AtenCosOp>(resultType, rangeAngular);
|
|
|
|
Value cosTwoRangeAngular =
|
|
|
|
b.create<Torch::AtenCosOp>(resultType, twoRangeAngular);
|
|
|
|
|
|
|
|
Value a1Component =
|
|
|
|
b.create<Torch::AtenMulScalarOp>(resultType, cosRangeAngular, a1);
|
|
|
|
Value a2Component =
|
|
|
|
b.create<Torch::AtenMulScalarOp>(resultType, cosTwoRangeAngular, a2);
|
|
|
|
|
|
|
|
// AtenSubScalarOp actually requires a tensor operand as the LHS, that
|
|
|
|
// is, operand #1. Therefore, to avoid errors, the onnx implementation
|
|
|
|
// has been modified. a1 has been changed to negative half, and the
|
|
|
|
// AtenSubScalarOp has been replaced with AtenAddScalarOp, as the add
|
|
|
|
// operation is commutative.
|
|
|
|
Value subA1Component =
|
|
|
|
b.create<Torch::AtenAddScalarOp>(resultType, a1Component, a0, one);
|
|
|
|
Value result = b.create<Torch::AtenAddTensorOp>(resultType, subA1Component,
|
|
|
|
a2Component, one);
|
|
|
|
|
|
|
|
std::optional<int64_t> dtypeIntTorch =
|
|
|
|
onnxDtypeIntToTorchDtypeInt(output_datatype);
|
|
|
|
if (!dtypeIntTorch.has_value()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unimplemented support for the given dtype conversion");
|
|
|
|
}
|
|
|
|
Value outputDtype = b.create<Torch::ConstantIntOp>(
|
|
|
|
rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64),
|
|
|
|
dtypeIntTorch.value()));
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenToDtypeOp>(
|
|
|
|
binder.op, resultType, result, outputDtype,
|
|
|
|
/*non_blocking=*/cstFalse, /*copy=*/cstFalse,
|
|
|
|
/*memory_format=*/noneVal);
|
|
|
|
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
2023-11-22 13:02:55 +08:00
|
|
|
// Simple rewrites for the default domain.
|
|
|
|
// See: https://onnx.ai/onnx/operators/
|
|
|
|
// For operators that are effectively version invariant, we register with
|
|
|
|
// sinceVersion==1. We interpret this to include the following spec
|
|
|
|
// diffs that are irrelevant to this level of lowering:
|
|
|
|
// * Supported element types.
|
|
|
|
// * Limited broadcasting to full broadcasting support.
|
|
|
|
//
|
|
|
|
// There are a lot of spec revisions that basically generalized elementwise
|
|
|
|
// to be more normal and a direct translation vs a special case. This
|
|
|
|
// results in a lot of ONNX test cases that all reduce to the exact same
|
|
|
|
// thing here, so we simplify.
|
|
|
|
void mlir::torch::onnx_c::populateDefaultDomainAtoF(
|
|
|
|
OnnxCustomOpConversionPattern &patterns) {
|
|
|
|
patterns.onOp("Abs", 1,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAbsOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
// Add became forward compatible with Torch in version 7.
|
|
|
|
patterns.onOp("Add", 7,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
Value const1 = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), 1));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAddTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs, const1);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
// TODO: AffineGrid
|
|
|
|
patterns.onOp("And", 1,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenLogicalAndOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp(
|
|
|
|
"ArgMax", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
bool keepDims;
|
|
|
|
int64_t axis;
|
|
|
|
bool selectLastIndex;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType) ||
|
|
|
|
binder.s64BoolAttr(keepDims, "keepdims", true) ||
|
|
|
|
binder.s64IntegerAttr(axis, "axis", 0) ||
|
|
|
|
binder.s64BoolAttr(selectLastIndex, "select_last_index", false))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
// ONNX allows negative axis.
|
2024-04-24 01:16:08 +08:00
|
|
|
auto operandSizes =
|
|
|
|
cast<Torch::ValueTensorType>(operand.getType()).getSizes();
|
2023-11-22 13:02:55 +08:00
|
|
|
if (axis < 0)
|
2024-04-24 01:16:08 +08:00
|
|
|
axis += operandSizes.size();
|
2023-11-22 13:02:55 +08:00
|
|
|
|
|
|
|
Value constAxis = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), axis));
|
|
|
|
Value constKeepDims = rewriter.create<Torch::ConstantBoolOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::BoolType>(),
|
|
|
|
rewriter.getBoolAttr(keepDims));
|
2024-04-24 01:16:08 +08:00
|
|
|
|
|
|
|
if (selectLastIndex) {
|
|
|
|
Value dims = createConstantIntList(binder, rewriter, {axis});
|
|
|
|
auto operandTy = dyn_cast<Torch::ValueTensorType>(operand.getType());
|
|
|
|
operand = rewriter.create<Torch::AtenFlipOp>(
|
|
|
|
binder.getLoc(), operandTy, operand, dims);
|
|
|
|
Value argmax = rewriter.create<Torch::AtenArgmaxOp>(
|
|
|
|
binder.getLoc(), resultType, operand, constAxis, constKeepDims);
|
|
|
|
Value offset = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getI64IntegerAttr(operandSizes[axis] - 1));
|
|
|
|
Value alpha = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1));
|
|
|
|
Value sub = rewriter.create<Torch::AtenSubScalarOp>(
|
|
|
|
binder.getLoc(), resultType, argmax, offset, alpha);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAbsOp>(binder.op, resultType,
|
|
|
|
sub);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2023-11-22 13:02:55 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenArgmaxOp>(
|
|
|
|
binder.op, resultType, operand, constAxis, constKeepDims);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp(
|
|
|
|
"ArgMin", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
bool keepDims;
|
|
|
|
int64_t axis;
|
|
|
|
bool selectLastIndex;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType) ||
|
|
|
|
binder.s64BoolAttr(keepDims, "keepdims", true) ||
|
|
|
|
binder.s64IntegerAttr(axis, "axis", 0) ||
|
|
|
|
binder.s64BoolAttr(selectLastIndex, "select_last_index", false))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
// ONNX allows negative axis.
|
2024-04-24 01:43:38 +08:00
|
|
|
auto operandSizes =
|
|
|
|
cast<Torch::ValueTensorType>(operand.getType()).getSizes();
|
2023-11-22 13:02:55 +08:00
|
|
|
if (axis < 0)
|
2024-04-24 01:43:38 +08:00
|
|
|
axis += operandSizes.size();
|
2023-11-22 13:02:55 +08:00
|
|
|
|
|
|
|
Value constAxis = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), axis));
|
|
|
|
Value constKeepDims = rewriter.create<Torch::ConstantBoolOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::BoolType>(),
|
|
|
|
rewriter.getBoolAttr(keepDims));
|
2024-04-24 01:43:38 +08:00
|
|
|
|
|
|
|
if (selectLastIndex) {
|
|
|
|
Value dims = createConstantIntList(binder, rewriter, {axis});
|
|
|
|
auto operandTy = dyn_cast<Torch::ValueTensorType>(operand.getType());
|
|
|
|
operand = rewriter.create<Torch::AtenFlipOp>(
|
|
|
|
binder.getLoc(), operandTy, operand, dims);
|
|
|
|
Value argmin = rewriter.create<Torch::AtenArgminOp>(
|
|
|
|
binder.getLoc(), resultType, operand, constAxis, constKeepDims);
|
|
|
|
Value offset = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getI64IntegerAttr(operandSizes[axis] - 1));
|
|
|
|
Value alpha = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1));
|
|
|
|
Value sub = rewriter.create<Torch::AtenSubScalarOp>(
|
|
|
|
binder.getLoc(), resultType, argmin, offset, alpha);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAbsOp>(binder.op, resultType,
|
|
|
|
sub);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2023-11-22 13:02:55 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenArgminOp>(
|
|
|
|
binder.op, resultType, operand, constAxis, constKeepDims);
|
|
|
|
return success();
|
|
|
|
});
|
2024-02-14 14:28:09 +08:00
|
|
|
patterns.onOp("Asin", 7,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAsinOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-05-04 00:06:44 +08:00
|
|
|
patterns.onOp("Asinh", 9,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAsinhOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2023-11-27 21:44:16 +08:00
|
|
|
patterns.onOp("Atan", 7,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAtanOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-05-04 00:06:44 +08:00
|
|
|
patterns.onOp("Atanh", 9,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAtanhOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-06 18:26:13 +08:00
|
|
|
patterns.onOp("Acos", 7,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAcosOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-05-04 00:06:44 +08:00
|
|
|
patterns.onOp("Acosh", 9,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAcoshOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-22 00:04:02 +08:00
|
|
|
patterns.onOp("BatchNormalization", 15,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input, weight, bias, runningMean, runningVar;
|
|
|
|
bool training;
|
|
|
|
float momentum, eps;
|
|
|
|
if (binder.s64BoolAttr(training, "training_mode", 0))
|
|
|
|
return failure();
|
|
|
|
if (training) {
|
|
|
|
// TODO: Add support for training = true
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: training = true");
|
|
|
|
}
|
|
|
|
|
|
|
|
if (binder.tensorOperandAtIndex(input, 0) ||
|
|
|
|
binder.tensorOperandAtIndex(weight, 1) ||
|
|
|
|
binder.tensorOperandAtIndex(bias, 2) ||
|
|
|
|
binder.tensorOperandAtIndex(runningMean, 3) ||
|
|
|
|
binder.tensorOperandAtIndex(runningVar, 4) ||
|
2024-01-09 06:26:38 +08:00
|
|
|
binder.f32FloatAttr(momentum, "momentum", 0.9f) ||
|
|
|
|
binder.f32FloatAttr(eps, "epsilon", 1e-05f) ||
|
2023-12-22 00:04:02 +08:00
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
Value cstFalse = rewriter.create<Torch::ConstantBoolOp>(
|
|
|
|
binder.getLoc(), false);
|
|
|
|
Value cstMomentum = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(), rewriter.getF64FloatAttr(momentum));
|
|
|
|
Value cstEps = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(), rewriter.getF64FloatAttr(eps));
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBatchNormOp>(
|
|
|
|
binder.op, resultType, input, weight, bias, runningMean,
|
|
|
|
runningVar, /*training=*/cstFalse, cstMomentum, cstEps,
|
|
|
|
/*cudnn_enabled=*/cstFalse);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-19 08:17:11 +08:00
|
|
|
patterns.onOp(
|
2024-04-02 00:44:14 +08:00
|
|
|
"AveragePool", 11,
|
2023-12-19 08:17:11 +08:00
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
std::string autoPad;
|
|
|
|
SmallVector<int64_t> dilation;
|
|
|
|
if (binder.customOpNameStringAttr(autoPad, "auto_pad", "NOTSET"))
|
|
|
|
return failure();
|
|
|
|
if (autoPad != "NOTSET") {
|
|
|
|
// TODO: Add support for `auto_pad` != "NOTSET"
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: auto_pad != NOTSET");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(dilation, "dilations", {})) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (dilation.size() > 0) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "dilation is not supported by torch.aten.avgpool op");
|
|
|
|
}
|
|
|
|
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
bool ceilMode, countIncludePad;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.s64BoolAttr(ceilMode, "ceil_mode", false) ||
|
|
|
|
binder.s64BoolAttr(countIncludePad, "count_include_pad", false) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
// Determine the rank of input tensor.
|
|
|
|
std::optional<unsigned> maybeRank = Torch::getTensorRank(operand);
|
|
|
|
if (!maybeRank)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"Unimplemented: unranked tensor");
|
|
|
|
unsigned rank = *maybeRank;
|
|
|
|
|
|
|
|
SmallVector<int64_t> kernel, padding, strides;
|
|
|
|
if (binder.s64IntegerArrayAttr(kernel, "kernel_shape", {})) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (kernel.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "kernel list size does not match the number of axes");
|
|
|
|
}
|
onnx: fix checks in TorchOnnxToTorch pass to match the ONNX spec (#2848)
This PR contains three commits to update the validation checks in the
ONNX -> Torch conversion pass for the AveragePool, Pad, and Slice operators:
> onnx: fix preconditions for lowering AveragePool ops
>
> The `pads` attribute of the AveragePool operator specifies the value to
> pad at both the beginning as well as the end of the axis (see
> https://onnx.ai/onnx/operators/onnx__AveragePool.html#attributes), so
> the size of this attribute should be twice the rank of the input tensor.
> However, our TorchOnnxToTorch bails out early since it incorrectly
> compares the pads attribute with the rank (not twice the rank) of the
> input tensor.
>
> This patch fixes the code to match the spec and adds a lit test.
> onnx: allow optional constant value for Pad operator
>
> The `constant_value` input of the onnx.Pad operator is optional (see
> https://onnx.ai/onnx/operators/onnx__Pad.html#inputs), but the
existing
> logic for lowering the operator into the Torch dialect assumes that it
> is mandatory.
>
> This patch makes the attribute optional and constructs a default value
> (a list of zeros the size of the input tensor) if the attribute was not
> specified.
> onnx: fix checks for axes and steps inputs of Slice operator
>
> The ONNX Spec for the Slice operator allows the `starts` and `ends`
> inputs to have fewer indices that the dimensions of the `data` tensor
> (see https://onnx.ai/onnx/operators/onnx__Slice.html), but our code
> expects these inputs to be as many as the `data` tensor's dimensions.
>
> More precisely, the spec requires that the `starts` and `ends` inputs
> are only as long as the `axes` input, but since the `axes` input is
> optional, the default type for the `axes` input has to match the type
> for the `starts` and `ends` inputs. Moreover, the number of indices in
> the `steps` input also has to match those in the `axes` inputs (instad
> of matching the dimensions of the `data` input).
>
> This patch fixes the checks in the TorchOnnxToTorch conversion so that
> they match the ONNX spec.
2024-02-08 13:19:27 +08:00
|
|
|
SmallVector<int64_t> defaultPadding(2 * (rank - 2), 0);
|
|
|
|
if (binder.s64IntegerArrayAttr(padding, "pads", defaultPadding)) {
|
2023-12-19 08:17:11 +08:00
|
|
|
return failure();
|
|
|
|
}
|
onnx: fix checks in TorchOnnxToTorch pass to match the ONNX spec (#2848)
This PR contains three commits to update the validation checks in the
ONNX -> Torch conversion pass for the AveragePool, Pad, and Slice operators:
> onnx: fix preconditions for lowering AveragePool ops
>
> The `pads` attribute of the AveragePool operator specifies the value to
> pad at both the beginning as well as the end of the axis (see
> https://onnx.ai/onnx/operators/onnx__AveragePool.html#attributes), so
> the size of this attribute should be twice the rank of the input tensor.
> However, our TorchOnnxToTorch bails out early since it incorrectly
> compares the pads attribute with the rank (not twice the rank) of the
> input tensor.
>
> This patch fixes the code to match the spec and adds a lit test.
> onnx: allow optional constant value for Pad operator
>
> The `constant_value` input of the onnx.Pad operator is optional (see
> https://onnx.ai/onnx/operators/onnx__Pad.html#inputs), but the
existing
> logic for lowering the operator into the Torch dialect assumes that it
> is mandatory.
>
> This patch makes the attribute optional and constructs a default value
> (a list of zeros the size of the input tensor) if the attribute was not
> specified.
> onnx: fix checks for axes and steps inputs of Slice operator
>
> The ONNX Spec for the Slice operator allows the `starts` and `ends`
> inputs to have fewer indices that the dimensions of the `data` tensor
> (see https://onnx.ai/onnx/operators/onnx__Slice.html), but our code
> expects these inputs to be as many as the `data` tensor's dimensions.
>
> More precisely, the spec requires that the `starts` and `ends` inputs
> are only as long as the `axes` input, but since the `axes` input is
> optional, the default type for the `axes` input has to match the type
> for the `starts` and `ends` inputs. Moreover, the number of indices in
> the `steps` input also has to match those in the `axes` inputs (instad
> of matching the dimensions of the `data` input).
>
> This patch fixes the checks in the TorchOnnxToTorch conversion so that
> they match the ONNX spec.
2024-02-08 13:19:27 +08:00
|
|
|
if (padding.size() != 2 * (rank - 2)) {
|
2023-12-19 08:17:11 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
onnx: fix checks in TorchOnnxToTorch pass to match the ONNX spec (#2848)
This PR contains three commits to update the validation checks in the
ONNX -> Torch conversion pass for the AveragePool, Pad, and Slice operators:
> onnx: fix preconditions for lowering AveragePool ops
>
> The `pads` attribute of the AveragePool operator specifies the value to
> pad at both the beginning as well as the end of the axis (see
> https://onnx.ai/onnx/operators/onnx__AveragePool.html#attributes), so
> the size of this attribute should be twice the rank of the input tensor.
> However, our TorchOnnxToTorch bails out early since it incorrectly
> compares the pads attribute with the rank (not twice the rank) of the
> input tensor.
>
> This patch fixes the code to match the spec and adds a lit test.
> onnx: allow optional constant value for Pad operator
>
> The `constant_value` input of the onnx.Pad operator is optional (see
> https://onnx.ai/onnx/operators/onnx__Pad.html#inputs), but the
existing
> logic for lowering the operator into the Torch dialect assumes that it
> is mandatory.
>
> This patch makes the attribute optional and constructs a default value
> (a list of zeros the size of the input tensor) if the attribute was not
> specified.
> onnx: fix checks for axes and steps inputs of Slice operator
>
> The ONNX Spec for the Slice operator allows the `starts` and `ends`
> inputs to have fewer indices that the dimensions of the `data` tensor
> (see https://onnx.ai/onnx/operators/onnx__Slice.html), but our code
> expects these inputs to be as many as the `data` tensor's dimensions.
>
> More precisely, the spec requires that the `starts` and `ends` inputs
> are only as long as the `axes` input, but since the `axes` input is
> optional, the default type for the `axes` input has to match the type
> for the `starts` and `ends` inputs. Moreover, the number of indices in
> the `steps` input also has to match those in the `axes` inputs (instad
> of matching the dimensions of the `data` input).
>
> This patch fixes the checks in the TorchOnnxToTorch conversion so that
> they match the ONNX spec.
2024-02-08 13:19:27 +08:00
|
|
|
binder.op,
|
|
|
|
"padding list size does not match twice the number of axes");
|
2023-12-19 08:17:11 +08:00
|
|
|
}
|
2024-04-02 00:44:14 +08:00
|
|
|
if (binder.s64IntegerArrayAttr(
|
|
|
|
strides, "strides", llvm::SmallVector<int64_t>(rank - 2, 1))) {
|
2023-12-19 08:17:11 +08:00
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (strides.size() != 1 && strides.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "strides list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
|
|
|
|
SmallVector<Value> cstKernel, cstPadding, cstStrides;
|
|
|
|
for (int64_t i : kernel) {
|
|
|
|
cstKernel.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
for (int64_t i : padding) {
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
for (int64_t i : strides) {
|
|
|
|
cstStrides.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
Value kernelSizeList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstKernel);
|
|
|
|
Value paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstPadding);
|
|
|
|
Value stridesList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstStrides);
|
|
|
|
Value cstCeilMode =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), ceilMode);
|
|
|
|
Value cstCountIncludePad = rewriter.create<Torch::ConstantBoolOp>(
|
|
|
|
binder.getLoc(), countIncludePad);
|
|
|
|
Value cstNone = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
|
|
|
|
if (rank == 3) {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAvgPool1dOp>(
|
|
|
|
binder.op, resultType, operand, kernelSizeList, stridesList,
|
|
|
|
paddingList, cstCeilMode, cstCountIncludePad);
|
|
|
|
return success();
|
|
|
|
} else if (rank == 4) {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAvgPool2dOp>(
|
|
|
|
binder.op, resultType, operand, kernelSizeList, stridesList,
|
|
|
|
paddingList, cstCeilMode, cstCountIncludePad,
|
|
|
|
/*divisor_override=*/cstNone);
|
|
|
|
return success();
|
|
|
|
} else if (rank == 5) {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAvgPool3dOp>(
|
|
|
|
binder.op, resultType, operand, kernelSizeList, stridesList,
|
|
|
|
paddingList, cstCeilMode, cstCountIncludePad,
|
|
|
|
/*divisor_override=*/cstNone);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
return failure();
|
|
|
|
});
|
2024-01-04 23:12:51 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Bernoulli", 15,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input;
|
2024-04-23 00:58:07 +08:00
|
|
|
int64_t dtypeIntOnnx;
|
2024-01-04 23:12:51 +08:00
|
|
|
if (binder.tensorOperand(input) ||
|
|
|
|
binder.s64IntegerAttr(dtypeIntOnnx, "dtype", -1) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
SmallString<64> name("torch.onnx.");
|
|
|
|
name.append("seed");
|
|
|
|
auto attr = binder.op->getAttr(name);
|
|
|
|
if (attr) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unimplemented: support not present for seed attribute");
|
|
|
|
}
|
|
|
|
|
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
Value bernoulli = rewriter.create<Torch::AtenBernoulliOp>(
|
|
|
|
binder.getLoc(), input.getType(), input, /*generator=*/none);
|
|
|
|
|
|
|
|
if (dtypeIntOnnx == -1) {
|
|
|
|
// True, if dtype attribute value is not present.
|
|
|
|
rewriter.replaceOp(binder.op, bernoulli);
|
|
|
|
return success();
|
|
|
|
}
|
2024-04-23 00:58:07 +08:00
|
|
|
std::optional<int64_t> dtypeIntTorch =
|
|
|
|
onnxDtypeIntToTorchDtypeInt(dtypeIntOnnx);
|
|
|
|
if (!dtypeIntTorch.has_value()) {
|
2024-01-04 23:12:51 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unimplemented support for the given dtype conversion");
|
|
|
|
}
|
|
|
|
Value constDtype = rewriter.create<Torch::ConstantIntOp>(
|
2024-04-23 00:58:07 +08:00
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(dtypeIntTorch.value()));
|
2024-01-04 23:12:51 +08:00
|
|
|
Value cstFalse =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), false);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenToDtypeOp>(
|
|
|
|
binder.op, resultType, bernoulli, constDtype,
|
|
|
|
/*non_blocking=*/cstFalse, /*copy=*/cstFalse,
|
|
|
|
/*memory_format=*/none);
|
|
|
|
return success();
|
|
|
|
});
|
2023-11-27 21:44:16 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"BitShift", 11, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType) ||
|
|
|
|
binder.customOpNameStringAttr(direction, "direction", ""))
|
|
|
|
return failure();
|
|
|
|
if (direction == "LEFT") {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseLeftShiftTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
} else {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseRightShiftTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
}
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-23 23:36:25 +08:00
|
|
|
patterns.onOp("BitwiseAnd", 18,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseAndTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp("BitwiseOr", 18,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseOrTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
2023-11-27 21:44:16 +08:00
|
|
|
patterns.onOp("BitwiseNot", 18,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseNotOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-23 23:36:25 +08:00
|
|
|
patterns.onOp("BitwiseXor", 18,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBitwiseXorTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-05 13:55:51 +08:00
|
|
|
patterns.onOp(
|
2023-12-28 02:08:09 +08:00
|
|
|
"Cast", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
2023-12-05 13:55:51 +08:00
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
2024-04-23 00:58:07 +08:00
|
|
|
int64_t dtypeIntOnnx;
|
2023-12-05 13:55:51 +08:00
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.s64IntegerAttr(dtypeIntOnnx, "to") ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
2024-04-23 00:58:07 +08:00
|
|
|
std::optional<int64_t> dtypeIntTorch =
|
|
|
|
onnxDtypeIntToTorchDtypeInt(dtypeIntOnnx);
|
|
|
|
if (!dtypeIntTorch.has_value()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unimplemented support for the given dtype conversion");
|
2023-12-05 13:55:51 +08:00
|
|
|
}
|
|
|
|
Value constDtype = rewriter.create<Torch::ConstantIntOp>(
|
2024-04-23 00:58:07 +08:00
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(dtypeIntTorch.value()));
|
2023-12-05 13:55:51 +08:00
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
Value cstFalse =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), false);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenToDtypeOp>(
|
|
|
|
binder.op, resultType, operand, constDtype,
|
|
|
|
/*non_blocking=*/cstFalse, /*copy=*/cstFalse,
|
|
|
|
/*memory_format=*/none);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-04 23:12:51 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"CastLike", 15, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input, target;
|
|
|
|
if (binder.tensorOperands(input, target) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
// TODO: Add support to handle the `saturate` attribute.
|
|
|
|
// Ignoring it right now, since it's only using during the float8
|
|
|
|
// conversions which are not supported in Torch-MLIR right now.
|
|
|
|
|
|
|
|
Torch::ValueTensorType targetTy =
|
2024-04-28 05:00:56 +08:00
|
|
|
cast<Torch::ValueTensorType>(target.getType());
|
2024-01-04 23:12:51 +08:00
|
|
|
if (!targetTy.hasDtype()) {
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"target tensor must have a dtype");
|
|
|
|
}
|
|
|
|
Type targetDtype = targetTy.getDtype();
|
|
|
|
Value constDtype = Torch::getDtypeIntValueForType(
|
|
|
|
rewriter, binder.getLoc(), targetDtype);
|
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
Value cstFalse =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), false);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenToDtypeOp>(
|
|
|
|
binder.op, resultType, input, constDtype,
|
|
|
|
/*non_blocking=*/cstFalse, /*copy=*/cstFalse,
|
|
|
|
/*memory_format=*/none);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-05 13:55:51 +08:00
|
|
|
patterns.onOp("Ceil", 13,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenCeilOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-03-13 23:04:10 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Celu", 12, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
float alpha;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType) ||
|
|
|
|
binder.f32FloatAttr(alpha, "alpha", 1.0f))
|
|
|
|
return failure();
|
|
|
|
// exp(x/alpha)
|
|
|
|
Value constAlpha = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::FloatType>(),
|
|
|
|
rewriter.getF64FloatAttr(alpha));
|
|
|
|
Value xDivAlpha = rewriter.create<Torch::AtenDivScalarOp>(
|
|
|
|
binder.getLoc(), resultType, operand, constAlpha);
|
|
|
|
Value expXDivAlpha = rewriter.create<Torch::AtenExpOp>(
|
|
|
|
binder.getLoc(), resultType, xDivAlpha);
|
|
|
|
// alpha * (exp(x/alpha) - 1)
|
|
|
|
Value constantOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1));
|
|
|
|
Value subOne = rewriter.create<Torch::AtenSubScalarOp>(
|
|
|
|
binder.getLoc(), resultType, expXDivAlpha, constantOne,
|
|
|
|
constantOne);
|
|
|
|
Value mulAlpha = rewriter.create<Torch::AtenMulScalarOp>(
|
|
|
|
binder.getLoc(), resultType, subOne, constAlpha);
|
|
|
|
Value constantZero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
Value zeroTensor = createRank0Tensor(rewriter, binder.getLoc(),
|
|
|
|
resultType, constantZero);
|
|
|
|
// min(0, alpha * (exp(x/alpha) - 1))
|
|
|
|
Value minExpression = rewriter.create<Torch::AtenMinimumOp>(
|
|
|
|
binder.getLoc(), resultType, zeroTensor, mulAlpha);
|
|
|
|
|
|
|
|
// max(0, x)
|
|
|
|
Value maxExpression = rewriter.create<Torch::AtenMaximumOp>(
|
|
|
|
binder.getLoc(), resultType, zeroTensor, operand);
|
|
|
|
// max(0,x) + min(0, alpha * (exp(x/alpha) - 1))
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenAddTensorOp>(
|
|
|
|
binder.op, resultType, maxExpression, minExpression, constantOne);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-05 13:55:51 +08:00
|
|
|
patterns.onOp(
|
2024-03-08 01:25:14 +08:00
|
|
|
"Clip", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
// https://onnx.ai/onnx/operators/onnx__Clip.html
|
|
|
|
|
|
|
|
// Inputs and outputs must be tensors.
|
|
|
|
Value source;
|
2023-12-05 13:55:51 +08:00
|
|
|
Torch::ValueTensorType resultType;
|
2024-03-08 01:25:14 +08:00
|
|
|
if (binder.tensorOperandAtIndex(source, 0) ||
|
|
|
|
binder.tensorResultType(resultType)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Min and max can be args (version 11+) or attributes (version 6-).
|
|
|
|
// They default to numeric_limits::lowest() and numeric_limits::max().
|
|
|
|
Value min;
|
|
|
|
Value max;
|
|
|
|
if (binder.op->getNumOperands() >= 2)
|
|
|
|
min = binder.op->getOperand(1);
|
|
|
|
if (binder.op->getNumOperands() == 3)
|
|
|
|
max = binder.op->getOperand(2);
|
|
|
|
|
|
|
|
// Note: attribute versions of the op only support float types.
|
|
|
|
auto resultDtype = resultType.getDtype();
|
|
|
|
if (!min && binder.op->hasAttr("torch.onnx.min")) {
|
|
|
|
float minValue;
|
|
|
|
if (binder.f32FloatAttr(minValue, "min",
|
|
|
|
std::numeric_limits<float>::lowest()))
|
2023-12-05 13:55:51 +08:00
|
|
|
return failure();
|
2024-03-08 01:25:14 +08:00
|
|
|
auto minSplatAttr = SplatElementsAttr::get(
|
2024-06-08 09:36:32 +08:00
|
|
|
resultType.toBuiltinTensor(),
|
2024-03-08 01:25:14 +08:00
|
|
|
rewriter.getFloatAttr(resultDtype, minValue));
|
|
|
|
min = rewriter.create<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.getLoc(), resultType, minSplatAttr);
|
|
|
|
}
|
|
|
|
if (!max && binder.op->hasAttr("torch.onnx.max")) {
|
|
|
|
float maxValue;
|
|
|
|
if (binder.f32FloatAttr(maxValue, "max",
|
|
|
|
std::numeric_limits<float>::max()))
|
2023-12-05 13:55:51 +08:00
|
|
|
return failure();
|
2024-03-08 01:25:14 +08:00
|
|
|
auto maxSplatAttr = SplatElementsAttr::get(
|
2024-06-08 09:36:32 +08:00
|
|
|
resultType.toBuiltinTensor(),
|
2024-03-08 01:25:14 +08:00
|
|
|
rewriter.getFloatAttr(resultDtype, maxValue));
|
|
|
|
max = rewriter.create<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.getLoc(), resultType, maxSplatAttr);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!min && !max) {
|
|
|
|
// Cliping with no limits is a no-op.
|
|
|
|
rewriter.replaceOp(binder.op, source);
|
2023-12-05 13:55:51 +08:00
|
|
|
return success();
|
2024-03-08 01:25:14 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!max) {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenClampMinTensorOp>(
|
|
|
|
binder.op, resultType, source, min);
|
2023-12-05 13:55:51 +08:00
|
|
|
return success();
|
|
|
|
}
|
2024-03-08 01:25:14 +08:00
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenClampTensorOp>(
|
|
|
|
binder.op, resultType, source, min, max);
|
|
|
|
return success();
|
2023-12-05 13:55:51 +08:00
|
|
|
});
|
2024-03-21 01:12:08 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Compress", 9, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand, conditionTensor;
|
|
|
|
int64_t axis;
|
|
|
|
if (binder.tensorOperands(operand, conditionTensor) ||
|
|
|
|
binder.s64IntegerAttr(axis, "axis", INT64_MAX) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
auto shapeSizes =
|
|
|
|
dyn_cast<Torch::ValueTensorType>(operand.getType()).getSizes();
|
2024-04-13 06:18:22 +08:00
|
|
|
auto resultSizes = resultType.getSizes();
|
|
|
|
|
|
|
|
// flatten input tensor if using default axis
|
2024-03-21 01:12:08 +08:00
|
|
|
if (axis == INT64_MAX) {
|
2024-04-13 06:18:22 +08:00
|
|
|
SmallVector<int64_t> nonzeroShape = {resultSizes[0]};
|
|
|
|
auto dtype =
|
|
|
|
dyn_cast<Torch::ValueTensorType>(conditionTensor.getType())
|
|
|
|
.getDtype();
|
|
|
|
auto nonzeroType =
|
|
|
|
rewriter.getType<Torch::ValueTensorType>(nonzeroShape, dtype);
|
|
|
|
Value indexVal = rewriter.create<Torch::AtenNonzeroOp>(
|
|
|
|
binder.getLoc(), nonzeroType, conditionTensor);
|
2024-03-21 01:12:08 +08:00
|
|
|
Value cstZero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
Value cstNegOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(-1));
|
|
|
|
int64_t numElements = 1;
|
|
|
|
for (auto i : shapeSizes) {
|
|
|
|
numElements *= i;
|
|
|
|
}
|
|
|
|
SmallVector<int64_t> flattenShape = {numElements};
|
2024-04-13 06:18:22 +08:00
|
|
|
auto flattenType = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
flattenShape, resultType.getDtype());
|
2024-03-21 01:12:08 +08:00
|
|
|
Value flattenTensor = rewriter.create<Torch::AtenFlattenUsingIntsOp>(
|
|
|
|
binder.getLoc(), flattenType, operand, cstZero, cstNegOne);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenIndexSelectOp>(
|
|
|
|
binder.op, resultType, flattenTensor, cstZero, indexVal);
|
|
|
|
return success();
|
|
|
|
}
|
2024-04-13 06:18:22 +08:00
|
|
|
|
|
|
|
// Negative axis value means counting dimensions from the back
|
|
|
|
if (axis < 0)
|
|
|
|
axis += shapeSizes.size();
|
|
|
|
SmallVector<int64_t> nonzeroShape = {resultSizes[axis]};
|
|
|
|
auto dtype = dyn_cast<Torch::ValueTensorType>(conditionTensor.getType())
|
|
|
|
.getDtype();
|
|
|
|
auto nonzeroType =
|
|
|
|
rewriter.getType<Torch::ValueTensorType>(nonzeroShape, dtype);
|
|
|
|
Value indexVal = rewriter.create<Torch::AtenNonzeroOp>(
|
|
|
|
binder.getLoc(), nonzeroType, conditionTensor);
|
|
|
|
Value dimVal = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(axis));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenIndexSelectOp>(
|
|
|
|
binder.op, resultType, operand, dimVal, indexVal);
|
2024-03-21 01:12:08 +08:00
|
|
|
return success();
|
|
|
|
});
|
2023-12-22 00:04:02 +08:00
|
|
|
patterns.onOp(
|
2024-06-09 14:37:20 +08:00
|
|
|
"Concat", 11, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
2023-12-22 00:04:02 +08:00
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
SmallVector<Value> tensors;
|
|
|
|
int64_t dim;
|
|
|
|
if (binder.tensorOperands(tensors, binder.op->getNumOperands()) ||
|
|
|
|
binder.s64IntegerAttr(dim, "axis", 0) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
Type listElemType =
|
2024-04-28 05:00:56 +08:00
|
|
|
cast<Torch::BaseTensorType>(tensors[0].getType())
|
2023-12-22 00:04:02 +08:00
|
|
|
.getWithSizesAndDtype(/*optionalSizes=*/std::nullopt,
|
|
|
|
/*optionalDtype=*/nullptr);
|
|
|
|
Type listType = Torch::ListType::get(listElemType);
|
|
|
|
Value tensorList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.op->getLoc(), listType, tensors);
|
|
|
|
Value cstDim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(dim));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenCatOp>(binder.op, resultType,
|
|
|
|
tensorList, cstDim);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-16 01:31:22 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Constant", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
if (binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
auto dtype = resultType.getDtype();
|
|
|
|
|
|
|
|
float floatValue;
|
|
|
|
if (binder.op->hasAttr("torch.onnx.value_float") &&
|
|
|
|
!binder.f32FloatAttr(floatValue, "value_float", 0.0)) {
|
|
|
|
auto splatAttr =
|
2024-06-08 09:36:32 +08:00
|
|
|
SplatElementsAttr::get(resultType.toBuiltinTensor(),
|
2024-01-16 01:31:22 +08:00
|
|
|
rewriter.getFloatAttr(dtype, floatValue));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, splatAttr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
int64_t intValue;
|
|
|
|
if (binder.op->hasAttr("torch.onnx.value_int") &&
|
|
|
|
!binder.s64IntegerAttr(intValue, "value_int", 0)) {
|
|
|
|
auto splatAttr =
|
2024-06-08 09:36:32 +08:00
|
|
|
SplatElementsAttr::get(resultType.toBuiltinTensor(),
|
2024-01-16 01:31:22 +08:00
|
|
|
rewriter.getIntegerAttr(dtype, intValue));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, splatAttr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2024-02-01 03:40:53 +08:00
|
|
|
if (DenseResourceElementsAttr attr =
|
2024-05-24 00:01:47 +08:00
|
|
|
dyn_cast_or_null<DenseResourceElementsAttr>(
|
|
|
|
binder.op->getAttr("torch.onnx.value"))) {
|
2024-02-01 03:40:53 +08:00
|
|
|
// Bytes are stored in little endian order. Big endian support will
|
|
|
|
// require swizzling.
|
|
|
|
if (!Endian::little) {
|
|
|
|
binder.op->emitError(
|
|
|
|
"unimplemented: importing on big endian systems");
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
|
|
|
|
auto ty = cast<ShapedType>(attr.getType());
|
2024-03-06 05:55:13 +08:00
|
|
|
ElementsAttr denseAttr;
|
2024-05-30 07:56:23 +08:00
|
|
|
auto ptr = attr.getRawHandle().getBlob();
|
|
|
|
if (!ptr) {
|
|
|
|
denseAttr = DenseResourceElementsAttr::get(
|
|
|
|
ty, "__onnx_constant_not_found_possibly_due_to_being_elided__",
|
|
|
|
AsmResourceBlob());
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, denseAttr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
auto data = ptr->getData();
|
2024-03-06 05:55:13 +08:00
|
|
|
if (cast<ShapedType>(attr.getType()).getElementType().isInteger(1)) {
|
|
|
|
llvm::SmallVector<APInt> newContents;
|
2024-05-30 07:56:23 +08:00
|
|
|
for (auto val : data) {
|
2024-03-06 05:55:13 +08:00
|
|
|
APInt apval(1, val);
|
|
|
|
newContents.push_back(apval);
|
|
|
|
}
|
|
|
|
denseAttr = DenseElementsAttr::get(ty, newContents);
|
|
|
|
} else {
|
2024-05-30 07:56:23 +08:00
|
|
|
denseAttr = DenseElementsAttr::getFromRawBuffer(ty, data);
|
2024-03-06 05:55:13 +08:00
|
|
|
}
|
|
|
|
|
2024-02-01 03:40:53 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, denseAttr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2024-05-24 00:01:47 +08:00
|
|
|
if (ElementsAttr attr = dyn_cast_or_null<ElementsAttr>(
|
|
|
|
binder.op->getAttr("torch.onnx.value"))) {
|
2024-01-16 01:31:22 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, attr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
llvm::SmallVector<int64_t> intValues;
|
|
|
|
if (!binder.s64IntegerArrayAttr(intValues, "value_ints", {}) &&
|
|
|
|
!intValues.empty()) {
|
|
|
|
llvm::SmallVector<APInt> apValues;
|
|
|
|
for (auto intVal : intValues) {
|
|
|
|
apValues.push_back(APInt(dtype.getIntOrFloatBitWidth(), intVal));
|
|
|
|
}
|
2024-06-08 09:36:32 +08:00
|
|
|
auto attr =
|
|
|
|
DenseElementsAttr::get(resultType.toBuiltinTensor(), apValues);
|
2024-01-16 01:31:22 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::ValueTensorLiteralOp>(
|
|
|
|
binder.op, resultType, attr);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
return failure();
|
|
|
|
});
|
2023-12-19 20:29:23 +08:00
|
|
|
patterns.onOp(
|
2024-02-06 08:09:41 +08:00
|
|
|
"Conv", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
2023-12-19 20:29:23 +08:00
|
|
|
std::string autoPad;
|
|
|
|
if (binder.customOpNameStringAttr(autoPad, "auto_pad", "NOTSET"))
|
|
|
|
return failure();
|
|
|
|
if (autoPad != "NOTSET") {
|
|
|
|
// TODO: Add support for `auto_pad` != "NOTSET"
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: auto_pad != NOTSET");
|
|
|
|
}
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input, weight;
|
|
|
|
int64_t group;
|
2023-12-22 16:01:13 +08:00
|
|
|
if (binder.tensorOperandAtIndex(input, 0) ||
|
|
|
|
binder.tensorOperandAtIndex(weight, 1) ||
|
2023-12-19 20:29:23 +08:00
|
|
|
binder.s64IntegerAttr(group, "group", 1) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
2024-04-28 05:00:56 +08:00
|
|
|
auto weightTensorType = cast<Torch::ValueTensorType>(weight.getType());
|
2023-12-19 20:29:23 +08:00
|
|
|
if (!weightTensorType || !weightTensorType.hasSizes()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Expected weight type having sizes");
|
|
|
|
}
|
|
|
|
ArrayRef<int64_t> weightShape = weightTensorType.getSizes();
|
|
|
|
SmallVector<int64_t> kernelShape;
|
|
|
|
if (binder.s64IntegerArrayAttr(kernelShape, "kernel_shape", {}))
|
|
|
|
return failure();
|
|
|
|
if (kernelShape.size()) {
|
|
|
|
if (kernelShape.size() != weightShape.size() - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: kernel_shape list size should have "
|
|
|
|
"number of values equal to weight_rank - 2");
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0; i < kernelShape.size(); i++) {
|
|
|
|
if (weightShape[i + 2] != kernelShape[i]) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: kernel_shape value "
|
|
|
|
"should be equal to the weight tensor shape");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Determine the rank of input tensor.
|
|
|
|
std::optional<unsigned> maybeRank = Torch::getTensorRank(input);
|
|
|
|
if (!maybeRank)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"Unimplemented: unranked tensor");
|
|
|
|
unsigned rank = *maybeRank;
|
|
|
|
|
|
|
|
SmallVector<int64_t> padding, strides, dilations;
|
|
|
|
SmallVector<int64_t> defaultPadding, defaultStrides, defaultDilations;
|
|
|
|
for (unsigned i = 0; i < rank - 2; i++) {
|
|
|
|
defaultPadding.push_back(0);
|
|
|
|
defaultStrides.push_back(1);
|
|
|
|
defaultDilations.push_back(1);
|
|
|
|
}
|
|
|
|
// Padding for the beginning and ending along each spatial axis, it can
|
|
|
|
// take any value greater than or equal to 0. The value represent the
|
|
|
|
// number of pixels added to the beginning and end part of the
|
|
|
|
// corresponding axis. pads format should be as follow [x1_begin,
|
|
|
|
// x2_begin…x1_end, x2_end,…], where xi_begin the number of pixels added
|
|
|
|
// at the beginning of axis i and xi_end, the number of pixels added at
|
|
|
|
// the end of axis i.
|
|
|
|
if (binder.s64IntegerArrayAttr(padding, "pads", defaultPadding)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (padding.size() != rank - 2 && padding.size() != 2 * (rank - 2)) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "padding list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(dilations, "dilations",
|
|
|
|
defaultDilations)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (dilations.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"dilations list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(strides, "strides", defaultStrides)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (strides.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "strides list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
|
|
|
|
SmallVector<Value> cstPadding, cstStrides, cstDilations,
|
|
|
|
cstOutputPadding;
|
2024-06-08 00:54:39 +08:00
|
|
|
Value paddedInput = input;
|
|
|
|
Value paddingList;
|
2023-12-19 20:29:23 +08:00
|
|
|
if (padding.size() != 2 * (rank - 2)) {
|
|
|
|
for (int64_t i : padding) {
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
2024-06-08 00:54:39 +08:00
|
|
|
paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(
|
|
|
|
Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstPadding);
|
2023-12-19 20:29:23 +08:00
|
|
|
} else {
|
2024-06-08 00:54:39 +08:00
|
|
|
// ONNX offers pads in the format listing all starting dims, then all
|
|
|
|
// ending dims, e.g. {t, l, b, r} for conv2d. Torch by default accepts
|
|
|
|
// only starting dims, e.g. {t, l}. However, we can support padding at
|
|
|
|
// the beginning and end of each dimension by first performing
|
|
|
|
// torch.nn.functional.pad on the input. But this requires the pad
|
|
|
|
// values to be rearranged since torch pad() takes pads in the order
|
|
|
|
// rightmost dim start and end, then next to last, and so on, e.g. {l,
|
|
|
|
// r, t, b}.
|
|
|
|
bool matchedPads = true;
|
2023-12-19 20:29:23 +08:00
|
|
|
for (unsigned i = 0; i < padding.size() / 2; i++) {
|
|
|
|
if (padding[i] != padding[i + (padding.size() / 2)]) {
|
2024-06-08 00:54:39 +08:00
|
|
|
matchedPads = false;
|
|
|
|
break;
|
2023-12-19 20:29:23 +08:00
|
|
|
}
|
2024-06-08 00:54:39 +08:00
|
|
|
}
|
|
|
|
if (matchedPads) {
|
|
|
|
for (unsigned i = 0; i < padding.size() / 2; i++) {
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(padding[i])));
|
|
|
|
}
|
|
|
|
paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(
|
|
|
|
Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstPadding);
|
|
|
|
} else {
|
|
|
|
SmallVector<Value> padsRearrange;
|
|
|
|
SmallVector<Value> inputPaddingList;
|
|
|
|
for (uint32_t i = 0; i < padding.size() / 2; i++) {
|
|
|
|
padsRearrange.emplace_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(padding[i])));
|
|
|
|
padsRearrange.emplace_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(
|
|
|
|
padding[(padding.size() / 2) + i])));
|
|
|
|
inputPaddingList.emplace_back(
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0)));
|
|
|
|
}
|
|
|
|
// The conv op itself will have no padding since the actual padding
|
|
|
|
// is performed using the torch.pad preceding it.
|
|
|
|
paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(
|
|
|
|
Torch::IntType::get(binder.op->getContext())),
|
|
|
|
inputPaddingList);
|
|
|
|
Value padsSizeList =
|
|
|
|
rewriter
|
|
|
|
.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(
|
|
|
|
rewriter.getType<Torch::IntType>()),
|
|
|
|
padsRearrange)
|
|
|
|
.getResult();
|
|
|
|
Value modeVal = rewriter.create<Torch::ConstantStrOp>(
|
|
|
|
binder.getLoc(), rewriter.getStringAttr("constant"));
|
|
|
|
Value constantValue;
|
|
|
|
auto inputTensorType =
|
|
|
|
cast<Torch::ValueTensorType>(input.getType());
|
|
|
|
if (isa<IntegerType>(inputTensorType.getDtype()))
|
|
|
|
constantValue = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
if (isa<FloatType>(inputTensorType.getDtype()))
|
|
|
|
constantValue = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(), rewriter.getF64FloatAttr(0.0f));
|
|
|
|
// Pad output shape must be computed explicitly from the pad values
|
|
|
|
SmallVector<int64_t> newInputShape(inputTensorType.getSizes());
|
|
|
|
for (uint32_t i = 0; i < padding.size() / 2; i++) {
|
|
|
|
newInputShape[2 + i] +=
|
|
|
|
padding[i] + padding[(padding.size() / 2) + i];
|
|
|
|
}
|
|
|
|
auto padTy = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
newInputShape, inputTensorType.getDtype());
|
|
|
|
paddedInput = rewriter.create<Torch::AtenPadOp>(
|
|
|
|
binder.getLoc(), padTy, input, padsSizeList, modeVal,
|
|
|
|
constantValue);
|
2023-12-19 20:29:23 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
for (int64_t i : dilations) {
|
|
|
|
cstDilations.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
for (int64_t i : strides) {
|
|
|
|
cstStrides.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
Value cstZero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
cstOutputPadding = {cstZero, cstZero};
|
|
|
|
|
|
|
|
Value dilationsList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstDilations);
|
|
|
|
Value stridesList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstStrides);
|
|
|
|
Value outputPaddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstOutputPadding);
|
|
|
|
Value transposed =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), false);
|
|
|
|
Value bias;
|
|
|
|
if (binder.op->getNumOperands() == 3) {
|
|
|
|
if (binder.tensorOperandAtIndex(bias, 2)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
bias = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
}
|
|
|
|
Value cstGroup = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(group));
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenConvolutionOp>(
|
2024-06-08 00:54:39 +08:00
|
|
|
binder.op, resultType, paddedInput, weight, bias, stridesList,
|
2024-04-24 00:42:02 +08:00
|
|
|
paddingList, dilationsList, transposed, outputPaddingList,
|
|
|
|
cstGroup);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp(
|
|
|
|
"ConvInteger", 10,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
std::string autoPad;
|
|
|
|
if (binder.customOpNameStringAttr(autoPad, "auto_pad", "NOTSET"))
|
|
|
|
return failure();
|
|
|
|
if (autoPad != "NOTSET")
|
|
|
|
// TODO: Add support for `auto_pad` != "NOTSET"
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: auto_pad != NOTSET");
|
|
|
|
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input, weight, inputZp, weightZp;
|
|
|
|
int64_t group;
|
|
|
|
if (binder.tensorOperandAtIndex(input, 0) ||
|
|
|
|
binder.tensorOperandAtIndex(weight, 1) ||
|
|
|
|
binder.s64IntegerAttr(group, "group", 1) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
auto inputTy = dyn_cast<Torch::ValueTensorType>(input.getType());
|
|
|
|
auto weightTy = dyn_cast<Torch::ValueTensorType>(weight.getType());
|
|
|
|
if (!weightTy || !weightTy.hasSizes())
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Expected weight type having sizes");
|
|
|
|
ArrayRef<int64_t> weightShape = weightTy.getSizes();
|
|
|
|
SmallVector<int64_t> kernelShape;
|
|
|
|
if (binder.s64IntegerArrayAttr(kernelShape, "kernel_shape", {}))
|
|
|
|
return failure();
|
|
|
|
if (kernelShape.size()) {
|
|
|
|
if (kernelShape.size() != weightShape.size() - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: kernel_shape list size should have "
|
|
|
|
"number of values equal to weight_rank - 2");
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0; i < kernelShape.size(); i++) {
|
|
|
|
if (weightShape[i + 2] != kernelShape[i])
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: kernel_shape value "
|
|
|
|
"should be equal to the weight tensor shape");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Determine the rank of input tensor.
|
|
|
|
std::optional<unsigned> maybeRank = Torch::getTensorRank(input);
|
|
|
|
if (!maybeRank)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"Unimplemented: unranked tensor");
|
|
|
|
unsigned rank = *maybeRank;
|
|
|
|
|
|
|
|
SmallVector<int64_t> padding, strides, dilations;
|
|
|
|
SmallVector<int64_t> defaultPadding(rank - 2, 0),
|
|
|
|
defaultStrides(rank - 2, 1), defaultDilations(rank - 2, 1);
|
|
|
|
// Padding for the beginning and ending along each spatial axis, it can
|
|
|
|
// take any value greater than or equal to 0. The value represent the
|
|
|
|
// number of pixels added to the beginning and end part of the
|
|
|
|
// corresponding axis. pads format should be as follow [x1_begin,
|
|
|
|
// x2_begin…x1_end, x2_end,…], where xi_begin the number of pixels added
|
|
|
|
// at the beginning of axis i and xi_end, the number of pixels added at
|
|
|
|
// the end of axis i.
|
|
|
|
if (binder.s64IntegerArrayAttr(padding, "pads", defaultPadding))
|
|
|
|
return failure();
|
|
|
|
if (padding.size() != rank - 2 && padding.size() != 2 * (rank - 2))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "padding list size does not match the number of axes");
|
|
|
|
if (binder.s64IntegerArrayAttr(dilations, "dilations",
|
|
|
|
defaultDilations))
|
|
|
|
return failure();
|
|
|
|
if (dilations.size() != rank - 2)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"dilations list size does not match the number of axes");
|
|
|
|
if (binder.s64IntegerArrayAttr(strides, "strides", defaultStrides))
|
|
|
|
return failure();
|
|
|
|
if (strides.size() != rank - 2)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "strides list size does not match the number of axes");
|
|
|
|
|
|
|
|
Value scale = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::FloatType>(),
|
|
|
|
rewriter.getF64FloatAttr(1.0));
|
|
|
|
if (binder.tensorOperandAtIndex(inputZp, 2)) {
|
|
|
|
inputZp = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
} else {
|
|
|
|
inputZp = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(), inputZp);
|
|
|
|
}
|
|
|
|
if (binder.tensorOperandAtIndex(weightZp, 3))
|
|
|
|
weightZp = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
// TODO: support per channel quantization if weightZp is a 1-D tensor
|
|
|
|
if (auto zpTy = dyn_cast<Torch::ValueTensorType>(weightZp.getType())) {
|
|
|
|
for (auto dim : zpTy.getSizes())
|
|
|
|
if (dim != 1)
|
|
|
|
return failure();
|
|
|
|
weightZp = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(), weightZp);
|
|
|
|
}
|
|
|
|
|
|
|
|
SmallVector<Value> cstPadding;
|
|
|
|
if (padding.size() != 2 * (rank - 2)) {
|
|
|
|
for (int64_t i : padding) {
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0; i < padding.size() / 2; i++) {
|
|
|
|
if (padding[i] != padding[i + (padding.size() / 2)])
|
|
|
|
// TODO: Add support for different padding values for the
|
|
|
|
// beginning and ending along each spatial axis
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: padding values for the beginning "
|
|
|
|
"and ending along each spatial axis must be equal");
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(padding[i])));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Value paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getType<Torch::ListType>(
|
|
|
|
rewriter.getType<Torch::IntType>()),
|
|
|
|
cstPadding);
|
|
|
|
Value dilationsList =
|
|
|
|
createConstantIntList(binder, rewriter, dilations);
|
|
|
|
Value stridesList = createConstantIntList(binder, rewriter, strides);
|
|
|
|
Value outputPaddingList =
|
|
|
|
createConstantIntList(binder, rewriter, {0, 0});
|
|
|
|
Value transposed =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), false);
|
|
|
|
Value bias = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
Value cstGroup = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(group));
|
|
|
|
|
|
|
|
Type inputQTy = getQTorchTypeFromTorchIntType(inputTy);
|
|
|
|
Type weightQTy = getQTorchTypeFromTorchIntType(weightTy);
|
|
|
|
input = rewriter.create<Torch::Aten_MakePerTensorQuantizedTensorOp>(
|
|
|
|
binder.getLoc(), inputQTy, input, scale, inputZp);
|
|
|
|
weight = rewriter.create<Torch::Aten_MakePerTensorQuantizedTensorOp>(
|
|
|
|
binder.getLoc(), weightQTy, weight, scale, weightZp);
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenConvolutionOp>(
|
|
|
|
binder.op, resultType, input, weight, bias, stridesList,
|
2023-12-19 20:29:23 +08:00
|
|
|
paddingList, dilationsList, transposed, outputPaddingList,
|
|
|
|
cstGroup);
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp(
|
|
|
|
"ConvTranspose", 11,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
std::string autoPad;
|
|
|
|
if (binder.customOpNameStringAttr(autoPad, "auto_pad", "NOTSET"))
|
|
|
|
return failure();
|
|
|
|
if (autoPad != "NOTSET") {
|
|
|
|
// TODO: Add support for `auto_pad` != "NOTSET"
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: auto_pad != NOTSET");
|
|
|
|
}
|
|
|
|
SmallVector<int64_t> outputShape;
|
|
|
|
if (binder.s64IntegerArrayAttr(outputShape, "output_shape", {}))
|
|
|
|
return failure();
|
|
|
|
if (outputShape.size()) {
|
|
|
|
// TODO: Add support for non-None output_shape value.
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: output_shape should be absent");
|
|
|
|
}
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input, weight;
|
|
|
|
int64_t group;
|
2023-12-22 16:01:13 +08:00
|
|
|
if (binder.tensorOperandAtIndex(input, 0) ||
|
|
|
|
binder.tensorOperandAtIndex(weight, 1) ||
|
2023-12-19 20:29:23 +08:00
|
|
|
binder.s64IntegerAttr(group, "group", 1) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
2024-04-28 05:00:56 +08:00
|
|
|
auto weightTensorType = cast<Torch::ValueTensorType>(weight.getType());
|
2023-12-19 20:29:23 +08:00
|
|
|
if (!weightTensorType || !weightTensorType.hasSizes()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Expected weight type having sizes");
|
|
|
|
}
|
|
|
|
ArrayRef<int64_t> weightShape = weightTensorType.getSizes();
|
|
|
|
SmallVector<int64_t> kernelShape;
|
|
|
|
if (binder.s64IntegerArrayAttr(kernelShape, "kernel_shape", {}))
|
|
|
|
return failure();
|
|
|
|
if (kernelShape.size()) {
|
|
|
|
if (kernelShape.size() != weightShape.size() - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: kernel_shape list size should have "
|
|
|
|
"number of values equal to weight_rank - 2");
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0; i < kernelShape.size(); i++) {
|
|
|
|
if (weightShape[i + 2] != kernelShape[i]) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "unsupported conversion: kernel_shape value "
|
|
|
|
"should be equal to the weight tensor shape");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Determine the rank of input tensor.
|
|
|
|
std::optional<unsigned> maybeRank = Torch::getTensorRank(input);
|
|
|
|
if (!maybeRank)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"Unimplemented: unranked tensor");
|
|
|
|
unsigned rank = *maybeRank;
|
|
|
|
|
|
|
|
SmallVector<int64_t> padding, strides, dilations, outputPadding;
|
2024-01-23 23:36:25 +08:00
|
|
|
SmallVector<int64_t> defaultPadding, defaultStrides, defaultDilations,
|
|
|
|
defaultOutputPadding;
|
2023-12-19 20:29:23 +08:00
|
|
|
for (unsigned i = 0; i < rank - 2; i++) {
|
|
|
|
defaultPadding.push_back(0);
|
|
|
|
defaultStrides.push_back(1);
|
|
|
|
defaultDilations.push_back(1);
|
|
|
|
defaultOutputPadding.push_back(0);
|
|
|
|
}
|
|
|
|
// Padding for the beginning and ending along each spatial axis, it can
|
|
|
|
// take any value greater than or equal to 0. The value represent the
|
|
|
|
// number of pixels added to the beginning and end part of the
|
|
|
|
// corresponding axis. pads format should be as follow [x1_begin,
|
|
|
|
// x2_begin…x1_end, x2_end,…], where xi_begin the number of pixels added
|
|
|
|
// at the beginning of axis i and xi_end, the number of pixels added at
|
|
|
|
// the end of axis i.
|
|
|
|
if (binder.s64IntegerArrayAttr(padding, "pads", defaultPadding)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (padding.size() != rank - 2 && padding.size() != 2 * (rank - 2)) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "padding list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(dilations, "dilations",
|
|
|
|
defaultDilations)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (dilations.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"dilations list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(strides, "strides", defaultStrides)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (strides.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "strides list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
if (binder.s64IntegerArrayAttr(outputPadding, "output_padding",
|
|
|
|
defaultOutputPadding)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
if (outputPadding.size() != rank - 2) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"output_padding list size does not match the number of axes");
|
|
|
|
}
|
|
|
|
|
|
|
|
SmallVector<Value> cstPadding, cstStrides, cstDilations,
|
|
|
|
cstOutputPadding;
|
|
|
|
if (padding.size() != 2 * (rank - 2)) {
|
|
|
|
for (int64_t i : padding) {
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0; i < padding.size() / 2; i++) {
|
|
|
|
if (padding[i] != padding[i + (padding.size() / 2)]) {
|
|
|
|
// TODO: Add support for different padding values for the
|
|
|
|
// beginning and ending along each spatial axis
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unsupported conversion: padding values for the beginning "
|
|
|
|
"and ending along each spatial axis must be equal");
|
|
|
|
}
|
|
|
|
cstPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(padding[i])));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for (int64_t i : dilations) {
|
|
|
|
cstDilations.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
for (int64_t i : strides) {
|
|
|
|
cstStrides.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
for (int64_t i : outputPadding) {
|
|
|
|
cstOutputPadding.push_back(rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(i)));
|
|
|
|
}
|
|
|
|
|
|
|
|
Value paddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstPadding);
|
|
|
|
Value dilationsList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstDilations);
|
|
|
|
Value stridesList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstStrides);
|
|
|
|
Value outputPaddingList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
cstOutputPadding);
|
|
|
|
Value transposed =
|
|
|
|
rewriter.create<Torch::ConstantBoolOp>(binder.getLoc(), true);
|
|
|
|
Value bias;
|
|
|
|
if (binder.op->getNumOperands() == 3) {
|
|
|
|
if (binder.tensorOperandAtIndex(bias, 2)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
bias = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
}
|
|
|
|
Value cstGroup = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(group));
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenConvolutionOp>(
|
|
|
|
binder.op, resultType, input, weight, bias, stridesList,
|
|
|
|
paddingList, dilationsList, transposed, outputPaddingList,
|
|
|
|
cstGroup);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-05 13:55:51 +08:00
|
|
|
patterns.onOp("Cos", 7,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenCosOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-05-04 00:06:44 +08:00
|
|
|
patterns.onOp("Cosh", 9,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenCoshOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-04 01:52:59 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"CumSum", 11, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
2024-05-11 02:09:01 +08:00
|
|
|
Value operand, axisTensor;
|
|
|
|
int64_t exclusive, reverse;
|
2024-01-04 01:52:59 +08:00
|
|
|
if (binder.tensorOperands(operand, axisTensor) ||
|
2024-05-11 02:09:01 +08:00
|
|
|
binder.s64IntegerAttr(exclusive, "exclusive", 0) ||
|
|
|
|
binder.s64IntegerAttr(reverse, "reverse", 0) ||
|
2024-01-04 01:52:59 +08:00
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
2024-05-11 02:09:01 +08:00
|
|
|
Torch::BaseTensorType resultTensorType =
|
|
|
|
cast<Torch::BaseTensorType>(resultType);
|
|
|
|
if (!resultTensorType.hasDtype()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "expected result type to have a dtype");
|
|
|
|
}
|
2024-01-04 01:52:59 +08:00
|
|
|
|
|
|
|
// deal with neg axis: if (axis < 0) axis += rank
|
|
|
|
int64_t rank =
|
|
|
|
cast<Torch::ValueTensorType>(operand.getType()).getSizes().size();
|
|
|
|
Value rankVal = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
2024-01-23 23:36:25 +08:00
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), rank));
|
2024-05-11 02:09:01 +08:00
|
|
|
Value cstZero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
Value cstOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1));
|
2024-01-23 23:36:25 +08:00
|
|
|
|
2024-01-04 01:52:59 +08:00
|
|
|
Value axisScalar = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(), axisTensor);
|
2024-01-23 23:36:25 +08:00
|
|
|
Value isNegative = rewriter.create<Torch::AtenLtIntOp>(
|
2024-05-11 02:09:01 +08:00
|
|
|
binder.getLoc(), axisScalar, cstZero);
|
2024-01-23 23:36:25 +08:00
|
|
|
isNegative =
|
|
|
|
rewriter.create<Torch::AtenIntBoolOp>(binder.getLoc(), isNegative);
|
2024-01-04 01:52:59 +08:00
|
|
|
Value finalOffset = rewriter.create<Torch::AtenMulIntOp>(
|
|
|
|
binder.getLoc(), isNegative, rankVal);
|
2024-05-11 02:09:01 +08:00
|
|
|
Value axis = rewriter.create<Torch::AtenAddIntOp>(
|
2024-01-04 01:52:59 +08:00
|
|
|
binder.getLoc(), axisScalar, finalOffset);
|
2024-05-11 02:09:01 +08:00
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
2024-01-04 01:52:59 +08:00
|
|
|
|
2024-05-11 02:09:01 +08:00
|
|
|
Value res;
|
|
|
|
if (reverse) {
|
|
|
|
Value dims = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getType<Torch::ListType>(
|
|
|
|
rewriter.getType<Torch::IntType>()),
|
|
|
|
SmallVector<Value>{axis});
|
|
|
|
Value flip = rewriter.create<Torch::AtenFlipOp>(
|
|
|
|
binder.getLoc(), resultType, operand, dims);
|
|
|
|
Value cumsum = rewriter.create<Torch::AtenCumsumOp>(
|
|
|
|
binder.getLoc(), resultType, flip, axis, none);
|
|
|
|
res = rewriter.create<Torch::AtenFlipOp>(binder.getLoc(), resultType,
|
|
|
|
cumsum, dims);
|
|
|
|
} else {
|
|
|
|
res = rewriter.create<Torch::AtenCumsumOp>(
|
|
|
|
binder.getLoc(), resultType, operand, axis, none);
|
2024-01-04 01:52:59 +08:00
|
|
|
}
|
2024-05-11 02:09:01 +08:00
|
|
|
|
|
|
|
if (exclusive)
|
|
|
|
res = rewriter.create<Torch::AtenSubTensorOp>(
|
|
|
|
binder.getLoc(), resultType, res, operand, cstOne);
|
|
|
|
rewriter.replaceOp(binder.op, res);
|
2024-01-04 01:52:59 +08:00
|
|
|
return success();
|
|
|
|
});
|
2024-01-03 20:55:56 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"DepthToSpace", 1,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input;
|
|
|
|
int64_t blockSize;
|
|
|
|
std::string mode;
|
|
|
|
if (binder.tensorOperand(input) ||
|
|
|
|
binder.s64IntegerAttr(blockSize, "blocksize") ||
|
|
|
|
binder.customOpNameStringAttr(mode, "mode", "DCR") ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
2024-04-28 05:00:56 +08:00
|
|
|
auto inputTy = dyn_cast<Torch::BaseTensorType>(input.getType());
|
2024-01-03 20:55:56 +08:00
|
|
|
if (!inputTy || !inputTy.hasSizes()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Expected input type having sizes");
|
|
|
|
}
|
|
|
|
SmallVector<int64_t> inputSizes{inputTy.getSizes()};
|
|
|
|
if (inputSizes.size() != 4) {
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"Expected input rank to be 4");
|
|
|
|
}
|
|
|
|
Value b = rewriter.create<Torch::AtenSizeIntOp>(
|
|
|
|
binder.getLoc(), input,
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0)));
|
|
|
|
Value c = rewriter.create<Torch::AtenSizeIntOp>(
|
|
|
|
binder.getLoc(), input,
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1)));
|
|
|
|
Value h = rewriter.create<Torch::AtenSizeIntOp>(
|
|
|
|
binder.getLoc(), input,
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(2)));
|
|
|
|
Value w = rewriter.create<Torch::AtenSizeIntOp>(
|
|
|
|
binder.getLoc(), input,
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(3)));
|
|
|
|
Value cstBlockSize = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(blockSize));
|
|
|
|
Value cstBlockSizeSquare = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(blockSize * blockSize));
|
|
|
|
Value cDivBlockSizeSquare = rewriter.create<Torch::AtenDivIntOp>(
|
|
|
|
binder.getLoc(), c, cstBlockSizeSquare);
|
|
|
|
cDivBlockSizeSquare = rewriter.create<Torch::AtenIntFloatOp>(
|
|
|
|
binder.getLoc(), cDivBlockSizeSquare);
|
|
|
|
Value reshapeSizesList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(input.getContext())),
|
|
|
|
llvm::SmallVector<Value>{b, cstBlockSize, cstBlockSize,
|
|
|
|
cDivBlockSizeSquare, h, w});
|
|
|
|
int64_t cDivBlockSizeSquareInt =
|
|
|
|
inputSizes[1] == Torch::kUnknownSize
|
|
|
|
? Torch::kUnknownSize
|
|
|
|
: inputSizes[1] / (blockSize * blockSize);
|
|
|
|
SmallVector<int64_t, 6> reshapeSizesInt{
|
|
|
|
inputSizes[0], blockSize, blockSize,
|
|
|
|
cDivBlockSizeSquareInt, inputSizes[2], inputSizes[3]};
|
|
|
|
Value reshapedInput = rewriter.create<Torch::AtenReshapeOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
inputTy.getWithSizesAndDtype(reshapeSizesInt,
|
|
|
|
inputTy.getOptionalDtype()),
|
|
|
|
input, reshapeSizesList);
|
|
|
|
|
|
|
|
Value transposedInput;
|
|
|
|
if (mode == "DCR") {
|
|
|
|
if (failed(createTorchTransposeOp(
|
|
|
|
rewriter, binder.getLoc(), reshapedInput,
|
|
|
|
/*dimA=*/1, /*dimB=*/3, transposedInput)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Failed to create TorchTranspose op");
|
|
|
|
if (failed(createTorchTransposeOp(
|
|
|
|
rewriter, binder.getLoc(), transposedInput,
|
|
|
|
/*dimA=*/2, /*dimB=*/4, transposedInput)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Failed to create TorchTranspose op");
|
|
|
|
} else {
|
|
|
|
// mode == "CRD"
|
|
|
|
if (failed(createTorchTransposeOp(
|
|
|
|
rewriter, binder.getLoc(), reshapedInput,
|
|
|
|
/*dimA=*/2, /*dimB=*/4, transposedInput)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Failed to create TorchTranspose op");
|
|
|
|
if (failed(createTorchTransposeOp(
|
|
|
|
rewriter, binder.getLoc(), transposedInput,
|
|
|
|
/*dimA=*/3, /*dimB=*/4, transposedInput)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Failed to create TorchTranspose op");
|
|
|
|
}
|
|
|
|
if (failed(createTorchTransposeOp(
|
|
|
|
rewriter, binder.getLoc(), transposedInput,
|
|
|
|
/*dimA=*/4, /*dimB=*/5, transposedInput)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op, "Failed to create TorchTranspose op");
|
|
|
|
|
|
|
|
Value hMulBlockSize = rewriter.create<Torch::AtenMulIntOp>(
|
|
|
|
binder.getLoc(), h, cstBlockSize);
|
|
|
|
Value wMulBlockSize = rewriter.create<Torch::AtenMulIntOp>(
|
|
|
|
binder.getLoc(), w, cstBlockSize);
|
|
|
|
reshapeSizesList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(input.getContext())),
|
|
|
|
llvm::SmallVector<Value>{b, cDivBlockSizeSquare, hMulBlockSize,
|
|
|
|
wMulBlockSize});
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenReshapeOp>(
|
|
|
|
binder.op, resultType, transposedInput, reshapeSizesList);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-19 08:47:21 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"DequantizeLinear", 1,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
llvm::SmallVector<Value> operands;
|
|
|
|
if (binder.tensorOperands(operands, 3) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
Value operand = operands[0];
|
|
|
|
Value scale = operands[1];
|
|
|
|
Value zeropoint = operands[2];
|
|
|
|
|
2024-04-28 05:00:56 +08:00
|
|
|
auto operandTy = cast<Torch::ValueTensorType>(operand.getType());
|
2024-01-19 08:47:21 +08:00
|
|
|
|
2024-04-28 05:00:56 +08:00
|
|
|
auto scaleTy = dyn_cast<Torch::ValueTensorType>(scale.getType());
|
2024-01-19 08:47:21 +08:00
|
|
|
if (!scaleTy || !scaleTy.hasSizes())
|
|
|
|
return rewriter.notifyMatchFailure(binder.op, "requires known rank");
|
|
|
|
if (!resultType.hasDtype())
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
2024-03-21 01:58:25 +08:00
|
|
|
"requires known result dtype");
|
2024-04-02 07:21:05 +08:00
|
|
|
if (scaleTy.getSizes().size() == 0 ||
|
|
|
|
(scaleTy.getSizes().size() == 1 && scaleTy.getSizes()[0] == 1)) {
|
2024-06-12 13:07:22 +08:00
|
|
|
auto qTensorTy = getQTorchTypeFromTorchIntType(operandTy);
|
|
|
|
if (!qTensorTy) {
|
2024-01-19 08:47:21 +08:00
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"unsupported result dtype");
|
|
|
|
}
|
|
|
|
|
|
|
|
scale = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::FloatType>(), scale);
|
|
|
|
zeropoint = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(), zeropoint);
|
|
|
|
|
|
|
|
auto quantize =
|
|
|
|
rewriter.create<Torch::Aten_MakePerTensorQuantizedTensorOp>(
|
|
|
|
binder.getLoc(), qTensorTy, operand, scale, zeropoint);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenDequantizeSelfOp>(
|
|
|
|
binder.op, resultType, quantize);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2024-04-02 07:21:05 +08:00
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"unimplemented: non-scalar scale");
|
2024-01-19 08:47:21 +08:00
|
|
|
});
|
2024-03-20 04:35:05 +08:00
|
|
|
patterns.onOp("Div", 7,
|
2023-12-05 13:55:51 +08:00
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenDivTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-08 22:38:49 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Dropout", 12, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Location loc = binder.getLoc();
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
int64_t numOperands = binder.op->getNumOperands();
|
|
|
|
SmallVector<Value> operands;
|
|
|
|
int64_t seed;
|
|
|
|
if (binder.tensorOperands(operands, numOperands) ||
|
|
|
|
binder.s64IntegerAttr(seed, "seed", 0) ||
|
|
|
|
binder.tensorResultTypeAtIndex(resultType, 0))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
// Global Seed value is 0.
|
|
|
|
if (seed != 0) {
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"expected seed value to be 0");
|
|
|
|
}
|
|
|
|
|
|
|
|
Value ratio, trainingMode;
|
|
|
|
if (numOperands == 3) {
|
|
|
|
ratio = rewriter.create<Torch::AtenFloatImplicitOp>(loc, operands[1]);
|
2024-02-27 13:32:05 +08:00
|
|
|
Value trainVal = operands[2];
|
|
|
|
auto trainTensorType =
|
2024-04-28 05:00:56 +08:00
|
|
|
dyn_cast<Torch::BaseTensorType>(trainVal.getType());
|
2024-02-27 13:32:05 +08:00
|
|
|
if (!trainTensorType)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"train tensor must have a type");
|
|
|
|
|
|
|
|
Type inputDtype = trainTensorType.getOptionalDtype();
|
|
|
|
if (!inputDtype || !inputDtype.isInteger(1))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"train tensor must have an integer dtype of width 1");
|
|
|
|
|
|
|
|
std::optional<unsigned> inputRank = Torch::getTensorRank(trainVal);
|
|
|
|
if (!inputRank || *inputRank != 0)
|
|
|
|
return rewriter.notifyMatchFailure(binder.op,
|
|
|
|
"train tensor must have rank 0");
|
|
|
|
|
|
|
|
if (auto valueTensorLiteralOp =
|
|
|
|
trainVal.getDefiningOp<Torch::ValueTensorLiteralOp>()) {
|
2024-04-28 05:00:56 +08:00
|
|
|
auto val = cast<DenseElementsAttr>(valueTensorLiteralOp.getValue())
|
2024-02-27 13:32:05 +08:00
|
|
|
.getSplatValue<bool>();
|
|
|
|
trainingMode = rewriter.create<Torch::ConstantBoolOp>(loc, val);
|
|
|
|
} else {
|
|
|
|
Value trainingModeScalar =
|
|
|
|
rewriter.create<Torch::AtenIntImplicitOp>(loc, operands[2]);
|
|
|
|
Value cstOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(1));
|
|
|
|
trainingMode = rewriter.create<Torch::AtenEqIntOp>(
|
|
|
|
loc, trainingModeScalar, cstOne);
|
|
|
|
}
|
2024-01-08 22:38:49 +08:00
|
|
|
} else if (numOperands == 2) {
|
|
|
|
ratio = rewriter.create<Torch::AtenFloatImplicitOp>(loc, operands[1]);
|
|
|
|
trainingMode = rewriter.create<Torch::ConstantBoolOp>(loc, false);
|
|
|
|
} else {
|
|
|
|
ratio = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(0.5));
|
|
|
|
trainingMode = rewriter.create<Torch::ConstantBoolOp>(loc, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
Value dropout = rewriter.create<Torch::AtenDropoutOp>(
|
|
|
|
loc, resultType, /*input=*/operands[0], ratio, trainingMode);
|
|
|
|
|
|
|
|
if (binder.op->getNumResults() == 1) {
|
|
|
|
rewriter.replaceOp(binder.op, dropout);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
Torch::ValueTensorType maskType;
|
|
|
|
if (binder.tensorResultTypeAtIndex(maskType, 1))
|
|
|
|
return failure();
|
|
|
|
Value dtype = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(
|
|
|
|
(int64_t)torch_upstream::ScalarType::Bool));
|
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(loc);
|
|
|
|
Value mask = rewriter.create<Torch::AtenOnesLikeOp>(
|
|
|
|
loc, maskType, operands[0], dtype, /*layout=*/none,
|
|
|
|
/*device=*/none, /*pin_memory=*/none, /*memory_format=*/none);
|
|
|
|
rewriter.replaceOp(binder.op, {dropout, mask});
|
|
|
|
return success();
|
|
|
|
});
|
2024-03-21 01:58:25 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"DynamicQuantizeLinear", 11,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Location loc = binder.getLoc();
|
|
|
|
Value input;
|
|
|
|
Torch::ValueTensorType resultType, scaleType, zeroPointType;
|
|
|
|
if (binder.tensorOperand(input) ||
|
|
|
|
binder.tensorResultTypeAtIndex(resultType, 0) ||
|
|
|
|
binder.tensorResultTypeAtIndex(scaleType, 1) ||
|
|
|
|
binder.tensorResultTypeAtIndex(zeroPointType, 2))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
Value scale, zeroPoint;
|
|
|
|
|
|
|
|
// scale = ( max(0, max(input)) - min(0, min(input)) ) / 255
|
|
|
|
Value inputMax =
|
|
|
|
rewriter.create<Torch::AtenMaxOp>(loc, scaleType, input);
|
|
|
|
Value inputMin =
|
|
|
|
rewriter.create<Torch::AtenMinOp>(loc, scaleType, input);
|
|
|
|
Value constantZero = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(0));
|
|
|
|
Value constantOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(1));
|
|
|
|
Value zeroTensor =
|
|
|
|
createRank0Tensor(rewriter, loc, scaleType, constantZero);
|
|
|
|
Value inputMaxW0 = rewriter.create<Torch::AtenMaximumOp>(
|
|
|
|
loc, scaleType, inputMax, zeroTensor);
|
|
|
|
Value inputMinW0 = rewriter.create<Torch::AtenMinimumOp>(
|
|
|
|
loc, scaleType, inputMin, zeroTensor);
|
|
|
|
Value scaleTensor = rewriter.create<Torch::AtenSubTensorOp>(
|
|
|
|
loc, scaleType, inputMaxW0, inputMinW0, constantOne);
|
|
|
|
// Note: the following is hard-coded for ui8
|
|
|
|
Value width = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(255));
|
|
|
|
Value widthTensor = createRank0Tensor(rewriter, loc, scaleType, width);
|
|
|
|
scaleTensor = rewriter.create<Torch::AtenDivTensorOp>(
|
|
|
|
loc, scaleType, scaleTensor, widthTensor);
|
|
|
|
// compute the preZeroPoint = 0 - (inputMin/scale)
|
|
|
|
// compute the zeroPoint = cast ( round (clip or saturate
|
|
|
|
// (preZeroPoint)))
|
|
|
|
Value preZeroPoint = rewriter.create<Torch::AtenDivTensorOp>(
|
|
|
|
loc, scaleType, inputMin, scaleTensor);
|
|
|
|
preZeroPoint = rewriter.create<Torch::AtenSubTensorOp>(
|
|
|
|
loc, scaleType, zeroTensor, preZeroPoint, constantOne);
|
|
|
|
// saturate to interval [0, 255]
|
|
|
|
preZeroPoint = rewriter.create<Torch::AtenClampOp>(
|
|
|
|
loc, scaleType, preZeroPoint, /*min=*/constantZero, /*max=*/width);
|
|
|
|
// round, then cast to uint8
|
|
|
|
preZeroPoint =
|
|
|
|
rewriter.create<Torch::AtenRoundOp>(loc, scaleType, preZeroPoint);
|
|
|
|
Type qTy = rewriter.getType<Torch::QUInt8Type>();
|
|
|
|
auto qTensorTy = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
resultType.getOptionalSizes(), qTy);
|
|
|
|
auto torchqTy = Torch::getScalarTypeForType(qTy);
|
|
|
|
Value tyConst = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64),
|
|
|
|
static_cast<int64_t>(torchqTy)));
|
|
|
|
Value none = rewriter.create<Torch::ConstantNoneOp>(loc);
|
|
|
|
Value cstFalse = rewriter.create<Torch::ConstantBoolOp>(loc, false);
|
|
|
|
Value zeroPointTensor = rewriter.create<Torch::AtenToDtypeOp>(
|
|
|
|
loc, zeroPointType, preZeroPoint, tyConst,
|
|
|
|
/*non_blocking=*/cstFalse, /*copy=*/cstFalse,
|
|
|
|
/*memory_format=*/none);
|
|
|
|
// extract scale and zeroPoint scalars to pass to
|
|
|
|
// AtenQuantizePerTensorOp
|
|
|
|
zeroPoint = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
loc, rewriter.getType<Torch::IntType>(), zeroPointTensor);
|
|
|
|
scale = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
loc, rewriter.getType<Torch::FloatType>(), scaleTensor);
|
|
|
|
Value quantizedTensor = rewriter.create<Torch::AtenQuantizePerTensorOp>(
|
|
|
|
loc, qTensorTy, input, scale, zeroPoint, tyConst);
|
|
|
|
// get uint8 tensor output
|
|
|
|
Value output = rewriter.create<Torch::AtenIntReprOp>(loc, resultType,
|
|
|
|
quantizedTensor);
|
|
|
|
rewriter.replaceOp(binder.op, {output, scaleTensor, zeroPointTensor});
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-28 02:08:09 +08:00
|
|
|
patterns.onOp("Equal", 1,
|
2023-12-05 13:55:51 +08:00
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value lhs, rhs;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperands(lhs, rhs) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenEqTensorOp>(
|
|
|
|
binder.op, resultType, lhs, rhs);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-08 22:38:49 +08:00
|
|
|
patterns.onOp("Elu", 6,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Location loc = binder.getLoc();
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value input;
|
|
|
|
float alpha;
|
|
|
|
if (binder.tensorOperand(input) ||
|
|
|
|
binder.f32FloatAttr(alpha, "alpha") ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
Value cstAlpha = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(alpha));
|
|
|
|
Value cstOne = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(1.0));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenEluOp>(
|
|
|
|
binder.op, resultType, input, cstAlpha, /*scale=*/cstOne,
|
|
|
|
/*input_scale=*/cstOne);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-20 00:07:27 +08:00
|
|
|
patterns.onOp("Erf", 13,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
std::string direction;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenErfOp>(
|
|
|
|
binder.op, resultType, operand);
|
2024-01-24 05:45:00 +08:00
|
|
|
return success();
|
|
|
|
});
|
|
|
|
patterns.onOp("Exp", 6,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenExpOp>(
|
|
|
|
binder.op, resultType, operand);
|
2023-12-20 00:07:27 +08:00
|
|
|
return success();
|
|
|
|
});
|
2024-01-11 05:05:37 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Expand", 1, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
// uses ideas and code from onnx.Reshape
|
2024-03-09 08:23:07 +08:00
|
|
|
auto loc = binder.getLoc();
|
2024-01-11 05:05:37 +08:00
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value data, shape;
|
|
|
|
if (binder.tensorOperands(data, shape) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
2024-03-09 08:23:07 +08:00
|
|
|
|
|
|
|
auto dataType = cast<Torch::BaseTensorType>(data.getType());
|
|
|
|
auto shapeType = cast<Torch::BaseTensorType>(shape.getType());
|
|
|
|
if (!dataType.hasSizes() || !shapeType.hasSizes())
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
auto shapeSizes = shapeType.getSizes();
|
|
|
|
int64_t dataRank = dataType.getSizes().size();
|
|
|
|
int64_t shapeRank = shapeSizes.size();
|
|
|
|
if (shapeRank != 1 || shapeSizes[0] == Torch::kUnknownSize)
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
auto rankDifference = dataRank - shapeSizes[0];
|
|
|
|
|
2024-01-11 05:05:37 +08:00
|
|
|
SmallVector<int64_t> selectSizes;
|
|
|
|
Type selectResultType = shapeType.getWithSizesAndDtype(
|
|
|
|
llvm::ArrayRef(selectSizes), shapeType.getOptionalDtype());
|
|
|
|
// Variable to store 1-D onnx shape tensor, shapeSizes[0] has the
|
|
|
|
// dimension size
|
|
|
|
// A constant zero value
|
|
|
|
Value zero = rewriter.create<Torch::ConstantIntOp>(
|
2024-03-09 08:23:07 +08:00
|
|
|
loc, rewriter.getI64IntegerAttr(0));
|
2024-01-11 05:05:37 +08:00
|
|
|
// Variable to store pytorch int list of shape (dimension)
|
|
|
|
SmallVector<Value> dimList;
|
|
|
|
|
|
|
|
// Convert the shape tensor from vector of int64_t to torch int list as
|
|
|
|
// we are using torch implementation Torch::AtenBroadcastToOp which
|
|
|
|
// takes list of int
|
|
|
|
for (int i = 0; i < shapeSizes[0]; i++) {
|
|
|
|
Value selectIndex = rewriter.create<Torch::ConstantIntOp>(
|
2024-03-09 08:23:07 +08:00
|
|
|
loc, rewriter.getType<Torch::IntType>(),
|
2024-01-11 05:05:37 +08:00
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), i));
|
|
|
|
Value extract = rewriter.create<Torch::AtenSelectIntOp>(
|
2024-03-09 08:23:07 +08:00
|
|
|
loc, selectResultType, shape, zero, selectIndex);
|
2024-01-11 05:05:37 +08:00
|
|
|
Value dim = rewriter.create<Torch::AtenItemOp>(
|
2024-03-09 08:23:07 +08:00
|
|
|
loc, rewriter.getType<Torch::IntType>(), extract);
|
|
|
|
|
|
|
|
if (i + rankDifference >= 0) {
|
|
|
|
Value iv =
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(loc, i + rankDifference);
|
|
|
|
auto sz = rewriter.create<Torch::AtenSizeIntOp>(
|
|
|
|
loc, rewriter.getType<Torch::IntType>(), data, iv);
|
|
|
|
dim = rewriter.create<Torch::PrimMaxIntOp>(loc, dim, sz);
|
|
|
|
}
|
|
|
|
|
2024-01-11 05:05:37 +08:00
|
|
|
dimList.push_back(dim);
|
|
|
|
}
|
|
|
|
Value dimValueList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
dimList);
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenBroadcastToOp>(
|
|
|
|
binder.op, resultType, data, dimValueList);
|
|
|
|
return success();
|
|
|
|
});
|
2024-04-16 00:23:26 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"EyeLike", 9, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
int64_t dtypeIntOnnx, diagonalIndex;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.s64IntegerAttr(dtypeIntOnnx, "dtype", 1) ||
|
|
|
|
binder.s64IntegerAttr(diagonalIndex, "k", 0) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
auto operandTy = cast<Torch::ValueTensorType>(operand.getType());
|
|
|
|
SmallVector<int64_t> shape(operandTy.getSizes());
|
|
|
|
for (unsigned i = 0; i < shape.size(); i++) {
|
|
|
|
if (shape[i] == ShapedType::kDynamic)
|
|
|
|
shape[i] = Torch::kUnknownSize;
|
|
|
|
}
|
|
|
|
|
|
|
|
Value cst0 = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
Value cst1 = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(1));
|
|
|
|
Value nVal = rewriter.create<Torch::AtenSizeIntOp>(binder.getLoc(),
|
|
|
|
operand, cst0);
|
|
|
|
Value mVal = rewriter.create<Torch::AtenSizeIntOp>(binder.getLoc(),
|
|
|
|
operand, cst1);
|
|
|
|
Value noneVal = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
2024-04-23 00:58:07 +08:00
|
|
|
std::optional<int64_t> dtypeIntTorch =
|
|
|
|
onnxDtypeIntToTorchDtypeInt(dtypeIntOnnx);
|
|
|
|
if (!dtypeIntTorch.has_value()) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"unimplemented support for the given dtype conversion");
|
|
|
|
}
|
2024-04-16 00:23:26 +08:00
|
|
|
Value dtypeVal = rewriter.create<Torch::ConstantIntOp>(
|
2024-04-23 00:58:07 +08:00
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(dtypeIntTorch.value()));
|
2024-04-16 00:23:26 +08:00
|
|
|
|
|
|
|
// diagonalIndex = 0 populates the main diagonal
|
|
|
|
// diagonalIndex > 0 populates an upper diagonal
|
|
|
|
// diagonalIndex < 0 populates a lower diagonal
|
|
|
|
if (diagonalIndex == 0) {
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenEyeMOp>(
|
|
|
|
binder.op, resultType, nVal, mVal, dtypeVal, noneVal, noneVal,
|
|
|
|
noneVal);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
Value diagVal = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getI64IntegerAttr(std::abs(diagonalIndex)));
|
|
|
|
Value newN, newM, dimVal, startVal;
|
|
|
|
// get shapes of main diag eye op and zeros op
|
|
|
|
if (diagonalIndex > 0) {
|
|
|
|
newN = nVal;
|
|
|
|
newM = rewriter.create<Torch::AtenSubIntOp>(binder.getLoc(), mVal,
|
|
|
|
diagVal);
|
|
|
|
if (shape[1] != Torch::kUnknownSize) {
|
|
|
|
shape[1] -= diagonalIndex;
|
|
|
|
}
|
|
|
|
dimVal = cst1;
|
|
|
|
startVal = mVal;
|
|
|
|
} else {
|
|
|
|
newN = rewriter.create<Torch::AtenSubIntOp>(binder.getLoc(), nVal,
|
|
|
|
diagVal);
|
|
|
|
newM = mVal;
|
|
|
|
if (shape[0] != Torch::kUnknownSize) {
|
|
|
|
shape[0] += diagonalIndex;
|
|
|
|
}
|
|
|
|
dimVal = cst0;
|
|
|
|
startVal = nVal;
|
|
|
|
}
|
|
|
|
|
|
|
|
// create main diag eye op
|
|
|
|
auto eyeResultType = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
shape, resultType.getOptionalDtype());
|
|
|
|
Value eyeOp = rewriter.create<Torch::AtenEyeMOp>(
|
|
|
|
binder.getLoc(), eyeResultType, newN, newM, dtypeVal, noneVal,
|
|
|
|
noneVal, noneVal);
|
|
|
|
// create zeros op
|
|
|
|
SmallVector<Value> zerosShapeValues = {nVal, mVal};
|
|
|
|
Value zerosShapeList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getType<Torch::ListType>(
|
|
|
|
rewriter.getType<Torch::IntType>()),
|
|
|
|
zerosShapeValues);
|
|
|
|
Value zerosOp = rewriter.create<Torch::AtenZerosOp>(
|
|
|
|
binder.getLoc(), resultType, zerosShapeList, dtypeVal, noneVal,
|
|
|
|
noneVal, noneVal);
|
|
|
|
|
|
|
|
// embeds the values of the eye matrix into zeros
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenSliceScatterOp>(
|
|
|
|
binder.op, resultType, zerosOp, eyeOp, dimVal,
|
|
|
|
/*start=*/diagVal, /*end=*/startVal, /*step=*/cst1);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-20 08:18:16 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Flatten", 13, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
// Flatten means to partition the input tensor's dimensions
|
|
|
|
// into a "left range" spanning 0 to axis - 1 and a "right range"
|
|
|
|
// spanning axis to rank - 1. Each range is then collapsed
|
|
|
|
// into a single dimension, resulting in a 2-D tensor.
|
|
|
|
// If either range is empty, it is replaced with a single
|
|
|
|
// dimension of size 1.
|
|
|
|
//
|
|
|
|
// For example, for a 4-D input tensor of shape (a, b, c, d)
|
|
|
|
// and axis==2, flatten produces a 2-D tensor of shape
|
|
|
|
// (a*b, c*d).
|
|
|
|
//
|
|
|
|
// If instead axis==0, the left range is empty, and the result
|
|
|
|
// is (1, a*b*c*d).
|
|
|
|
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
int64_t axis;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.s64IntegerAttr(axis, "axis", 1) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
2024-02-06 06:23:46 +08:00
|
|
|
auto operandTy = cast<Torch::ValueTensorType>(operand.getType());
|
|
|
|
llvm::SmallVector<int64_t> shape(operandTy.getSizes());
|
|
|
|
int64_t rank = shape.size();
|
|
|
|
|
2024-01-20 08:18:16 +08:00
|
|
|
// If axis is negative, count from the right instead of left
|
|
|
|
if (axis < 0)
|
|
|
|
axis = rank + axis;
|
|
|
|
|
2024-02-06 06:23:46 +08:00
|
|
|
// We collapse in the dimensions to the right of the axis.
|
|
|
|
for (int i = axis + 1; i < rank; ++i) {
|
|
|
|
bool dynamic = shape[axis] == Torch::kUnknownSize ||
|
|
|
|
shape[i] == Torch::kUnknownSize;
|
|
|
|
if (dynamic) {
|
|
|
|
shape[axis] = Torch::kUnknownSize;
|
|
|
|
} else {
|
|
|
|
shape[axis] = shape[axis] * shape[i];
|
|
|
|
}
|
|
|
|
}
|
2024-01-20 08:18:16 +08:00
|
|
|
|
2024-02-06 06:23:46 +08:00
|
|
|
shape.resize(axis + 1, 1);
|
|
|
|
|
|
|
|
auto baseType = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
shape, operandTy.getDtype());
|
|
|
|
Value collapsedRight;
|
2024-01-20 08:18:16 +08:00
|
|
|
if (axis >= rank) {
|
|
|
|
// If the right range is empty, add a dim of size 1 to the
|
|
|
|
// right side of the shape:
|
|
|
|
// cr = torch.unsqueeze(x, x.ndim)
|
|
|
|
Value rankConst = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(rank));
|
|
|
|
collapsedRight = rewriter.create<Torch::AtenUnsqueezeOp>(
|
|
|
|
binder.getLoc(), baseType, operand, rankConst);
|
|
|
|
} else {
|
|
|
|
// Otherwise, collapse the right range into a single dimension:
|
|
|
|
// cr = torch._prims.collapse(x, axis, x.ndim - 1)
|
|
|
|
Value axisConst = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(axis));
|
|
|
|
Value rankLess1Const = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(rank - 1));
|
|
|
|
collapsedRight = rewriter.create<Torch::PrimsCollapseOp>(
|
|
|
|
binder.getLoc(), baseType, operand, axisConst, rankLess1Const);
|
|
|
|
}
|
|
|
|
|
|
|
|
Value zero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(0));
|
|
|
|
|
|
|
|
if (axis <= 0) {
|
|
|
|
// If the left range is empty, add a dim of size 1 to the
|
|
|
|
// left side of the shape:
|
|
|
|
// torch.unsqueeze(cr, 0)
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenUnsqueezeOp>(
|
|
|
|
binder.op, resultType, collapsedRight, zero);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, collapse the left range into a single dimension:
|
|
|
|
// torch._prims.collapse(cr, 0, axis - 1)
|
|
|
|
Value axisLess1Const = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getI64IntegerAttr(axis - 1));
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::PrimsCollapseOp>(
|
|
|
|
binder.op, resultType, collapsedRight, zero, axisLess1Const);
|
|
|
|
return success();
|
|
|
|
});
|
2023-12-05 13:55:51 +08:00
|
|
|
patterns.onOp("Floor", 13,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value operand;
|
|
|
|
if (binder.tensorOperand(operand) ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenFloorOp>(
|
|
|
|
binder.op, resultType, operand);
|
|
|
|
return success();
|
|
|
|
});
|
2024-01-27 01:36:39 +08:00
|
|
|
patterns.onOp(
|
2024-02-22 13:34:43 +08:00
|
|
|
"ConstantOfShape", 1,
|
2024-01-27 01:36:39 +08:00
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
Value shape;
|
|
|
|
if (binder.tensorOperand(shape) || binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
// convert shape tensor to list of ints
|
|
|
|
auto shapeSizes =
|
|
|
|
dyn_cast<Torch::ValueTensorType>(shape.getType()).getSizes();
|
|
|
|
SmallVector<Value> dimList;
|
|
|
|
Torch::BaseTensorType shapeType =
|
2024-04-28 05:00:56 +08:00
|
|
|
cast<Torch::BaseTensorType>(shape.getType());
|
2024-02-22 13:34:43 +08:00
|
|
|
Type selectResultType = rewriter.getType<Torch::ValueTensorType>(
|
|
|
|
ArrayRef<int64_t>({}), shapeType.getOptionalDtype());
|
2024-01-27 01:36:39 +08:00
|
|
|
Value zero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), 0));
|
2024-02-22 13:34:43 +08:00
|
|
|
|
2024-01-27 01:36:39 +08:00
|
|
|
for (int i = 0; i < shapeSizes[0]; i++) {
|
|
|
|
Value selectIndex = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(),
|
|
|
|
rewriter.getIntegerAttr(rewriter.getIntegerType(64), i));
|
|
|
|
Value extract = rewriter.create<Torch::AtenSelectIntOp>(
|
|
|
|
binder.getLoc(), selectResultType, shape, zero, selectIndex);
|
|
|
|
Value dim = rewriter.create<Torch::AtenItemOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::IntType>(), extract);
|
|
|
|
dimList.push_back(dim);
|
|
|
|
}
|
2024-02-22 13:34:43 +08:00
|
|
|
|
2024-01-27 01:36:39 +08:00
|
|
|
Value dimValueList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
Torch::ListType::get(Torch::IntType::get(binder.op->getContext())),
|
|
|
|
dimList);
|
|
|
|
Value noneVal = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
|
|
|
|
// Get fill_value if it is present.
|
|
|
|
// Assumption : resultDType and value attr type match.
|
|
|
|
auto attr = binder.op->getAttr("torch.onnx.value");
|
|
|
|
auto resultDType = resultType.getDtype();
|
|
|
|
|
|
|
|
// Extract the fill value and dtype
|
|
|
|
// ONNX requires value attr to be a tensor
|
|
|
|
if (!attr) {
|
2024-06-08 09:36:32 +08:00
|
|
|
attr =
|
|
|
|
DenseElementsAttr::get(resultType.toBuiltinTensor(),
|
|
|
|
rewriter.getFloatAttr(resultDType, 0.0));
|
2024-01-27 01:36:39 +08:00
|
|
|
}
|
2024-02-22 13:34:43 +08:00
|
|
|
|
|
|
|
// If its a dense resource attr we need to convert to a dense type:
|
|
|
|
if (DenseResourceElementsAttr rattr =
|
2024-04-11 21:47:35 +08:00
|
|
|
dyn_cast_or_null<DenseResourceElementsAttr>(attr)) {
|
2024-02-22 13:34:43 +08:00
|
|
|
// Bytes are stored in little endian order. Big endian support will
|
|
|
|
// require swizzling.
|
|
|
|
if (!Endian::little) {
|
|
|
|
binder.op->emitError(
|
|
|
|
"unimplemented: importing on big endian systems");
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
|
|
|
|
auto ty = cast<ShapedType>(rattr.getType());
|
|
|
|
auto ptr = rattr.getRawHandle().getBlob()->getData();
|
|
|
|
auto denseAttr = DenseElementsAttr::getFromRawBuffer(ty, ptr);
|
|
|
|
attr = dyn_cast_or_null<SplatElementsAttr>(denseAttr);
|
|
|
|
}
|
|
|
|
|
|
|
|
Attribute splattr;
|
|
|
|
if (isa<SplatElementsAttr>(attr)) {
|
2024-04-11 21:47:35 +08:00
|
|
|
auto denseAttr = cast<DenseElementsAttr>(attr);
|
2024-02-22 13:34:43 +08:00
|
|
|
splattr = denseAttr.getSplatValue<Attribute>();
|
2024-01-27 01:36:39 +08:00
|
|
|
}
|
|
|
|
|
2024-02-22 13:34:43 +08:00
|
|
|
if (!isa<FloatAttr, IntegerAttr>(splattr)) {
|
2024-01-27 01:36:39 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
binder.op,
|
|
|
|
"`value` attr tensor only supports types int and float for now.");
|
|
|
|
}
|
|
|
|
|
2024-02-22 13:34:43 +08:00
|
|
|
Value splatvalue;
|
|
|
|
if (auto intattr = dyn_cast<IntegerAttr>(splattr)) {
|
|
|
|
IntegerType intty = cast<IntegerType>(intattr.getType());
|
|
|
|
int64_t value;
|
|
|
|
if (intty.isUnsignedInteger()) {
|
|
|
|
value = intattr.getUInt();
|
|
|
|
} else if (intty.isSignedInteger()) {
|
|
|
|
value = intattr.getSInt();
|
|
|
|
} else {
|
|
|
|
value = intattr.getInt();
|
|
|
|
}
|
|
|
|
splatvalue =
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(binder.getLoc(), value);
|
2024-01-27 01:36:39 +08:00
|
|
|
}
|
|
|
|
|
2024-02-22 13:34:43 +08:00
|
|
|
if (auto fpattr = dyn_cast<FloatAttr>(splattr))
|
|
|
|
splatvalue = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
binder.getLoc(),
|
|
|
|
rewriter.getF64FloatAttr(fpattr.getValueAsDouble()));
|
|
|
|
|
2024-01-27 01:36:39 +08:00
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenFullOp>(
|
2024-02-22 13:34:43 +08:00
|
|
|
binder.op, resultType, dimValueList, splatvalue, /*dtype=*/noneVal,
|
2024-01-27 01:36:39 +08:00
|
|
|
/*layout=*/noneVal, /*device=*/noneVal, /*pin_memory=*/noneVal);
|
|
|
|
return success();
|
|
|
|
});
|
2024-04-09 01:08:01 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"Einsum", 12, [](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
SmallVector<Value> tensors;
|
|
|
|
std::string equation;
|
|
|
|
if (binder.tensorOperands(tensors, binder.op->getNumOperands()) ||
|
|
|
|
binder.customOpNameStringAttr(equation, "equation") ||
|
|
|
|
binder.tensorResultType(resultType))
|
|
|
|
return failure();
|
|
|
|
Type listElemType =
|
2024-05-24 00:01:47 +08:00
|
|
|
cast<Torch::BaseTensorType>(tensors[0].getType())
|
2024-04-09 01:08:01 +08:00
|
|
|
.getWithSizesAndDtype(/*optionalSizes=*/std::nullopt,
|
|
|
|
/*optionalDtype=*/nullptr);
|
|
|
|
Type listType = Torch::ListType::get(listElemType);
|
|
|
|
Value tensorList = rewriter.create<Torch::PrimListConstructOp>(
|
|
|
|
binder.op->getLoc(), listType, tensors);
|
|
|
|
Value cstEquation = rewriter.create<Torch::ConstantStrOp>(
|
|
|
|
binder.getLoc(), rewriter.getType<Torch::StringType>(),
|
|
|
|
rewriter.getStringAttr(equation));
|
|
|
|
Value cstNone = rewriter.create<Torch::ConstantNoneOp>(binder.getLoc());
|
|
|
|
rewriter.replaceOpWithNewOp<Torch::AtenEinsumOp>(
|
|
|
|
binder.op, resultType, cstEquation, tensorList, /*path=*/cstNone);
|
|
|
|
return success();
|
|
|
|
});
|
2024-05-01 00:21:27 +08:00
|
|
|
patterns.onOp(
|
|
|
|
"BlackmanWindow", 17,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Value size;
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
int64_t periodic, output_datatype;
|
|
|
|
if (binder.tensorOperand(size) ||
|
|
|
|
binder.s64IntegerAttr(output_datatype, "output_datatype", 1) ||
|
|
|
|
binder.s64IntegerAttr(periodic, "periodic", 1) ||
|
|
|
|
binder.tensorResultType(resultType)) {
|
|
|
|
return failure();
|
|
|
|
}
|
2024-05-07 01:21:45 +08:00
|
|
|
|
|
|
|
Location loc = binder.getLoc();
|
2024-05-01 00:21:27 +08:00
|
|
|
Value a0 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(0.42));
|
2024-05-01 00:21:27 +08:00
|
|
|
Value a1 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(-0.5));
|
2024-05-01 00:21:27 +08:00
|
|
|
Value a2 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(0.08));
|
2024-05-01 00:21:27 +08:00
|
|
|
|
2024-05-04 00:04:57 +08:00
|
|
|
auto windowFunctionResult =
|
|
|
|
windowFunctionImpl(binder, rewriter, size, a0, a1, a2, resultType,
|
|
|
|
output_datatype, periodic);
|
2024-05-01 00:21:27 +08:00
|
|
|
|
2024-05-04 00:04:57 +08:00
|
|
|
if (failed(windowFunctionResult))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
|
|
|
|
patterns.onOp(
|
|
|
|
"HannWindow", 17,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Value size;
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
int64_t periodic, output_datatype;
|
|
|
|
if (binder.tensorOperand(size) ||
|
|
|
|
binder.s64IntegerAttr(output_datatype, "output_datatype", 1) ||
|
|
|
|
binder.s64IntegerAttr(periodic, "periodic", 1) ||
|
|
|
|
binder.tensorResultType(resultType)) {
|
|
|
|
return failure();
|
2024-05-01 00:21:27 +08:00
|
|
|
}
|
2024-05-07 01:21:45 +08:00
|
|
|
|
|
|
|
Location loc = binder.getLoc();
|
2024-05-04 00:04:57 +08:00
|
|
|
Value a0 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(0.5));
|
2024-05-04 00:04:57 +08:00
|
|
|
Value a1 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(-0.5));
|
|
|
|
Value a2 = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(0.0));
|
|
|
|
|
|
|
|
auto windowFunctionResult =
|
|
|
|
windowFunctionImpl(binder, rewriter, size, a0, a1, a2, resultType,
|
|
|
|
output_datatype, periodic);
|
|
|
|
|
|
|
|
if (failed(windowFunctionResult))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
return success();
|
|
|
|
});
|
|
|
|
|
|
|
|
patterns.onOp(
|
|
|
|
"HammingWindow", 17,
|
|
|
|
[](OpBinder binder, ConversionPatternRewriter &rewriter) {
|
|
|
|
Value size;
|
|
|
|
Torch::ValueTensorType resultType;
|
|
|
|
int64_t periodic, output_datatype;
|
|
|
|
if (binder.tensorOperand(size) ||
|
|
|
|
binder.s64IntegerAttr(output_datatype, "output_datatype", 1) ||
|
|
|
|
binder.s64IntegerAttr(periodic, "periodic", 1) ||
|
|
|
|
binder.tensorResultType(resultType)) {
|
|
|
|
return failure();
|
|
|
|
}
|
|
|
|
|
|
|
|
Location loc = binder.getLoc();
|
|
|
|
Value a0 = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(0.543478));
|
|
|
|
Value a1 = rewriter.create<Torch::ConstantFloatOp>(
|
|
|
|
loc, rewriter.getF64FloatAttr(-0.456522));
|
2024-05-04 00:04:57 +08:00
|
|
|
Value a2 = rewriter.create<Torch::ConstantFloatOp>(
|
2024-05-07 01:21:45 +08:00
|
|
|
loc, rewriter.getF64FloatAttr(0.0));
|
2024-05-04 00:04:57 +08:00
|
|
|
|
|
|
|
auto windowFunctionResult =
|
|
|
|
windowFunctionImpl(binder, rewriter, size, a0, a1, a2, resultType,
|
|
|
|
output_datatype, periodic);
|
|
|
|
|
|
|
|
if (failed(windowFunctionResult))
|
|
|
|
return failure();
|
2024-05-01 00:21:27 +08:00
|
|
|
|
|
|
|
return success();
|
|
|
|
});
|
2023-11-22 13:02:55 +08:00
|
|
|
}
|