//===----------------------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // Also available under a BSD-style license. See LICENSE. // //===----------------------------------------------------------------------===// #include "torch-mlir/Conversion/TorchToStablehlo/TorchToStablehlo.h" #include "../PassDetail.h" #include "PopulatePatterns.h" #include "mlir/Dialect/Arith/IR/Arith.h" #include "mlir/Dialect/Tensor/IR/Tensor.h" #include "stablehlo/dialect/ChloOps.h" #include "stablehlo/dialect/StablehloOps.h" #include "torch-mlir/Conversion/TorchToStablehlo/StablehloLegalizeUtils.h" #include "torch-mlir/Conversion/Utils/Utils.h" #include "torch-mlir/Dialect/Torch/IR/TorchOps.h" #include "torch-mlir/Dialect/Torch/Utils/TorchUpstream.h" #include "torch-mlir/Dialect/Torch/Utils/Utils.h" #include #include using namespace mlir; using namespace mlir::torch; using namespace mlir::torch::Torch; using namespace mlir::torch::torch_to_stablehlo; static SmallVector getReduceOutputShape(ArrayRef inputShape, ArrayRef dims) { std::unordered_set dimsSet(dims.begin(), dims.end()); SmallVector reduceResultShape; for (size_t i = 0; i < inputShape.size(); i++) { if (dimsSet.find(i) == dimsSet.end()) { reduceResultShape.push_back(inputShape[i]); } } return reduceResultShape; } static Value createInitialValueForReduceOp(Operation *op, Type elementTy, PatternRewriter &rewriter) { auto constType = RankedTensorType::get({}, elementTy); if (isa(op)) { if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APFloat::getZero( cast(elementTy).getFloatSemantics(), /*negative=*/false)}); return rewriter.create(op->getLoc(), constType, constAttr); } else if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APInt::getZero(elementTy.getIntOrFloatBitWidth())}); return rewriter.create(op->getLoc(), constType, constAttr); } } if (isa(op)) { if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APFloat::getInf(cast(elementTy).getFloatSemantics(), /*negative=*/true)}); return rewriter.create(op->getLoc(), constType, constAttr); } else if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APInt::getSignedMinValue(elementTy.getIntOrFloatBitWidth())}); return rewriter.create(op->getLoc(), constType, constAttr); } } if (isa(op)) { if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APFloat::getInf(cast(elementTy).getFloatSemantics(), /*negative=*/false)}); return rewriter.create(op->getLoc(), constType, constAttr); } else if (isa(elementTy)) { auto constAttr = DenseElementsAttr::get( constType, {APInt::getSignedMaxValue(elementTy.getIntOrFloatBitWidth())}); return rewriter.create(op->getLoc(), constType, constAttr); } } if (isa(op)) { if (isa(elementTy)) { APFloat one(cast(elementTy).getFloatSemantics(), 1); auto constAttr = DenseElementsAttr::get(constType, one); return rewriter.create(op->getLoc(), constType, constAttr); } else if (isa(elementTy)) { APInt one(elementTy.getIntOrFloatBitWidth(), 1); auto constAttr = DenseElementsAttr::get(constType, one); return rewriter.create(op->getLoc(), constType, constAttr); } } if (isa(op)) { auto constAttr = DenseElementsAttr::get(constType, {APInt(/*numBits=*/1, 1)}); return rewriter.create(op->getLoc(), constType, constAttr); } if (isa(op)) { auto constAttr = DenseElementsAttr::get(constType, {APInt(/*numBits=*/1, 0)}); return rewriter.create(op->getLoc(), constType, constAttr); } op->emitError("unimplemented lowering in " "createInitialValueForReduceOp"); return nullptr; } static Value createReduceOpWithSingleRegionOp(Operation *op, Value input, Type outTy, ArrayRef dims, PatternRewriter &rewriter) { auto inputTy = dyn_cast(input.getType()); if (!inputTy) return nullptr; Value initValue = createInitialValueForReduceOp(op, inputTy.getElementType(), rewriter); if (!initValue) return nullptr; stablehlo::ReduceOp reduce = rewriter.create( op->getLoc(), outTy, input, initValue, rewriter.getDenseI64ArrayAttr(dims)); Block &block = reduce.getBody().emplaceBlock(); auto blockArgumentTy = RankedTensorType::get({}, inputTy.getElementType()); block.addArgument(blockArgumentTy, op->getLoc()); block.addArgument(blockArgumentTy, op->getLoc()); auto *firstArgument = block.args_begin(); auto secondArgument = block.args_rbegin(); { OpBuilder::InsertionGuard guard(rewriter); rewriter.setInsertionPointToStart(&block); Value result; if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else if (isa(op)) { result = rewriter.create( op->getLoc(), blockArgumentTy, *firstArgument, *secondArgument); } else { op->emitError("unimplemented lowering in " "createReduceOpWithSingleRegionOp"); return nullptr; } rewriter.create(op->getLoc(), result); } return reduce.getResults()[0]; } // Util for converting AtenMaxDimOp/AtenMinDimOp static std::optional createReduceOpReturnIndices(ConversionPatternRewriter &rewriter, Operation *op, Value &input, ArrayRef inputShapeVec, int64_t dim, size_t dimSizeIndexBits) { auto inputTy = cast(input.getType()); if (!inputTy) { return std::nullopt; } if (!inputTy.getElementType().isIntOrFloat()) { return std::nullopt; } auto inputShape = inputTy.getShape(); auto inputElemTy = inputTy.getElementType(); Value initValue = createInitialValueForReduceOp(op, inputElemTy, rewriter); if (!initValue) return std::nullopt; Value initIndex; if (dimSizeIndexBits == 32) { initIndex = hlo::getConstTensor(rewriter, op, {0}, {}).value(); } else { initIndex = hlo::getConstTensor(rewriter, op, {0}, {}).value(); } auto outputShape = getReduceOutputShape(inputShape, {dim}); auto outputTy = RankedTensorType::get(outputShape, inputElemTy); auto outputIndexTy = RankedTensorType::get(outputShape, rewriter.getIntegerType(64)); auto inputShapeTensor = rewriter.create( op->getLoc(), inputShapeVec); auto indexTensor = rewriter.create( op->getLoc(), RankedTensorType::get(inputShape, rewriter.getIntegerType(dimSizeIndexBits)), inputShapeTensor, static_cast(dim)); auto stablehloReduceOp = rewriter.create( op->getLoc(), TypeRange{outputTy, outputIndexTy}, ValueRange{input, indexTensor}, ValueRange{ initValue, initIndex, }, rewriter.getDenseI64ArrayAttr(dim)); Block &block = stablehloReduceOp.getBody().emplaceBlock(); // Add block arguments auto blockValArgumentType = RankedTensorType::get({}, inputTy.getElementType()); auto blockIdxArgumentType = RankedTensorType::get({}, rewriter.getIntegerType(dimSizeIndexBits)); auto compareResultType = RankedTensorType::get({}, rewriter.getI1Type()); block.addArgument(blockValArgumentType, op->getLoc()); block.addArgument(blockIdxArgumentType, op->getLoc()); block.addArgument(blockValArgumentType, op->getLoc()); block.addArgument(blockIdxArgumentType, op->getLoc()); auto *firstValArg = block.args_begin(); auto *firstIdxArg = std::next(firstValArg); auto *secondValArg = std::next(firstIdxArg); auto *secondIdxArg = std::next(secondValArg); stablehlo::ComparisonTypeAttr compareTypeAttr; if (isa(inputTy.getElementType())) { compareTypeAttr = stablehlo::ComparisonTypeAttr::get( rewriter.getContext(), stablehlo::ComparisonType::FLOAT); } else if (isa(inputTy.getElementType())) { compareTypeAttr = stablehlo::ComparisonTypeAttr::get( rewriter.getContext(), stablehlo::ComparisonType::SIGNED); } stablehlo::ComparisonDirectionAttr compareGeDirectionAttr = stablehlo::ComparisonDirectionAttr::get( rewriter.getContext(), stablehlo::ComparisonDirection::GE); stablehlo::ComparisonDirectionAttr compareLeDirectionAttr = stablehlo::ComparisonDirectionAttr::get( rewriter.getContext(), stablehlo::ComparisonDirection::LE); stablehlo::ComparisonDirectionAttr compareEqDirectionAttr = stablehlo::ComparisonDirectionAttr::get( rewriter.getContext(), stablehlo::ComparisonDirection::EQ); { OpBuilder::InsertionGuard guard(rewriter); rewriter.setInsertionPointToStart(&block); Value compareResult; if (isa(op)) { compareResult = rewriter.create( op->getLoc(), compareResultType, *firstValArg, *secondValArg, compareGeDirectionAttr, compareTypeAttr); } else if (isa(op)) { compareResult = rewriter.create( op->getLoc(), compareResultType, *firstValArg, *secondValArg, compareLeDirectionAttr, compareTypeAttr); } else { op->emitError("unimplement lowering of createReduceOpReturnIndices"); return std::nullopt; } Value retValResult = rewriter.create( op->getLoc(), compareResult, *firstValArg, *secondValArg); // get smaller index value if compared nums are equal. Value compareEqResult = rewriter.create( op->getLoc(), compareResultType, *firstValArg, *secondValArg, compareEqDirectionAttr, compareTypeAttr); Value minIdx = rewriter.create(op->getLoc(), *firstIdxArg, *secondIdxArg); Value idxWithGeVal = rewriter.create( op->getLoc(), compareResult, *firstIdxArg, *secondIdxArg); Value retIdxResult = rewriter.create( op->getLoc(), compareEqResult, minIdx, idxWithGeVal); rewriter.create( op->getLoc(), ValueRange{retValResult, retIdxResult}); } return stablehloReduceOp.getResults(); } static Value reshapeReduceResultWhenKeepDim(ConversionPatternRewriter &rewriter, Location loc, Value reduceResult, ArrayRef inputShapeVec, Type outType, ArrayRef dims) { SmallVector outShapeVec(inputShapeVec); Value one = rewriter.create( loc, rewriter.getIntegerAttr(rewriter.getIndexType(), 1)); for (auto dim : dims) { outShapeVec[dim] = one; } auto outShapeTensor = rewriter.create(loc, outShapeVec); return rewriter.create( loc, outType, reduceResult, outShapeTensor); } namespace { template class ConvertAtenReductionOp : public ConvertAtenOp { public: using ConvertAtenOp::ConvertAtenOp; using OpAdaptor = typename AtenOpT::Adaptor; LogicalResult matchAndRewrite(AtenOpT op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const override { assert(false && "Unimplemented"); return failure(); }; }; template class ConvertAtenReduceAllDimsOp : public ConvertAtenReductionOp { public: using ConvertAtenReductionOp::ConvertAtenReductionOp; using OpAdaptor = typename AtenOpT::Adaptor; LogicalResult matchAndRewrite(AtenOpT op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const override { Value input = adaptor.getSelf(); auto inputTy = dyn_cast(input.getType()); auto outTy = dyn_cast( ConvertAtenReductionOp::getTypeConverter()->convertType( op.getType())); if (!inputTy || !outTy) { return rewriter.notifyMatchFailure( op, "only Tensor types supported in StableHLO"); } auto inputElemTy = inputTy.getElementType(); if (!inputElemTy.isIntOrFloat()) { return op.emitError( "only floating-point or integer datatype legalization supported"); } if (inputElemTy != outTy.getElementType()) { // use output type as computation type input = rewriter.create(op->getLoc(), input, outTy.getElementType()); } SmallVector dims = llvm::to_vector(llvm::seq(0, inputTy.getRank())); Value result = createReduceOpWithSingleRegionOp(op, input, outTy, dims, rewriter); if (!result) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } rewriter.replaceOp(op, result); return success(); } }; template class ConvertAtenReduceOneDimOp : public ConvertAtenReductionOp { public: using ConvertAtenReductionOp::ConvertAtenReductionOp; using OpAdaptor = typename AtenOpT::Adaptor; LogicalResult matchAndRewrite(AtenOpT op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const override { Value input = adaptor.getSelf(); auto inputTy = dyn_cast(input.getType()); auto outTy = dyn_cast( ConvertAtenReductionOp::getTypeConverter()->convertType( op.getType())); if (!inputTy || !outTy) { return rewriter.notifyMatchFailure( op, "only Tensor types supported in StableHLO"); } auto inputElemTy = inputTy.getElementType(); if (!inputElemTy.isIntOrFloat()) { return op.emitError( "only floating-point or integer datatype legalization supported"); } if (inputElemTy != outTy.getElementType()) { // use output type as computation type input = rewriter.create(op->getLoc(), input, outTy.getElementType()); } bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure(op, "non-bool keepdim unsupported"); } int64_t dim; if (!matchPattern(op.getDim(), m_TorchConstantInt(&dim))) { return rewriter.notifyMatchFailure( op, "non-const integer `dim` is not supported"); } dim = toPositiveDim(dim, inputTy.getRank()); SmallVector reduceResultShape = getReduceOutputShape(inputTy.getShape(), {dim}); Value reduceResult = createReduceOpWithSingleRegionOp( op, input, RankedTensorType::get(reduceResultShape, outTy.getElementType()), {dim}, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } if (keepDim) { auto outShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(outShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } reduceResult = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), reduceResult, *outShapeInfo, outTy, {dim}); } rewriter.replaceOp(op, reduceResult); return success(); } }; template class ConvertAtenReduceDimsOp : public ConvertAtenReductionOp { public: using ConvertAtenReductionOp::ConvertAtenReductionOp; using OpAdaptor = typename AtenOpT::Adaptor; LogicalResult matchAndRewrite(AtenOpT op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const override { Value input = adaptor.getSelf(); auto inputTy = dyn_cast(input.getType()); auto outTy = dyn_cast( ConvertAtenReductionOp::getTypeConverter()->convertType( op.getType())); if (!inputTy || !outTy) { return rewriter.notifyMatchFailure( op, "only Tensor types supported in StableHLO"); } auto inputElemTy = inputTy.getElementType(); if (!inputElemTy.isIntOrFloat()) { return op.emitError( "only floating-point or integer datatype legalization supported"); } if (inputElemTy != outTy.getElementType()) { // use output type as computation type input = rewriter.create(op->getLoc(), input, outTy.getElementType()); } bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure(op, "non-bool keepdim unsupported"); } SmallVector inputDims; SmallVector dims; if (!matchPattern(op.getDim(), m_TorchListOfConstantInts(inputDims))) { return rewriter.notifyMatchFailure( op, "non-const integer `dim` is not supported"); } if (inputDims.size() == 0) { dims = llvm::to_vector(llvm::seq(0, inputTy.getRank())); } else { for (auto d : inputDims) { d = toPositiveDim(d, inputTy.getRank()); // Drop invalid dims if (isValidDim(d, inputTy.getRank())) { dims.push_back(d); } } llvm::sort(dims.begin(), dims.end()); } SmallVector reduceResultShape = getReduceOutputShape(inputTy.getShape(), dims); Value reduceResult = createReduceOpWithSingleRegionOp( op, input, RankedTensorType::get(reduceResultShape, outTy.getElementType()), dims, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } if (keepDim) { auto outShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(outShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } reduceResult = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), reduceResult, *outShapeInfo, outTy, dims); } rewriter.replaceOp(op, reduceResult); return success(); } }; template class ConvertAtenReduceWithIndicesOp : public ConvertAtenReductionOp { public: using ConvertAtenReductionOp::ConvertAtenReductionOp; using OpAdaptor = typename AtenOpT::Adaptor; LogicalResult matchAndRewrite(AtenOpT op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const override { Value input = adaptor.getSelf(); auto inputTy = dyn_cast(input.getType()); if (!inputTy) { return rewriter.notifyMatchFailure( op, "only Tensor types supported in StableHLO"); } auto inputElemTy = inputTy.getElementType(); if (!inputElemTy.isIntOrFloat()) { return op.emitError( "Only floating-point or integer datatype legalization supported"); } RankedTensorType valResultType = cast( ConvertAtenReductionOp::getTypeConverter()->convertType( op.getResult(0).getType())); RankedTensorType idxResultType = cast( ConvertAtenReductionOp::getTypeConverter()->convertType( op.getResult(1).getType())); Type idxElementType = idxResultType.getElementType(); if (!isa(idxElementType)) { return op.emitError("indices result should to be integer tyep"); } int64_t dim; if (!matchPattern(op.getDim(), m_TorchConstantInt(&dim))) { return rewriter.notifyMatchFailure(op, "non-int dim unsupported"); } dim = toPositiveDim(dim, inputTy.getRank()); if (!isValidDim(dim, inputTy.getRank())) { return rewriter.notifyMatchFailure(op, "dim is not a valid dim"); } bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure(op, "non-bool keepdim unsupported"); } const auto &options = ConvertAtenReductionOp::getOptions(); auto inputShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(inputShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } auto inputShapeVec = *inputShapeInfo; if (op.getResult(1).use_empty()) { llvm::SmallVector outputShape(inputTy.getShape()); outputShape.erase(outputShape.begin() + dim); Value reduceResult = createReduceOpWithSingleRegionOp( op, input, RankedTensorType::get(outputShape, inputElemTy), ArrayRef{dim}, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } if (keepDim) { reduceResult = reshapeReduceResultWhenKeepDim(rewriter, op->getLoc(), reduceResult, inputShapeVec, valResultType, {dim}); } rewriter.replaceOp(op, {reduceResult, Value()}); return success(); } else { ValueRange stablehloReduceResults = createReduceOpReturnIndices(rewriter, op, input, inputShapeVec, dim, options.dimSizeIndexBits) .value(); SmallVector reduceResults(stablehloReduceResults); if (keepDim) { reduceResults[0] = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), reduceResults[0], inputShapeVec, valResultType, {dim}); reduceResults[1] = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), reduceResults[1], inputShapeVec, idxResultType, {dim}); } rewriter.replaceOp(op, reduceResults); return success(); } }; }; } // namespace // AtenSumDimIntListOp namespace { template <> LogicalResult ConvertAtenReductionOp::matchAndRewrite( AtenSumDimIntListOp op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const { Value input = adaptor.getSelf(); auto inputTy = dyn_cast(input.getType()); auto outTy = dyn_cast(getTypeConverter()->convertType(op.getType())); if (!inputTy) { return rewriter.notifyMatchFailure( op, "only Tensor types supported in StableHLO"); } if (inputTy.getElementType() != outTy.getElementType()) { // Use output element type as computation type. auto dstElemTy = outTy.getElementType(); input = rewriter.create(op->getLoc(), input, dstElemTy); inputTy = dyn_cast(input.getType()); } auto inputElemTy = inputTy.getElementType(); if (!inputElemTy.isIntOrFloat()) { return op.emitError( "Only floating-point or integer datatype legalization supported"); } SmallVector inputDims; SmallVector dims; if (failed(checkNotNone(rewriter, op, op.getDim()))) { inputDims = llvm::to_vector<4>(llvm::seq(0, inputTy.getRank())); } else { if (!matchPattern(op.getDim(), m_TorchListOfConstantInts(inputDims))) { return rewriter.notifyMatchFailure( op, "non-const integer `dim` is not supported"); } if (inputDims.size() == 0) { inputDims = llvm::to_vector<4>(llvm::seq(0, inputTy.getRank())); } } for (auto d : inputDims) { d = toPositiveDim(d, inputTy.getRank()); // Drop invalid dims if (isValidDim(d, inputTy.getRank())) { dims.push_back(d); } } llvm::sort(dims.begin(), dims.end()); SmallVector reduceResultShape = getReduceOutputShape(inputTy.getShape(), dims); bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure(op, "non-bool keepdim unsupported"); } Value reduceResult = createReduceOpWithSingleRegionOp( op, input, RankedTensorType::get(reduceResultShape, outTy.getElementType()), dims, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } if (keepDim) { auto outShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(outShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } reduceResult = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), reduceResult, *outShapeInfo, outTy, dims); } rewriter.replaceOp(op, reduceResult); return success(); } } // namespace // AtenFrobeniusNormDimOp // aten.frobenius_norm.dim => stablehlo.reduce(calculate square sum along given // dims) + stablehlo.sqrt namespace { template <> LogicalResult ConvertAtenReductionOp::matchAndRewrite( AtenFrobeniusNormDimOp op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const { Value input = adaptor.getSelf(); auto inputType = dyn_cast(input.getType()); if (!inputType) { return op.emitError( "only ranked tensor input supported in AtenFrobeniusNormDimOp"); } auto inputRank = inputType.getRank(); auto inputElemType = inputType.getElementType(); if (!isa(inputElemType)) { return op.emitError( "only float dtype allowed in input tensor of AtenFrobeniusNormDimOp"); } SmallVector dims; if (!matchPattern(op.getDim(), m_TorchListOfConstantInts(dims))) { return rewriter.notifyMatchFailure( op, "non-const integer `dim` is not supported"); } for (auto &dim : dims) { dim = toPositiveDim(dim, inputRank); if (!isValidDim(dim, inputRank)) { return rewriter.notifyMatchFailure(op, "invalid dimension detected in `dim`"); } } // Sort the dims in ascending order, making the conversion // stable with unordered dims. std::sort(dims.begin(), dims.end()); SmallVector reduceResultShape = getReduceOutputShape(inputType.getShape(), dims); bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure( op, "non-const bool `keepdim` is not supported"); } auto squareOp = rewriter.create(op->getLoc(), input, input); Value reduceResult = createReduceOpWithSingleRegionOp( op, squareOp.getResult(), RankedTensorType::get(reduceResultShape, inputElemType), dims, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } Value output = rewriter.create(op->getLoc(), reduceResult); if (keepDim) { auto outShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(outShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } output = reshapeReduceResultWhenKeepDim( rewriter, op->getLoc(), output, *outShapeInfo, getTypeConverter()->convertType(op.getType()), dims); } rewriter.replaceOp(op, output); return success(); } } // namespace // AtenLinalgVectorNormOp namespace { template <> LogicalResult ConvertAtenReductionOp::matchAndRewrite( AtenLinalgVectorNormOp op, OpAdaptor adaptor, ConversionPatternRewriter &rewriter) const { Value input = adaptor.getSelf(); auto inputType = dyn_cast(input.getType()); if (!inputType) { return op.emitError( "only ranked tensor input supported in AtenLinalgVectorNormOp"); } int64_t inputRank = inputType.getRank(); auto outType = cast(getTypeConverter()->convertType(op.getType())); auto outElemType = outType.getElementType(); if (!isa(outElemType)) { return op.emitError("only float dtype allowed in AtenLinalgVectorNormOp"); } if (inputType.getElementType() != outType.getElementType()) { input = rewriter.create(op->getLoc(), input, outElemType); } Value ord = hlo::scalarToStablehloTensor(rewriter, op, adaptor.getOrd(), outElemType); SmallVector dims; if (failed(checkNotNone(rewriter, op, op.getDim()))) { dims = llvm::to_vector<4>(llvm::seq(0, inputRank)); } else { if (!matchPattern(op.getDim(), m_TorchListOfConstantInts(dims))) { return rewriter.notifyMatchFailure( op, "non-const integer `dim` is not supported"); } for (auto &dim : dims) { dim = toPositiveDim(dim, inputRank); if (!isValidDim(dim, inputRank)) { return rewriter.notifyMatchFailure( op, "invalid dimension detected in `dim`"); } } // Sort the dims in ascending order, making the conversion // stable with unordered dims. std::sort(dims.begin(), dims.end()); } SmallVector reduceResultShape = getReduceOutputShape(inputType.getShape(), dims); bool keepDim = false; if (!matchPattern(op.getKeepdim(), m_TorchConstantBool(&keepDim))) { return rewriter.notifyMatchFailure( op, "non-const bool `keepdim` is not supported"); } Value absValue = rewriter.create(op->getLoc(), input); Value powValue = rewriter.create(op->getLoc(), absValue, ord, nullptr); Value reduceResult = createReduceOpWithSingleRegionOp( op, powValue, RankedTensorType::get(reduceResultShape, outElemType), dims, rewriter); if (!reduceResult) { return op->emitError("createReduceOpWithSingleRegionOp return nullptr"); } auto scalarType = RankedTensorType::get({}, outElemType); auto constantOne = rewriter.create( op->getLoc(), scalarType, DenseElementsAttr::get( scalarType, APFloat(cast(outElemType).getFloatSemantics(), 1))); auto reciprocalOrd = rewriter.create( op->getLoc(), scalarType, constantOne, ord); Value output = rewriter.create( op->getLoc(), reduceResult, reciprocalOrd, nullptr); if (keepDim) { auto outShapeInfo = hlo::getDimIndexOfTensor(rewriter, op, input); if (failed(outShapeInfo)) { return rewriter.notifyMatchFailure( op, "failed to get dimension sizes of the input"); } output = reshapeReduceResultWhenKeepDim(rewriter, op->getLoc(), output, *outShapeInfo, outType, dims); } rewriter.replaceOp(op, output); return success(); } } // namespace void mlir::torch::torch_to_stablehlo::populateReductionOpPatternsAndLegality( TypeConverter &typeConverter, RewritePatternSet &patterns, ConversionTarget &target, const TorchToStablehloOptions &options) { MLIRContext *context = patterns.getContext(); #define INSERT_ATEN_REDUCTION_OP_PATTERN(AtenOp) \ target.addIllegalOp(); \ patterns.add>(typeConverter, context, options) INSERT_ATEN_REDUCTION_OP_PATTERN(AtenSumDimIntListOp); INSERT_ATEN_REDUCTION_OP_PATTERN(AtenFrobeniusNormDimOp); INSERT_ATEN_REDUCTION_OP_PATTERN(AtenLinalgVectorNormOp); #undef INSERT_ATEN_REDUCTION_OP_PATTERN #define INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenOp) \ target.addIllegalOp(); \ patterns.add>(typeConverter, context, \ options) INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenMaxOp); INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenMinOp); INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenSumOp); INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenProdOp); INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenAllOp); INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN(AtenAnyOp); #undef INSERT_ATEN_REDUCTION_ALL_DIMS_OP_PATTERN #define INSERT_ATEN_REDUCTION_ONE_DIM_OP_PATTERN(AtenOp) \ target.addIllegalOp(); \ patterns.add>(typeConverter, context, \ options) INSERT_ATEN_REDUCTION_ONE_DIM_OP_PATTERN(AtenAnyDimOp); INSERT_ATEN_REDUCTION_ONE_DIM_OP_PATTERN(AtenAllDimOp); #undef INSERT_ATEN_REDUCTION_ONE_DIM_OP_PATTERN #define INSERT_ATEN_REDUCTION_DIMS_OP_PATTERN(AtenOp) \ target.addIllegalOp(); \ patterns.add>(typeConverter, context, options) INSERT_ATEN_REDUCTION_DIMS_OP_PATTERN(AtenAmaxOp); INSERT_ATEN_REDUCTION_DIMS_OP_PATTERN(AtenAminOp); #undef INSERT_ATEN_REDUCTION_DIMS_OP_PATTERN #define INSERT_ATEN_REDUCTION_WITH_INDICES_PATTERN(AtenOp) \ target.addIllegalOp(); \ patterns.add>(typeConverter, context, \ options) INSERT_ATEN_REDUCTION_WITH_INDICES_PATTERN(AtenMaxDimOp); INSERT_ATEN_REDUCTION_WITH_INDICES_PATTERN(AtenMinDimOp); #undef INSERT_ATEN_REDUCTION_WITH_INDICES_PATTERN }