2022-03-03 00:48:15 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
|
|
// Also available under a BSD-style license. See LICENSE.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "torch-mlir/Conversion/TorchToTMTensor/TorchToTMTensor.h"
|
|
|
|
|
|
|
|
#include "../PassDetail.h"
|
2022-10-05 21:28:06 +08:00
|
|
|
#include "mlir/Dialect/Arith/IR/Arith.h"
|
2022-03-16 18:44:23 +08:00
|
|
|
#include "mlir/Dialect/Func/IR/FuncOps.h"
|
2022-03-03 00:48:15 +08:00
|
|
|
#include "mlir/Dialect/Linalg/IR/Linalg.h"
|
2023-05-17 03:17:45 +08:00
|
|
|
#include "mlir/Dialect/Math/IR/Math.h"
|
2022-04-27 03:27:51 +08:00
|
|
|
#include "mlir/Dialect/Tensor/IR/Tensor.h"
|
2022-09-09 12:36:40 +08:00
|
|
|
#include "mlir/IR/Builders.h"
|
2023-05-17 03:17:45 +08:00
|
|
|
#include "mlir/IR/BuiltinTypeInterfaces.h"
|
2022-09-09 12:36:40 +08:00
|
|
|
#include "mlir/IR/BuiltinTypes.h"
|
2022-03-03 00:48:15 +08:00
|
|
|
#include "mlir/IR/MLIRContext.h"
|
2023-05-17 03:17:45 +08:00
|
|
|
#include "mlir/IR/Matchers.h"
|
2023-02-22 07:05:55 +08:00
|
|
|
#include "mlir/IR/ValueRange.h"
|
2022-03-03 00:48:15 +08:00
|
|
|
#include "torch-mlir-dialects/Dialect/TMTensor/IR/TMTensorDialect.h"
|
|
|
|
#include "torch-mlir-dialects/Dialect/TMTensor/IR/TMTensorOps.h"
|
|
|
|
#include "torch-mlir/Conversion/Utils/Utils.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchDialect.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchOps.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchTypes.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/Utils/TorchUpstream.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/Utils/Utils.h"
|
|
|
|
#include "torch-mlir/Dialect/TorchConversion/Transforms/BackendTypeConversion.h"
|
2023-02-22 07:05:55 +08:00
|
|
|
#include "llvm/ADT/APFloat.h"
|
|
|
|
#include "llvm/ADT/APInt.h"
|
|
|
|
#include "llvm/Support/ErrorHandling.h"
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
using namespace mlir;
|
|
|
|
using namespace mlir::torch;
|
|
|
|
using namespace mlir::torch::Torch;
|
|
|
|
using namespace mlir::torch::TorchConversion;
|
|
|
|
using namespace mlir::torch::TMTensor;
|
|
|
|
|
|
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// Patterns (as this grows, it should be organized into multiple files)
|
|
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// This is going to eventually be O(#aten ops), which is in the 100s.
|
|
|
|
//
|
|
|
|
// Most of these patterns consist of:
|
|
|
|
// 1. Checking that the operand/result types and other static properties are
|
|
|
|
// good-enough to create a valid linalg op (such as operands being of
|
|
|
|
// ranks/dtypes acceptable to the linalg op).
|
|
|
|
// 2. Creating dynamic error guards, usually checking a predicate on the
|
|
|
|
// compatibility of operand shapes.
|
|
|
|
// 3. Creating init tensors for the computation op. Usually this involves
|
|
|
|
// reifying IR for a shape transfer function based on the operand shapes.
|
|
|
|
// 4. Creating a named linalg op to replace the original op.
|
|
|
|
//
|
|
|
|
// TODO: Use linalg OpDSL to autogenerate at least 1)/2)/3) such
|
|
|
|
// that these patterns become mostly mechanical associations of
|
|
|
|
// "aten.foo -> linalg.foo".
|
|
|
|
|
2023-04-25 23:52:46 +08:00
|
|
|
static TypedAttr getNumericLimit(PatternRewriter &rewriter, Type elementType,
|
2023-02-22 07:05:55 +08:00
|
|
|
bool getMin = true) {
|
|
|
|
auto bitWidth = elementType.getIntOrFloatBitWidth();
|
|
|
|
if (llvm::isa<mlir::IntegerType>(elementType)) {
|
|
|
|
if (getMin) {
|
|
|
|
return rewriter.getIntegerAttr(elementType,
|
|
|
|
APInt::getSignedMinValue(bitWidth));
|
|
|
|
} else {
|
|
|
|
return rewriter.getIntegerAttr(elementType,
|
|
|
|
APInt::getSignedMaxValue(bitWidth));
|
|
|
|
}
|
|
|
|
} else if (mlir::FloatType floatType =
|
|
|
|
llvm::dyn_cast<mlir::FloatType>(elementType)) {
|
|
|
|
return rewriter.getFloatAttr(
|
|
|
|
elementType,
|
|
|
|
APFloat::getLargest(floatType.getFloatSemantics(), getMin));
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only float/integer types are supported!");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// This function will reformat the `index` and `src` from torch operations
|
|
|
|
// like `torch.scatter` or `torch.scatter_reduce` to match the expected
|
|
|
|
// input for the TMScatterOp. It will return the reformated `index` and `src`
|
|
|
|
// as a pair of mlir::Value that can be used as inputs for the TMScatterOp.
|
|
|
|
static std::pair<Value, Value>
|
|
|
|
convertTorchScatterIndexAndSrcToTMScatterIndexAndSrc(PatternRewriter &rewriter,
|
|
|
|
Value indices, Value src,
|
|
|
|
int64_t dim) {
|
|
|
|
// Get information on types for inputs
|
|
|
|
RankedTensorType indexType = indices.getType().cast<RankedTensorType>();
|
|
|
|
RankedTensorType srcSelf = src.getType().cast<RankedTensorType>();
|
|
|
|
|
|
|
|
// Store location for insertions
|
|
|
|
Location loc = src.getLoc();
|
|
|
|
|
|
|
|
Value indexSize = getTensorSize(rewriter, loc, indices);
|
|
|
|
indexSize = castIntToIndex(rewriter, loc, indexSize);
|
|
|
|
SmallVector<Value> indexShape = getTensorSizes(rewriter, loc, indices);
|
|
|
|
Value cstOne = rewriter.create<arith::ConstantIndexOp>(loc, 1);
|
|
|
|
|
|
|
|
// We flatten the `src` values from (i, j, k, ...) -> (i * j * k * ...)
|
|
|
|
SmallVector<Value> indSliceShape({indexSize, cstOne});
|
|
|
|
Value indSlice =
|
|
|
|
createZeroInitTensor(rewriter, loc, indSliceShape, rewriter.getI32Type());
|
|
|
|
|
|
|
|
// New output shape will be equal to the product of the dimensions of the
|
|
|
|
// updates
|
|
|
|
SmallVector<Value> outputs(indexType.getRank(), indSlice);
|
|
|
|
outputs.push_back(createZeroInitTensor(rewriter, loc, {indexSize},
|
|
|
|
srcSelf.getElementType()));
|
|
|
|
SmallVector<Type> outputsType(indexType.getRank(), indSlice.getType());
|
|
|
|
outputsType.push_back(outputs[indexType.getRank()].getType());
|
|
|
|
|
|
|
|
// Create mapping over flattened iteration space
|
|
|
|
SmallVector<AffineExpr> indSliceExpr = {rewriter.getAffineDimExpr(0),
|
|
|
|
rewriter.getAffineConstantExpr(0)};
|
|
|
|
SmallVector<AffineMap> mapping(
|
|
|
|
indexType.getRank(), AffineMap::get(/*dimCount=*/1, /*symbolCount=*/0,
|
|
|
|
indSliceExpr, src.getContext()));
|
|
|
|
// Mapping for updates
|
|
|
|
mapping.push_back(rewriter.getDimIdentityMap());
|
|
|
|
SmallVector<utils::IteratorType> iteratorTypes(
|
|
|
|
{utils::IteratorType::parallel});
|
|
|
|
|
|
|
|
// This function goes over the flattened iteration space of the `indices`
|
|
|
|
// and `src`. It will reconstruct the original induction variables based
|
|
|
|
// on the current flattened index. The flattened iteration space is required
|
|
|
|
// because TMTensorScatterOp expects a list of single element updates.
|
|
|
|
auto flattenedUpdates =
|
|
|
|
rewriter
|
|
|
|
.create<linalg::GenericOp>(
|
|
|
|
loc, outputsType, ValueRange(), outputs, mapping, iteratorTypes,
|
|
|
|
[&](OpBuilder &b, Location loc, ValueRange args) {
|
|
|
|
SmallVector<Value> indexValues(indexType.getRank());
|
|
|
|
Value ind = b.create<linalg::IndexOp>(loc, 0);
|
|
|
|
for (int i = indexType.getRank() - 1; i >= 0; i--) {
|
|
|
|
indexValues[i] =
|
|
|
|
b.create<arith::RemSIOp>(loc, ind, indexShape[i]);
|
|
|
|
ind = b.create<arith::DivSIOp>(loc, ind, indexShape[i]);
|
|
|
|
}
|
|
|
|
// Extract the scatter index and update value
|
|
|
|
Value extractIndexValue =
|
|
|
|
b.create<tensor::ExtractOp>(loc, indices, indexValues);
|
|
|
|
Value extractSrcValue =
|
|
|
|
b.create<tensor::ExtractOp>(loc, src, indexValues);
|
|
|
|
SmallVector<Value> yieldVals;
|
|
|
|
for (Value v : indexValues) {
|
|
|
|
Value scalar = castIndexToInt64(b, loc, v);
|
|
|
|
yieldVals.push_back(b.create<arith::TruncIOp>(
|
|
|
|
loc, rewriter.getI32Type(), scalar));
|
|
|
|
}
|
|
|
|
// Replace the original index with the index specified
|
|
|
|
// by the scatter.
|
|
|
|
yieldVals[dim] = b.create<arith::TruncIOp>(
|
|
|
|
loc, rewriter.getI32Type(), extractIndexValue);
|
|
|
|
yieldVals.push_back(extractSrcValue);
|
|
|
|
b.create<linalg::YieldOp>(loc, yieldVals);
|
|
|
|
})
|
|
|
|
.getResultTensors();
|
|
|
|
|
|
|
|
auto toOpFoldResult = [](Value v) -> OpFoldResult {
|
|
|
|
auto op = v.getDefiningOp<arith::ConstantIndexOp>();
|
|
|
|
if (!op)
|
|
|
|
return v;
|
|
|
|
return op.getValue();
|
|
|
|
};
|
|
|
|
|
|
|
|
// The result of the linalg::Generic operation gives us (rank(`src`) + 1)
|
|
|
|
// 1D-tensors where each contains a number of elements equal to the total
|
|
|
|
// number of elements in the `src` tensor. The indices must now be
|
|
|
|
// constructed by concatanating the first rank(`src`) tensors together. The
|
|
|
|
// new `src` tensor is the last tensor returned from the linalg::Generic
|
|
|
|
// operation.
|
|
|
|
SmallVector<Value> offsets = {
|
|
|
|
rewriter.create<arith::ConstantIndexOp>(loc, 0),
|
|
|
|
rewriter.create<arith::ConstantIndexOp>(loc, 0)};
|
|
|
|
SmallVector<Value> strides = {
|
|
|
|
rewriter.create<arith::ConstantIndexOp>(loc, 1),
|
|
|
|
rewriter.create<arith::ConstantIndexOp>(loc, 1)};
|
|
|
|
Value indicesRank =
|
|
|
|
rewriter.create<arith::ConstantIndexOp>(loc, indexType.getRank());
|
|
|
|
Value flattenedIndices = createZeroInitTensor(
|
|
|
|
rewriter, loc, SmallVector<Value>({indexSize, indicesRank}),
|
|
|
|
rewriter.getI32Type());
|
|
|
|
SmallVector<Value> scatterInputsVector(flattenedUpdates);
|
|
|
|
for (auto const slice : ArrayRef(scatterInputsVector).drop_back()) {
|
|
|
|
SmallVector<Value> sizes = getTensorSizes(rewriter, loc, slice);
|
|
|
|
flattenedIndices = rewriter.createOrFold<tensor::InsertSliceOp>(
|
|
|
|
loc, slice, flattenedIndices,
|
|
|
|
llvm::to_vector(llvm::map_range(offsets, toOpFoldResult)),
|
|
|
|
llvm::to_vector(llvm::map_range(sizes, toOpFoldResult)),
|
|
|
|
llvm::to_vector(llvm::map_range(strides, toOpFoldResult)));
|
|
|
|
// Increment offset to insert into next column
|
|
|
|
offsets[1] = rewriter.createOrFold<arith::AddIOp>(loc, offsets[1], cstOne);
|
|
|
|
}
|
|
|
|
|
|
|
|
return std::make_pair(flattenedIndices,
|
|
|
|
scatterInputsVector[indexType.getRank()]);
|
|
|
|
}
|
|
|
|
|
2022-04-01 00:27:21 +08:00
|
|
|
static Value createTMTensorScatterOp(
|
|
|
|
OpBuilder &b, Location loc, Value updates, Value indices, Value original,
|
|
|
|
bool uniqueIndices,
|
|
|
|
function_ref<void(OpBuilder &, Location, Value, Value)> bodyBuild) {
|
|
|
|
auto originalTensorType = original.getType().cast<RankedTensorType>();
|
|
|
|
Type originalElementType = originalTensorType.getElementType();
|
|
|
|
auto scatterOp = b.create<TMTensor::ScatterOp>(
|
|
|
|
loc, originalTensorType, ValueRange{updates, indices},
|
|
|
|
ValueRange{original}, uniqueIndices);
|
|
|
|
|
2022-12-08 04:20:41 +08:00
|
|
|
Region &scatterOpRegion = scatterOp.getRegion();
|
2022-04-01 00:27:21 +08:00
|
|
|
auto &scatterOpBlock = scatterOpRegion.emplaceBlock();
|
|
|
|
scatterOpBlock.addArguments({originalElementType, originalElementType},
|
|
|
|
{loc, loc});
|
|
|
|
OpBuilder regionBuilder(scatterOpRegion);
|
|
|
|
auto blockArgs = scatterOpBlock.getArguments();
|
|
|
|
Value updatesElement = blockArgs[0];
|
|
|
|
Value originalElement = blockArgs[1];
|
|
|
|
bodyBuild(regionBuilder, loc, updatesElement, originalElement);
|
|
|
|
return scatterOp->getResult(0);
|
|
|
|
}
|
|
|
|
|
2022-09-09 12:36:40 +08:00
|
|
|
static Value createTMTensorScanOp(
|
|
|
|
OpBuilder &b, Location loc, Value input, Value output, Value accumulator,
|
|
|
|
int64_t dim, bool inclusive,
|
|
|
|
function_ref<void(OpBuilder &, Location, Value, Value)> bodyBuild) {
|
|
|
|
auto inputType = input.getType().cast<RankedTensorType>();
|
|
|
|
auto accType = accumulator.getType().cast<RankedTensorType>();
|
|
|
|
Type elementType = inputType.getElementType();
|
|
|
|
auto scanOp = b.create<TMTensor::ScanOp>(
|
|
|
|
loc, TypeRange{inputType, accType}, input,
|
|
|
|
ValueRange{output, accumulator}, b.getI64IntegerAttr(dim),
|
|
|
|
b.getBoolAttr(inclusive));
|
|
|
|
|
2022-12-08 04:20:41 +08:00
|
|
|
Region &scanOpRegion = scanOp.getRegion();
|
2022-09-09 12:36:40 +08:00
|
|
|
auto &scanOpBlock = scanOpRegion.emplaceBlock();
|
|
|
|
scanOpBlock.addArguments({elementType, elementType}, {loc, loc});
|
|
|
|
OpBuilder regionBuilder(scanOpRegion);
|
|
|
|
auto blockArgs = scanOpBlock.getArguments();
|
|
|
|
Value inputElement = blockArgs[0];
|
|
|
|
Value accElement = blockArgs[1];
|
|
|
|
bodyBuild(regionBuilder, loc, inputElement, accElement);
|
|
|
|
return scanOp->getResult(0);
|
|
|
|
}
|
|
|
|
|
2023-04-03 18:49:01 +08:00
|
|
|
// Utility function to create a TMTensor::SortOp.
|
|
|
|
static FailureOr<SmallVector<Value>>
|
|
|
|
createTMTensorSortOp(PatternRewriter &rewriter, Location sortOpLoc,
|
|
|
|
llvm::ArrayRef<Value> operands,
|
|
|
|
llvm::ArrayRef<Type> elementTypes, int64_t dimension,
|
|
|
|
bool isStable, bool isDescending) {
|
|
|
|
// Step 1. Create TMTensor::SortOp structure.
|
|
|
|
SmallVector<Type> sortResultTypes;
|
|
|
|
for (Value val : operands) {
|
|
|
|
sortResultTypes.push_back(val.getType());
|
|
|
|
}
|
|
|
|
ValueRange inputs;
|
|
|
|
auto sortOp = rewriter.create<TMTensor::SortOp>(
|
|
|
|
sortOpLoc, sortResultTypes, inputs, operands,
|
|
|
|
rewriter.getI64IntegerAttr(dimension));
|
|
|
|
|
|
|
|
// Step 2. Add two arguments for each element type in the SortOp's block.
|
|
|
|
Region *body = &sortOp.getRegion();
|
|
|
|
Block *block = rewriter.createBlock(body);
|
|
|
|
Location loc = body->getLoc();
|
|
|
|
for (Type elementType : elementTypes) {
|
|
|
|
block->addArguments({elementType, elementType},
|
|
|
|
SmallVector<Location, 2>(2, loc));
|
|
|
|
}
|
|
|
|
|
|
|
|
// Step 3. Create comparison op which will be used as the sorting predicate.
|
|
|
|
Value compareOp;
|
|
|
|
if (auto intType = elementTypes[0].dyn_cast<mlir::IntegerType>()) {
|
|
|
|
// Case for using arith::CmpIOp.
|
|
|
|
arith::CmpIPredicate ge = arith::CmpIPredicate::sge;
|
|
|
|
arith::CmpIPredicate le = arith::CmpIPredicate::sle;
|
|
|
|
if (intType.isUnsignedInteger()) {
|
|
|
|
ge = arith::CmpIPredicate::uge;
|
|
|
|
le = arith::CmpIPredicate::ule;
|
|
|
|
}
|
|
|
|
arith::CmpIPredicate predicate = isDescending ? ge : le;
|
|
|
|
compareOp = rewriter.create<arith::CmpIOp>(
|
|
|
|
loc, predicate, block->getArgument(0), block->getArgument(1));
|
|
|
|
} else if (elementTypes[0].isa<mlir::FloatType>()) {
|
|
|
|
// Case for using arith::CmpFOp.
|
|
|
|
arith::CmpFPredicate predicate =
|
|
|
|
isDescending ? arith::CmpFPredicate::OGE : arith::CmpFPredicate::OLE;
|
|
|
|
compareOp = rewriter.create<arith::CmpFOp>(
|
|
|
|
loc, predicate, block->getArgument(0), block->getArgument(1));
|
|
|
|
} else {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
sortOpLoc, "Only Integer and Floating element type expected.");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Step 4. Create yield op for yielding the sorting predicate.
|
|
|
|
rewriter.create<TMTensor::YieldOp>(loc, compareOp);
|
|
|
|
return SmallVector<Value>(sortOp.getResults());
|
|
|
|
}
|
|
|
|
|
2022-03-03 00:48:15 +08:00
|
|
|
namespace {
|
|
|
|
// aten::bincount op counts the frequency of each value in a 1-d input tensor of
|
|
|
|
// non-negative ints.
|
|
|
|
class ConvertAtenBincountOp : public OpConversionPattern<AtenBincountOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenBincountOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
if (failed(verifyLinalgCompatibleTypes(op, rewriter)))
|
|
|
|
return failure();
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
MLIRContext *context = op->getContext();
|
|
|
|
TypeConverter *typeConverter = getTypeConverter();
|
2022-12-08 04:20:41 +08:00
|
|
|
Value input = adaptor.getSelf();
|
|
|
|
Value torchTypeInput = op.getSelf();
|
|
|
|
Value minlength = adaptor.getMinlength();
|
|
|
|
Value weights = adaptor.getWeights();
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
// TODO: Add a check to verify that the input tensor elements are all
|
|
|
|
// non-negative.
|
|
|
|
// Check whether the input is a 1-d tensor of integer type or not.
|
|
|
|
RankedTensorType inputType = input.getType().cast<RankedTensorType>();
|
|
|
|
if (inputType.getRank() != 1 ||
|
|
|
|
!inputType.getElementType().isa<mlir::IntegerType>())
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op,
|
|
|
|
"Input tensor has to be a one-dimensional tensor of integer type.");
|
|
|
|
|
|
|
|
// Check whether the input tensor element type is i64 or not.
|
|
|
|
IntegerType inputIntegerType =
|
|
|
|
inputType.getElementType().cast<IntegerType>();
|
|
|
|
if (inputIntegerType.getWidth() != 64)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op,
|
|
|
|
"Unimplemented: Integer width not equal to 64 are not supported.");
|
|
|
|
|
|
|
|
// TODO: Incorporate the weight argument.
|
|
|
|
if (!weights.getType().isa<mlir::torch::Torch::NoneType>())
|
|
|
|
return rewriter.notifyMatchFailure(
|
2022-04-01 00:27:21 +08:00
|
|
|
op, "Unimplemented: the weights operand is not incorporated.");
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
// Finding the maximum value in the input tensor.
|
|
|
|
SmallVector<int64_t> maxTensorSizes;
|
|
|
|
ValueTensorType maxTensorType = ValueTensorType::get(
|
2023-01-25 09:29:42 +08:00
|
|
|
context, llvm::ArrayRef(maxTensorSizes),
|
2022-03-03 00:48:15 +08:00
|
|
|
torchTypeInput.getType().cast<ValueTensorType>().getDtype());
|
|
|
|
Value maxTensor =
|
|
|
|
rewriter.create<AtenMaxOp>(loc, maxTensorType, torchTypeInput);
|
|
|
|
maxTensor = typeConverter->materializeTargetConversion(
|
|
|
|
rewriter, loc, typeConverter->convertType(maxTensor.getType()),
|
|
|
|
maxTensor);
|
|
|
|
|
|
|
|
// `maxTensor` is a 0-d tensor, extracting its only element and
|
|
|
|
// storing it in `maxInput`.
|
|
|
|
Value maxInput = rewriter.create<tensor::ExtractOp>(loc, maxTensor);
|
|
|
|
|
|
|
|
// Creating a tm_tensor.scatter op with the following mapping:
|
|
|
|
// 1.) `input` tensor maps to the indices in scatter op. `input` is
|
|
|
|
// expanded from 1-d to 2-d, and its element type is set to i32 as required
|
|
|
|
// for the scatter op.
|
|
|
|
// 2.) `updates` is a 1-d dummy tensor with the size equivalent to the
|
|
|
|
// `input`.
|
|
|
|
// 3.) `bincount` a 1-d tensor maps to the original in scatter op
|
|
|
|
// with size equal to the max(max(input) + 1, minlength).
|
2022-11-29 20:33:31 +08:00
|
|
|
SmallVector<int64_t> expandedInputSizes{
|
|
|
|
makeShapeTorchCompatible(inputType.getShape())[0], 1};
|
2022-03-03 00:48:15 +08:00
|
|
|
ValueTensorType expandInputType = ValueTensorType::get(
|
2023-01-25 09:29:42 +08:00
|
|
|
context, llvm::ArrayRef(expandedInputSizes),
|
2022-03-03 00:48:15 +08:00
|
|
|
torchTypeInput.getType().cast<ValueTensorType>().getDtype());
|
|
|
|
Value torchCstOne = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(1));
|
|
|
|
Value expandedInputTensor = rewriter.create<AtenUnsqueezeOp>(
|
|
|
|
loc, expandInputType, torchTypeInput, torchCstOne);
|
|
|
|
|
|
|
|
// Converting the input element type to i32.
|
|
|
|
Value indices = convertTensorToDtype(
|
|
|
|
rewriter, loc, expandedInputTensor,
|
|
|
|
mlir::IntegerType::get(context, 32, mlir::IntegerType::Signed));
|
|
|
|
indices = typeConverter->materializeTargetConversion(
|
|
|
|
rewriter, loc, typeConverter->convertType(indices.getType()), indices);
|
|
|
|
|
2022-04-01 00:27:21 +08:00
|
|
|
auto resultType = typeConverter->convertType(op->getResult(0).getType())
|
|
|
|
.cast<RankedTensorType>();
|
|
|
|
Type resultElemType = resultType.getElementType();
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
SmallVector<Value, 1> inputSizeDynamic =
|
|
|
|
getTensorSizesUntilDim(rewriter, loc, input, 0);
|
2022-10-18 12:22:53 +08:00
|
|
|
Value updatesTensor = rewriter.create<tensor::EmptyOp>(
|
2022-03-03 00:48:15 +08:00
|
|
|
loc, getAsOpFoldResult(inputSizeDynamic), resultElemType);
|
|
|
|
|
|
|
|
Value constantZero = rewriter.create<arith::ConstantOp>(
|
|
|
|
loc, rewriter.getZeroAttr(resultElemType));
|
|
|
|
Value constantOne = rewriter.create<arith::ConstantIntOp>(
|
|
|
|
loc, 1, resultElemType.getIntOrFloatBitWidth());
|
|
|
|
|
|
|
|
// Bincount size = max(max(input) + 1, minlength)
|
|
|
|
Value maxInputPlusOne =
|
|
|
|
rewriter.create<arith::AddIOp>(loc, maxInput, constantOne);
|
|
|
|
Value bincountSize =
|
|
|
|
rewriter.create<arith::MaxSIOp>(loc, maxInputPlusOne, minlength);
|
|
|
|
bincountSize = castIntToIndex(rewriter, loc, bincountSize);
|
|
|
|
Value bincountTensor = createInitTensor(rewriter, loc, {bincountSize},
|
|
|
|
resultElemType, constantZero);
|
|
|
|
|
2022-04-01 00:27:21 +08:00
|
|
|
Value scatterOp = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, updatesTensor, indices, bincountTensor,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value _, Value bincountElem) {
|
|
|
|
Value add = b.create<arith::AddIOp>(loc, bincountElem, constantOne);
|
|
|
|
b.create<TMTensor::YieldOp>(loc, add);
|
|
|
|
});
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::CastOp>(op, resultType, scatterOp);
|
2022-03-03 00:48:15 +08:00
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
// """Create a map from each dimension of the input tensor to the
|
|
|
|
// subspace that dimension corresponds to in the result shape one gets
|
|
|
|
// from indexing the tensor with the optional index tensors.
|
|
|
|
//
|
|
|
|
// Note: Index tensors are first broadcasted to a common shape before
|
|
|
|
// creating the mapping. So the index of every index tensor will map to
|
|
|
|
// the same dimensions in the result shape.
|
|
|
|
//
|
|
|
|
// For example:
|
|
|
|
// indices = [None, None, torch.randint(4, (6, 1)), torch.randint(5, (7,))]
|
|
|
|
// indexBroadcastShapeValue = [6, 7]
|
|
|
|
// map = {0: [0], 1: [1], 2: [2, 3], 3: [2, 3]}
|
|
|
|
static SmallVector<SmallVector<int64_t>>
|
|
|
|
getInputShapeToOutputShapeMap(SmallVector<Value> optionalIndices,
|
|
|
|
SmallVector<Value> indexBroadcastShapeValue) {
|
|
|
|
SmallVector<Value> indices;
|
|
|
|
for (Value index : optionalIndices) {
|
|
|
|
if (!index.getType().isa<Torch::NoneType>())
|
|
|
|
indices.push_back(index);
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned broadcastRank = indexBroadcastShapeValue.size();
|
|
|
|
unsigned numIndexTensors = indices.size();
|
|
|
|
int64_t indexOfFirstIndexTensor = -1;
|
|
|
|
SmallVector<SmallVector<int64_t>> result;
|
|
|
|
|
|
|
|
for (unsigned i = 0; i < optionalIndices.size(); i++) {
|
|
|
|
if (optionalIndices[i].getType().isa<Torch::NoneType>()) {
|
|
|
|
unsigned val = i;
|
|
|
|
if (indexOfFirstIndexTensor >= 0)
|
|
|
|
val += broadcastRank - numIndexTensors;
|
|
|
|
result.push_back({val});
|
|
|
|
} else {
|
|
|
|
if (indexOfFirstIndexTensor < 0)
|
|
|
|
indexOfFirstIndexTensor = i;
|
|
|
|
SmallVector<int64_t> outputIndices;
|
|
|
|
for (unsigned j = indexOfFirstIndexTensor;
|
|
|
|
j < (indexOfFirstIndexTensor + broadcastRank); j++)
|
|
|
|
outputIndices.push_back(j);
|
|
|
|
result.push_back(outputIndices);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
static std::tuple<SmallVector<Value>, SmallVector<int64_t>>
|
|
|
|
getIndicesFinalShape(ConversionPatternRewriter &rewriter, Location loc,
|
|
|
|
Value input, SmallVector<Value> optionalIndices,
|
|
|
|
SmallVector<int64_t> inputShapeInt,
|
|
|
|
SmallVector<Value> inputShapeValue,
|
|
|
|
SmallVector<int64_t> indexBroadcastShapeInt,
|
|
|
|
SmallVector<Value> indexBroadcastShapeValue) {
|
|
|
|
SmallVector<Value> result;
|
|
|
|
SmallVector<int64_t> resultInt;
|
|
|
|
bool handledIndexTensorSpace = false;
|
|
|
|
|
|
|
|
for (unsigned i = 0; i < inputShapeValue.size(); i++) {
|
|
|
|
if (optionalIndices[i].getType().isa<Torch::NoneType>()) {
|
|
|
|
result.push_back(inputShapeValue[i]);
|
|
|
|
resultInt.push_back(inputShapeInt[i]);
|
|
|
|
} else {
|
|
|
|
if (!handledIndexTensorSpace) {
|
|
|
|
handledIndexTensorSpace = true;
|
|
|
|
for (unsigned j = 0; j < indexBroadcastShapeValue.size(); j++) {
|
|
|
|
result.push_back(indexBroadcastShapeValue[j]);
|
|
|
|
resultInt.push_back(indexBroadcastShapeInt[j]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return std::make_tuple(result, resultInt);
|
|
|
|
}
|
|
|
|
|
|
|
|
static FailureOr<Value>
|
|
|
|
getScatterIndices(Aten_IndexPutImplOp op, ConversionPatternRewriter &rewriter,
|
|
|
|
Type indicesDtype, SmallVector<Value> optionalIndices,
|
|
|
|
SmallVector<int64_t> indexBroadcastShapeInt,
|
|
|
|
SmallVector<Value> indexBroadcastShapeValue) {
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
MLIRContext *context = op->getContext();
|
|
|
|
Value input = op.getSelf();
|
|
|
|
|
|
|
|
SmallVector<SmallVector<int64_t>> shapeMap =
|
|
|
|
getInputShapeToOutputShapeMap(optionalIndices, indexBroadcastShapeValue);
|
|
|
|
|
|
|
|
SmallVector<int64_t> inputShapeInt{
|
|
|
|
input.getType().cast<BaseTensorType>().getSizes()};
|
|
|
|
int64_t inputRank = inputShapeInt.size();
|
|
|
|
SmallVector<Value> inputShapeValue;
|
|
|
|
for (unsigned i = 0; i < inputShapeInt.size(); i++) {
|
|
|
|
Value dim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(i));
|
|
|
|
inputShapeValue.push_back(
|
|
|
|
rewriter.createOrFold<AtenSizeIntOp>(loc, input, dim));
|
|
|
|
}
|
|
|
|
|
|
|
|
auto finalShapeResult = getIndicesFinalShape(
|
|
|
|
rewriter, loc, input, optionalIndices, inputShapeInt, inputShapeValue,
|
|
|
|
indexBroadcastShapeInt, indexBroadcastShapeValue);
|
|
|
|
SmallVector<Value> finalShapeValue = std::get<0>(finalShapeResult);
|
|
|
|
SmallVector<int64_t> finalShapeInt = std::get<1>(finalShapeResult);
|
|
|
|
|
|
|
|
Value torchCstNone = rewriter.create<Torch::ConstantNoneOp>(loc);
|
|
|
|
Value torchCstZero =
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(loc, rewriter.getI64IntegerAttr(0));
|
|
|
|
Value torchCstOne =
|
|
|
|
rewriter.create<Torch::ConstantIntOp>(loc, rewriter.getI64IntegerAttr(1));
|
|
|
|
|
|
|
|
Value indexBroadcastShapeTorchList = rewriter.create<PrimListConstructOp>(
|
|
|
|
loc, Torch::ListType::get(Torch::IntType::get(context)),
|
|
|
|
indexBroadcastShapeValue);
|
|
|
|
|
|
|
|
// Calculating index count.
|
|
|
|
int64_t indexCount = 1;
|
|
|
|
if (llvm::all_of(finalShapeInt,
|
|
|
|
[](int64_t shape) { return shape != kUnknownSize; })) {
|
|
|
|
for (int64_t i : finalShapeInt)
|
|
|
|
indexCount *= i;
|
|
|
|
} else {
|
|
|
|
indexCount = kUnknownSize;
|
|
|
|
}
|
|
|
|
|
|
|
|
Value indexCountValue = finalShapeValue[0];
|
|
|
|
for (unsigned i = 1; i < finalShapeValue.size(); i++)
|
|
|
|
indexCountValue =
|
|
|
|
rewriter.create<AtenMulIntOp>(loc, indexCountValue, finalShapeValue[i]);
|
|
|
|
|
|
|
|
ValueTensorType flattenIndicesType =
|
|
|
|
ValueTensorType::get(context, llvm::ArrayRef(indexCount), indicesDtype);
|
|
|
|
Value flattenEndDim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(finalShapeInt.size() - 1));
|
|
|
|
|
|
|
|
SmallVector<Value> broadcastedIndices;
|
|
|
|
for (unsigned i = 0; i < optionalIndices.size(); i++) {
|
|
|
|
Value broadcastedIndexTensor;
|
|
|
|
if (optionalIndices[i].getType().isa<Torch::NoneType>()) {
|
|
|
|
Value torchCstDim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(i));
|
|
|
|
Value inputDim = rewriter.create<AtenSizeIntOp>(loc, input, torchCstDim);
|
|
|
|
ValueTensorType tensorType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef(inputShapeInt[i]), indicesDtype);
|
|
|
|
broadcastedIndexTensor = rewriter.create<AtenArangeStartStepOp>(
|
|
|
|
loc, tensorType, /*start=*/torchCstZero, /*end=*/inputDim,
|
|
|
|
/*step=*/torchCstOne,
|
|
|
|
/*dtype=*/torchCstNone,
|
|
|
|
/*layout=*/torchCstNone,
|
|
|
|
/*device=*/torchCstNone,
|
|
|
|
/*pin_memory=*/torchCstNone);
|
|
|
|
} else {
|
|
|
|
ValueTensorType tensorType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef(indexBroadcastShapeInt), indicesDtype);
|
|
|
|
broadcastedIndexTensor = rewriter.create<AtenBroadcastToOp>(
|
|
|
|
loc, tensorType, optionalIndices[i], indexBroadcastShapeTorchList);
|
|
|
|
}
|
|
|
|
|
|
|
|
// spotlight_indices(final_shape, shape_map[i]):
|
|
|
|
// Turn all values in `final_shape` to `1` except for those with index in
|
|
|
|
// `indices`.
|
|
|
|
// for j in range(len(final_shape)):
|
|
|
|
// if j not in indices:
|
|
|
|
// final_shape[j] = 1
|
|
|
|
// This is equivalent to unsqueezing the index tensor at the dimension `j`
|
|
|
|
// not in indices.
|
|
|
|
for (unsigned j = 0; j < finalShapeInt.size(); j++) {
|
|
|
|
if (llvm::find(shapeMap[i], j) == shapeMap[i].end()) {
|
|
|
|
Value unsqueezeDim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(j));
|
|
|
|
auto unsqueezedInfo =
|
|
|
|
unsqueezeTensor(rewriter, op, broadcastedIndexTensor,
|
|
|
|
/*dim=*/unsqueezeDim);
|
|
|
|
if (failed(unsqueezedInfo)) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "cannot generate unsqueeze tensor op");
|
|
|
|
}
|
|
|
|
broadcastedIndexTensor = *unsqueezedInfo;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Performing broadcast to final shape.
|
|
|
|
Value broadcastShapeTorchList = rewriter.create<PrimListConstructOp>(
|
|
|
|
loc, Torch::ListType::get(Torch::IntType::get(context)),
|
|
|
|
finalShapeValue);
|
|
|
|
ValueTensorType broadcastTensorType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef(finalShapeInt), indicesDtype);
|
|
|
|
broadcastedIndexTensor = rewriter.create<AtenBroadcastToOp>(
|
|
|
|
loc, broadcastTensorType, broadcastedIndexTensor,
|
|
|
|
broadcastShapeTorchList);
|
|
|
|
|
|
|
|
// Flattening the tensor.
|
|
|
|
broadcastedIndexTensor = rewriter.create<AtenFlattenUsingIntsOp>(
|
|
|
|
loc, flattenIndicesType, broadcastedIndexTensor, torchCstZero,
|
|
|
|
flattenEndDim);
|
|
|
|
|
|
|
|
broadcastedIndices.push_back(broadcastedIndexTensor);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stacking broadcasted indices.
|
|
|
|
Value scatterIndices;
|
|
|
|
// The operation torch.stack([a, b], dim=0) is decomposed into:
|
|
|
|
// torch.cat([a.unsqueeze(dim=0), b.unsqueeze(dim=0)], dim=0)
|
|
|
|
// Unsqueeze all tensors before concatenating.
|
|
|
|
SmallVector<Value> unsqueezedIndexTensors;
|
|
|
|
for (Value tensor : broadcastedIndices) {
|
|
|
|
auto unsqueezedInfo =
|
|
|
|
unsqueezeTensor(rewriter, op, tensor, /*dim=*/torchCstZero);
|
|
|
|
if (failed(unsqueezedInfo)) {
|
|
|
|
return rewriter.notifyMatchFailure(op,
|
|
|
|
"cannot generate unsqueeze tensor op");
|
|
|
|
}
|
|
|
|
unsqueezedIndexTensors.push_back(*unsqueezedInfo);
|
|
|
|
}
|
|
|
|
|
|
|
|
BaseTensorType unsqueezedTensorType =
|
|
|
|
unsqueezedIndexTensors[0].getType().cast<BaseTensorType>();
|
|
|
|
Value concatIndicesTorchList = rewriter.create<PrimListConstructOp>(
|
|
|
|
loc, Torch::ListType::get(unsqueezedTensorType), unsqueezedIndexTensors);
|
|
|
|
ValueTensorType concatIndicesType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef({inputRank, indexCount}), indicesDtype);
|
|
|
|
scatterIndices = rewriter.create<AtenCatOp>(
|
|
|
|
loc, concatIndicesType, concatIndicesTorchList, torchCstZero);
|
|
|
|
|
|
|
|
ValueTensorType transposedIndicesType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef({indexCount, inputRank}), indicesDtype);
|
|
|
|
scatterIndices = rewriter.create<AtenTransposeIntOp>(
|
|
|
|
loc, transposedIndicesType, scatterIndices, torchCstZero, torchCstOne);
|
|
|
|
return scatterIndices;
|
|
|
|
}
|
|
|
|
|
2022-03-09 21:05:13 +08:00
|
|
|
namespace {
|
2022-10-28 23:06:11 +08:00
|
|
|
class ConvertAten_IndexPutImplOp
|
|
|
|
: public OpConversionPattern<Aten_IndexPutImplOp> {
|
2022-03-09 21:05:13 +08:00
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
2022-10-28 23:06:11 +08:00
|
|
|
matchAndRewrite(Aten_IndexPutImplOp op, OpAdaptor adaptor,
|
2022-03-09 21:05:13 +08:00
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
if (failed(verifyLinalgCompatibleTypes(op, rewriter)))
|
|
|
|
return failure();
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
MLIRContext *context = op->getContext();
|
2022-12-08 04:20:41 +08:00
|
|
|
Value input = adaptor.getSelf();
|
|
|
|
Value values = adaptor.getValues();
|
2022-03-09 21:05:13 +08:00
|
|
|
RankedTensorType inputType = input.getType().cast<RankedTensorType>();
|
|
|
|
RankedTensorType valuesType = values.getType().cast<RankedTensorType>();
|
2022-12-08 13:46:54 +08:00
|
|
|
int64_t inputRank = inputType.getRank();
|
|
|
|
auto valuesTensorType = op.getValues().getType().cast<BaseTensorType>();
|
2022-04-01 00:27:21 +08:00
|
|
|
auto resultType = typeConverter->convertType(op->getResult(0).getType())
|
|
|
|
.cast<RankedTensorType>();
|
2022-03-09 21:05:13 +08:00
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
if (!valuesTensorType.hasSizes())
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: the values tensor type must have sizes.");
|
|
|
|
|
2022-03-09 21:05:13 +08:00
|
|
|
// The unsafe should be either `False` or `none`.
|
2022-12-08 04:20:41 +08:00
|
|
|
if (!op.getUnsafe().getType().isa<Torch::NoneType>()) {
|
2022-03-09 21:05:13 +08:00
|
|
|
bool unsafe;
|
2022-12-08 04:20:41 +08:00
|
|
|
if (!matchPattern(op.getUnsafe(), m_TorchConstantBool(&unsafe)))
|
2022-03-09 21:05:13 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: unsafe must be a constant");
|
|
|
|
else if (unsafe)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: unsafe is expected to be false");
|
|
|
|
}
|
|
|
|
|
|
|
|
// The accumulate should be a torch constant of boolean type.
|
|
|
|
bool accumulate;
|
2022-12-08 04:20:41 +08:00
|
|
|
if (!matchPattern(op.getAccumulate(), m_TorchConstantBool(&accumulate)))
|
2022-03-09 21:05:13 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "Expected accumulate to be constant bool.");
|
|
|
|
|
|
|
|
// The element type of the `input` and `values` should be same.
|
|
|
|
if (inputType.getElementType() != valuesType.getElementType())
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "Input element type should be same as the values element type.");
|
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
SmallVector<Value> optionalIndicesList;
|
|
|
|
getListConstructElements(op.getIndices(), optionalIndicesList);
|
2022-03-09 21:05:13 +08:00
|
|
|
// The size of the list of the index tensors should not be greater than the
|
|
|
|
// input rank.
|
2022-12-08 13:46:54 +08:00
|
|
|
if ((int64_t)optionalIndicesList.size() > inputRank)
|
2022-03-09 21:05:13 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "Indices list size should not be greater than the input rank.");
|
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
Value torchCstNone = rewriter.create<Torch::ConstantNoneOp>(loc);
|
|
|
|
unsigned sizeOptionalIndicesList = optionalIndicesList.size();
|
|
|
|
SmallVector<int64_t> nonNoneIndexTensorDim;
|
|
|
|
unsigned numNonNoneIndices;
|
|
|
|
|
|
|
|
if (sizeOptionalIndicesList == 0)
|
|
|
|
return rewriter.notifyMatchFailure(op, "Indices list must not be empty.");
|
|
|
|
|
|
|
|
for (unsigned i = 0; i < optionalIndicesList.size(); i++) {
|
|
|
|
if (!optionalIndicesList[i].getType().isa<Torch::NoneType>()) {
|
|
|
|
nonNoneIndexTensorDim.push_back(i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
numNonNoneIndices = nonNoneIndexTensorDim.size();
|
|
|
|
if (numNonNoneIndices > 2) {
|
2022-03-09 21:05:13 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
2022-12-08 13:46:54 +08:00
|
|
|
op, "unimplemented: non none index tensors less than or equal to 2 "
|
|
|
|
"supported only");
|
|
|
|
} else if (numNonNoneIndices == 2 &&
|
|
|
|
nonNoneIndexTensorDim[0] != nonNoneIndexTensorDim[1] - 1) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: case of 2 non none index tensors is supported "
|
|
|
|
"only when both the tensors are along consecutive dimensions");
|
|
|
|
}
|
2022-03-09 21:05:13 +08:00
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
// Padding the indices list with none values.
|
|
|
|
if (sizeOptionalIndicesList < inputRank) {
|
|
|
|
for (unsigned i = 0; i < (inputRank - sizeOptionalIndicesList); i++)
|
|
|
|
optionalIndicesList.push_back(torchCstNone);
|
|
|
|
}
|
2022-03-09 21:05:13 +08:00
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
SmallVector<int64_t> indexBroadcastShapeInt{
|
|
|
|
optionalIndicesList[nonNoneIndexTensorDim[0]]
|
|
|
|
.getType()
|
|
|
|
.cast<BaseTensorType>()
|
|
|
|
.getSizes()};
|
|
|
|
SmallVector<Value> indexBroadcastShapeValue;
|
|
|
|
if (numNonNoneIndices == 2) {
|
|
|
|
computeBroadcastShape(rewriter, loc,
|
|
|
|
optionalIndicesList[nonNoneIndexTensorDim[0]],
|
|
|
|
optionalIndicesList[nonNoneIndexTensorDim[1]],
|
|
|
|
indexBroadcastShapeInt, indexBroadcastShapeValue);
|
|
|
|
} else {
|
|
|
|
// It means there's only one index tensor and broadcast shape is same as
|
|
|
|
// that index tensor' shape.
|
|
|
|
for (unsigned i = 0; i < indexBroadcastShapeInt.size(); i++) {
|
|
|
|
Value dim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(i));
|
|
|
|
indexBroadcastShapeValue.push_back(rewriter.createOrFold<AtenSizeIntOp>(
|
|
|
|
loc, optionalIndicesList[nonNoneIndexTensorDim[0]], dim));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Type indicesDtype = optionalIndicesList[nonNoneIndexTensorDim[0]]
|
|
|
|
.getType()
|
|
|
|
.cast<BaseTensorType>()
|
|
|
|
.getDtype();
|
|
|
|
|
|
|
|
// This implementation is done to get the scatter indices:
|
|
|
|
|
|
|
|
// def get_broadcast_shape(tensors):
|
|
|
|
// return list(torch.broadcast_tensors(*tensors)[0].shape)
|
|
|
|
|
|
|
|
// def get_input_shape_to_output_shape_map(optional_index_tensors:
|
|
|
|
// list[Optional[torch.Tensor]]):
|
|
|
|
// index_tensors = list(filter(lambda x: x is not None,
|
|
|
|
// optional_index_tensors)) broadcast_rank =
|
|
|
|
// len(get_broadcast_shape(index_tensors)) num_of_index_tensors =
|
|
|
|
// len(index_tensors) index_of_first_index_tensor: Optional[int] = None
|
|
|
|
// result = {}
|
|
|
|
// for i, index in enumerate(optional_index_tensors):
|
|
|
|
// if index is None:
|
|
|
|
// val = i
|
|
|
|
// if index_of_first_index_tensor is not None:
|
|
|
|
// val += broadcast_rank - num_of_index_tensors
|
|
|
|
// result[i] = [val]
|
|
|
|
// else:
|
|
|
|
// if index_of_first_index_tensor is None:
|
|
|
|
// index_of_first_index_tensor = i
|
|
|
|
// output_indices = list(range(index_of_first_index_tensor,
|
|
|
|
// index_of_first_index_tensor +
|
|
|
|
// broadcast_rank))
|
|
|
|
// result[i] = output_indices
|
|
|
|
// return result
|
|
|
|
|
|
|
|
// def spotlight_indices(shape, indices: list[int]):
|
|
|
|
// """Turn all values in `shape` to `1` except for those with index in
|
|
|
|
// `indices`.""" shape = shape.copy() for i in range(len(shape)):
|
|
|
|
// if i not in indices:
|
|
|
|
// shape[i] = 1
|
|
|
|
// return shape
|
|
|
|
|
|
|
|
// def get_final_shape(input, optional_index_tensors:
|
|
|
|
// list[Optional[torch.Tensor]]):
|
|
|
|
// index_tensors = list(filter(lambda x: x is not None,
|
|
|
|
// optional_index_tensors)) index_tensors_broadcast_shape =
|
|
|
|
// get_broadcast_shape(index_tensors) result = []
|
|
|
|
// handled_index_tensor_space = False
|
|
|
|
// for e, i in enumerate(input.shape):
|
|
|
|
// if optional_index_tensors[e] is None:
|
|
|
|
// result.append(i)
|
|
|
|
// else:
|
|
|
|
// if not handled_index_tensor_space:
|
|
|
|
// handled_index_tensor_space = True
|
|
|
|
// result += index_tensors_broadcast_shape
|
|
|
|
// return result
|
|
|
|
|
|
|
|
// def get_scatter_indices(input, optional_index_tensors:
|
|
|
|
// list[Optional[torch.Tensor]]):
|
|
|
|
// assert len(input.size()) == len(optional_index_tensors), "Pad indices
|
|
|
|
// with None" shape_map =
|
|
|
|
// get_input_shape_to_output_shape_map(optional_index_tensors)
|
|
|
|
// index_tensors = list(filter(lambda x: x is not None,
|
|
|
|
// optional_index_tensors)) index_tensors_broadcast_shape =
|
|
|
|
// get_broadcast_shape(index_tensors) final_shape =
|
|
|
|
// get_final_shape(input, optional_index_tensors)
|
|
|
|
|
|
|
|
// broadcasted_index_tensors = []
|
|
|
|
// for e, optional_index_tensor in enumerate(optional_index_tensors):
|
|
|
|
// if optional_index_tensor is None:
|
|
|
|
// tensor_to_broadcast = torch.arange(0, input.size(e))
|
|
|
|
// else:
|
|
|
|
// tensor_to_broadcast =
|
|
|
|
// optional_index_tensor.broadcast_to(index_tensors_broadcast_shape)
|
|
|
|
|
|
|
|
// broadcasted_index_tensor = \
|
|
|
|
// tensor_to_broadcast.reshape(spotlight_indices(final_shape, shape_map[e]))\
|
|
|
|
// .broadcast_to(final_shape)\
|
|
|
|
// .flatten()
|
|
|
|
// broadcasted_index_tensors.append(broadcasted_index_tensor)
|
|
|
|
|
|
|
|
// return torch.stack(broadcasted_index_tensors, dim=0).t()
|
|
|
|
|
|
|
|
auto scatterIndicesInfo =
|
|
|
|
getScatterIndices(op, rewriter, indicesDtype, optionalIndicesList,
|
|
|
|
indexBroadcastShapeInt, indexBroadcastShapeValue);
|
|
|
|
if (failed(scatterIndicesInfo)) {
|
2022-04-01 00:27:21 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
2022-12-08 13:46:54 +08:00
|
|
|
op, "cannot generate scatter indices for index put op");
|
|
|
|
}
|
|
|
|
Value indexTensor = *scatterIndicesInfo;
|
|
|
|
|
|
|
|
// Flattening the values tensor.
|
|
|
|
Value torchCstZero = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc, rewriter.getI64IntegerAttr(0));
|
|
|
|
Value flattenedValuesTensorLastDim = rewriter.create<Torch::ConstantIntOp>(
|
|
|
|
loc,
|
|
|
|
rewriter.getI64IntegerAttr(valuesTensorType.getSizes().size() - 1));
|
|
|
|
SmallVector<int64_t> valuesShapeInt{valuesTensorType.getSizes()};
|
|
|
|
int64_t valuesCount = 1;
|
|
|
|
if (llvm::all_of(valuesShapeInt,
|
|
|
|
[](int64_t shape) { return shape != kUnknownSize; })) {
|
|
|
|
for (int64_t i : valuesShapeInt)
|
|
|
|
valuesCount *= i;
|
|
|
|
} else {
|
|
|
|
valuesCount = kUnknownSize;
|
|
|
|
}
|
|
|
|
auto flattenedValuesTensorType = ValueTensorType::get(
|
|
|
|
context, llvm::ArrayRef(valuesCount), valuesTensorType.getDtype());
|
|
|
|
Value flattenedValuesTensor = rewriter.create<AtenFlattenUsingIntsOp>(
|
|
|
|
loc, flattenedValuesTensorType, op.getValues(), torchCstZero,
|
|
|
|
flattenedValuesTensorLastDim);
|
|
|
|
values = typeConverter->materializeTargetConversion(
|
|
|
|
rewriter, loc,
|
|
|
|
typeConverter->convertType(flattenedValuesTensor.getType()),
|
|
|
|
flattenedValuesTensor);
|
2022-03-09 21:05:13 +08:00
|
|
|
|
2022-04-01 00:27:21 +08:00
|
|
|
// `TMTensor::ScatterOp` expects indices of element type i32.
|
2022-03-09 21:05:13 +08:00
|
|
|
Value indices = convertTensorToDtype(
|
2022-12-08 13:46:54 +08:00
|
|
|
rewriter, loc, indexTensor,
|
2022-03-09 21:05:13 +08:00
|
|
|
mlir::IntegerType::get(context, 32, mlir::IntegerType::Signed));
|
|
|
|
indices = typeConverter->materializeTargetConversion(
|
|
|
|
rewriter, loc, typeConverter->convertType(indices.getType()), indices);
|
|
|
|
|
2022-12-08 13:46:54 +08:00
|
|
|
// Creating a tm_tensor.scatter op with the following mapping:
|
|
|
|
// 1.) Index tensor from the `indicesList` maps to the indices in scatter
|
|
|
|
// op.
|
|
|
|
// 2.) `values` is mapped to `updates` in scatter op.
|
|
|
|
// 3.) `input` is mapped to `original` in scatter op.
|
2022-04-01 00:27:21 +08:00
|
|
|
bool invalidInputTypeFound = false;
|
|
|
|
Value scatterOp = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, values, indices, input, /*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value valuesElement,
|
|
|
|
Value inputElement) {
|
|
|
|
Value yieldValue = valuesElement;
|
|
|
|
if (accumulate) {
|
|
|
|
if (inputElement.getType().isa<mlir::IntegerType>()) {
|
|
|
|
yieldValue =
|
|
|
|
b.create<arith::AddIOp>(loc, inputElement, valuesElement);
|
|
|
|
} else if (inputElement.getType().isa<mlir::FloatType>()) {
|
|
|
|
yieldValue =
|
|
|
|
b.create<arith::AddFOp>(loc, inputElement, valuesElement);
|
|
|
|
} else {
|
|
|
|
invalidInputTypeFound = true;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
b.create<TMTensor::YieldOp>(loc, yieldValue);
|
|
|
|
});
|
|
|
|
|
|
|
|
if (invalidInputTypeFound) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op,
|
|
|
|
"unimplemented: input tensor must be of integer type or float type");
|
2022-03-09 21:05:13 +08:00
|
|
|
}
|
2022-04-01 00:27:21 +08:00
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::CastOp>(op, resultType, scatterOp);
|
2022-03-09 21:05:13 +08:00
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2022-04-14 20:16:39 +08:00
|
|
|
namespace {
|
|
|
|
// The original implementation of the op is as follows:
|
|
|
|
//
|
|
|
|
// Indices and GradOutput Layout: [N, C, H, W] or [C, H, W]
|
|
|
|
// Input Layout: [N, C, Hin, Win] or [C, Hin, Win]
|
|
|
|
//
|
|
|
|
// for i in range(N):
|
|
|
|
// for j in range(C):
|
|
|
|
// for k in range(H):
|
|
|
|
// for l in range(W):
|
|
|
|
// index = indices[i, j, k, l]
|
|
|
|
// result[i, j, index/Win, index%Win] += gradOutput[i, j, k, l]
|
|
|
|
//
|
|
|
|
// OR
|
|
|
|
//
|
|
|
|
// for i in range(C):
|
|
|
|
// for j in range(H):
|
|
|
|
// for k in range(W):
|
|
|
|
// index = indices[i, j, k]
|
|
|
|
// result[i, index/Win, index%Win] += gradOutput[i, j, k]
|
|
|
|
//
|
|
|
|
class ConvertAtenMaxPool2dWithIndicesBackwardOp
|
|
|
|
: public OpConversionPattern<AtenMaxPool2dWithIndicesBackwardOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenMaxPool2dWithIndicesBackwardOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
if (failed(verifyLinalgCompatibleTypes(op, rewriter)))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
MLIRContext *context = op->getContext();
|
2022-12-08 04:20:41 +08:00
|
|
|
Value gradOutput = adaptor.getGradOutput();
|
|
|
|
Value input = adaptor.getSelf();
|
2022-04-14 20:16:39 +08:00
|
|
|
RankedTensorType gradOutputType =
|
|
|
|
gradOutput.getType().cast<RankedTensorType>();
|
|
|
|
Type gradOutputElemType = gradOutputType.getElementType();
|
|
|
|
RankedTensorType inputType = input.getType().cast<RankedTensorType>();
|
|
|
|
Type inputElemType = inputType.getElementType();
|
|
|
|
int64_t tensorOperandRank = inputType.getRank();
|
|
|
|
|
|
|
|
// `TMTensor::ScatterOp` expects indices of element type i32.
|
|
|
|
Value indices = convertTensorToDtype(
|
2022-12-08 04:20:41 +08:00
|
|
|
rewriter, loc, op.getIndices(),
|
2022-04-14 20:16:39 +08:00
|
|
|
mlir::IntegerType::get(context, 32, mlir::IntegerType::Signed));
|
|
|
|
indices = typeConverter->materializeTargetConversion(
|
|
|
|
rewriter, loc, typeConverter->convertType(indices.getType()), indices);
|
|
|
|
RankedTensorType indicesType = indices.getType().cast<RankedTensorType>();
|
|
|
|
Type indicesElemType = indicesType.getElementType();
|
|
|
|
|
|
|
|
// The element type of the `input` and `grad_output` should be same.
|
|
|
|
if (inputElemType != gradOutputElemType)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op,
|
|
|
|
"Input element type should be same as the grad_output element type.");
|
|
|
|
|
|
|
|
// Since the scatter op requires indices to be a 2-d tensor, we create a new
|
|
|
|
// 5-d/4-d tensor (depending on the original indices layout) comprising the
|
|
|
|
// index values. We will collapse this tensor into a 2-d tensor. The
|
|
|
|
// algorithm for the creation of updated indices tensor is as follows:
|
|
|
|
//
|
|
|
|
// for i in range(N):
|
|
|
|
// for j in range(C):
|
|
|
|
// for k in range(H):
|
|
|
|
// for l in range(W):
|
|
|
|
// for m in range(4):
|
|
|
|
// if m == 0:
|
|
|
|
// updatedIndices[N][C][H][W][0] = i
|
|
|
|
// if m == 1:
|
|
|
|
// updatedIndices[N][C][H][W][1] = j
|
|
|
|
// if m == 2:
|
|
|
|
// updatedIndices[N][C][H][W][2] =
|
|
|
|
// originalIndices[i, j, k, l] / Win
|
|
|
|
// if m == 3:
|
|
|
|
// updatedIndices[N][C][H][W][3] =
|
|
|
|
// originalIndices[i, j, k, l] % Win
|
|
|
|
//
|
|
|
|
// OR
|
|
|
|
//
|
|
|
|
// for j in range(C):
|
|
|
|
// for k in range(H):
|
|
|
|
// for l in range(W):
|
|
|
|
// for m in range(3):
|
|
|
|
// if m == 0:
|
|
|
|
// updatedIndices[C][H][W][0] = i
|
|
|
|
// if m == 1:
|
|
|
|
// updatedIndices[C][H][W][1] = originalIndices[i, j, k, l] / Win
|
|
|
|
// if m == 2:
|
|
|
|
// updatedIndices[C][H][W][2] = originalIndices[i, j, k, l] % Win
|
|
|
|
|
|
|
|
SmallVector<Value> inputShape = getTensorSizes(rewriter, loc, input);
|
|
|
|
|
|
|
|
SmallVector<AffineExpr> originalIndicesDimExprs, updatedIndicesDimExprs;
|
|
|
|
for (int64_t i = 0; i < tensorOperandRank; i++) {
|
|
|
|
originalIndicesDimExprs.push_back(rewriter.getAffineDimExpr(i));
|
|
|
|
updatedIndicesDimExprs.push_back(rewriter.getAffineDimExpr(i));
|
|
|
|
}
|
|
|
|
updatedIndicesDimExprs.push_back(
|
|
|
|
rewriter.getAffineDimExpr(tensorOperandRank));
|
|
|
|
|
|
|
|
SmallVector<AffineMap> indexingMaps = AffineMap::inferFromExprList(
|
|
|
|
{originalIndicesDimExprs, updatedIndicesDimExprs});
|
2022-11-17 06:40:36 +08:00
|
|
|
SmallVector<utils::IteratorType> iteratorTypes(
|
|
|
|
tensorOperandRank + 1, utils::IteratorType::parallel);
|
2022-04-14 20:16:39 +08:00
|
|
|
|
|
|
|
SmallVector<OpFoldResult> updatedIndicesShape =
|
|
|
|
getAsOpFoldResult(getTensorSizes(rewriter, loc, indices));
|
|
|
|
updatedIndicesShape.push_back(rewriter.getIndexAttr(tensorOperandRank));
|
|
|
|
|
2022-10-18 12:22:53 +08:00
|
|
|
Value initTensor = rewriter.create<tensor::EmptyOp>(
|
2022-04-14 20:16:39 +08:00
|
|
|
loc, updatedIndicesShape, indicesElemType);
|
|
|
|
|
|
|
|
Value wIn = inputShape[tensorOperandRank - 1];
|
|
|
|
SmallVector<Value> cstValues;
|
|
|
|
for (int64_t i = 0; i < tensorOperandRank; i++)
|
|
|
|
cstValues.push_back(rewriter.create<arith::ConstantIndexOp>(loc, i));
|
|
|
|
|
|
|
|
Value updatedIndices =
|
|
|
|
rewriter
|
|
|
|
.create<linalg::GenericOp>(
|
|
|
|
loc, initTensor.getType(), indices, initTensor, indexingMaps,
|
|
|
|
iteratorTypes,
|
|
|
|
[tensorOperandRank, wIn, cstValues,
|
|
|
|
indicesElemType](OpBuilder &b, Location loc, ValueRange args) {
|
|
|
|
Value index = castIntToIndex(b, loc, args[0]);
|
|
|
|
Value updatedIndex = cstValues[0];
|
|
|
|
Value lastDim =
|
|
|
|
b.create<linalg::IndexOp>(loc, tensorOperandRank);
|
|
|
|
|
|
|
|
for (int64_t i = tensorOperandRank - 1; i >= 0; i--) {
|
|
|
|
Value result;
|
|
|
|
if (i == tensorOperandRank - 1)
|
|
|
|
result = b.create<arith::RemSIOp>(loc, index, wIn);
|
|
|
|
if (i == tensorOperandRank - 2)
|
|
|
|
result = b.create<arith::FloorDivSIOp>(loc, index, wIn);
|
|
|
|
if (i == tensorOperandRank - 3 ||
|
|
|
|
i == tensorOperandRank - 4)
|
|
|
|
result = b.create<linalg::IndexOp>(loc, i);
|
|
|
|
|
|
|
|
Value pred = b.create<arith::CmpIOp>(
|
|
|
|
loc, arith::CmpIPredicate::eq, lastDim, cstValues[i]);
|
|
|
|
Value addAmount = b.create<arith::SelectOp>(
|
|
|
|
loc, pred, result, cstValues[0]);
|
|
|
|
updatedIndex =
|
|
|
|
b.create<arith::AddIOp>(loc, updatedIndex, addAmount);
|
|
|
|
}
|
|
|
|
|
|
|
|
updatedIndex = b.create<arith::IndexCastOp>(
|
|
|
|
loc, indicesElemType, updatedIndex);
|
|
|
|
b.create<linalg::YieldOp>(loc, updatedIndex);
|
|
|
|
})
|
|
|
|
.getResult(0);
|
|
|
|
|
|
|
|
// Creating a new tensor initialized with zeros and size same as the input
|
|
|
|
// tensor.
|
|
|
|
Value outputTensor =
|
|
|
|
createZeroInitTensor(rewriter, loc, inputShape, inputElemType);
|
|
|
|
|
|
|
|
// Collapsing `gradOutput` into a 1-d tensor.
|
|
|
|
SmallVector<ReassociationIndices> reassociationCollapse(1);
|
|
|
|
for (auto i = 0; i < gradOutputType.getRank(); i++)
|
|
|
|
reassociationCollapse[0].push_back(i);
|
|
|
|
RankedTensorType gradOutputFlattenedType;
|
|
|
|
int64_t numelGradOutput = getNumberOfElements(gradOutputType);
|
2022-11-29 20:33:31 +08:00
|
|
|
gradOutputFlattenedType = RankedTensorType::get(
|
|
|
|
makeShapeLLVMCompatible({numelGradOutput}), gradOutputElemType);
|
2022-04-14 20:16:39 +08:00
|
|
|
Value gradOutputFlattened = rewriter.create<tensor::CollapseShapeOp>(
|
|
|
|
loc, gradOutputFlattenedType, gradOutput, reassociationCollapse);
|
|
|
|
|
|
|
|
// Collapsing updated indices into a 2-d tensor.
|
|
|
|
SmallVector<ReassociationIndices> reassociationCollapseIndices(2);
|
|
|
|
for (auto i = 0; i < tensorOperandRank; i++)
|
|
|
|
reassociationCollapseIndices[0].push_back(i);
|
|
|
|
reassociationCollapseIndices[1].push_back(tensorOperandRank);
|
|
|
|
int64_t numelIndices = getNumberOfElements(indicesType);
|
|
|
|
Value indicesCollapsed = rewriter.create<tensor::CollapseShapeOp>(
|
|
|
|
loc,
|
2022-11-29 20:33:31 +08:00
|
|
|
RankedTensorType::get(
|
|
|
|
makeShapeLLVMCompatible({numelIndices, tensorOperandRank}),
|
|
|
|
indicesElemType),
|
2022-04-14 20:16:39 +08:00
|
|
|
updatedIndices, reassociationCollapseIndices);
|
|
|
|
|
|
|
|
bool invalidInputTypeFound = false;
|
|
|
|
Value scatterOp = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, /*updates=*/gradOutputFlattened,
|
|
|
|
/*indices=*/indicesCollapsed, /*original=*/outputTensor,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value valuesElement,
|
|
|
|
Value inputElement) {
|
|
|
|
Value yieldValue = valuesElement;
|
|
|
|
if (inputElement.getType().isa<mlir::IntegerType>()) {
|
|
|
|
yieldValue =
|
|
|
|
b.create<arith::AddIOp>(loc, inputElement, valuesElement);
|
|
|
|
} else if (inputElement.getType().isa<mlir::FloatType>()) {
|
|
|
|
yieldValue =
|
|
|
|
b.create<arith::AddFOp>(loc, inputElement, valuesElement);
|
|
|
|
} else {
|
|
|
|
invalidInputTypeFound = true;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
b.create<TMTensor::YieldOp>(loc, yieldValue);
|
|
|
|
});
|
|
|
|
|
|
|
|
if (invalidInputTypeFound) {
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op,
|
|
|
|
"unimplemented: input tensor must be of integer type or float type");
|
|
|
|
}
|
|
|
|
|
|
|
|
Type newResultType = getTypeConverter()->convertType(op.getType());
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::CastOp>(op, newResultType, scatterOp);
|
|
|
|
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2023-02-22 07:05:55 +08:00
|
|
|
namespace {
|
|
|
|
class ConvertAtenScatterReduceTwoOp
|
|
|
|
: public OpConversionPattern<AtenScatterReduceTwoOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenScatterReduceTwoOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
if (failed(verifyLinalgCompatibleTypes(op, rewriter)))
|
|
|
|
return failure();
|
|
|
|
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
|
|
|
|
RankedTensorType selfType =
|
|
|
|
adaptor.getSelf().getType().cast<RankedTensorType>();
|
|
|
|
RankedTensorType indexType =
|
|
|
|
adaptor.getIndex().getType().cast<RankedTensorType>();
|
|
|
|
RankedTensorType srcType =
|
|
|
|
adaptor.getSrc().getType().cast<RankedTensorType>();
|
|
|
|
|
|
|
|
Value self = adaptor.getSelf();
|
|
|
|
|
|
|
|
if (selfType.getRank() != indexType.getRank() ||
|
|
|
|
indexType.getRank() != srcType.getRank())
|
|
|
|
return rewriter.notifyMatchFailure(op,
|
|
|
|
"'self', 'index' and 'src' should all "
|
|
|
|
"have the same number of dimensions.");
|
|
|
|
|
|
|
|
std::string reduceType;
|
|
|
|
if (!matchPattern(op.getReduce(), m_TorchConstantStr(reduceType)))
|
|
|
|
return rewriter.notifyMatchFailure(op,
|
|
|
|
"'reduce' must be a costant string");
|
|
|
|
|
|
|
|
int64_t dim;
|
|
|
|
if (!matchPattern(op.getDim(), m_TorchConstantInt(&dim)))
|
|
|
|
return rewriter.notifyMatchFailure(op, "'dim' is not constant");
|
|
|
|
|
|
|
|
bool includeSelf;
|
|
|
|
if (!matchPattern(op.getIncludeSelf(), m_TorchConstantBool(&includeSelf)))
|
|
|
|
return rewriter.notifyMatchFailure(op, "'include_self' is not constant");
|
|
|
|
|
|
|
|
// Get reduce string as the equivalent enum
|
|
|
|
auto reduceEnum = torch_upstream::get_reduction_enum(reduceType);
|
|
|
|
|
|
|
|
// Get the inputs reformatted for the TMScatterOp
|
|
|
|
auto [indices, updates] =
|
|
|
|
convertTorchScatterIndexAndSrcToTMScatterIndexAndSrc(
|
|
|
|
rewriter, adaptor.getIndex(), adaptor.getSrc(), dim);
|
|
|
|
|
|
|
|
// Value 'counts' will be used to tally the number of reductions into
|
|
|
|
// each unique index. The tally is used to calculate the average of the
|
|
|
|
// values scattered per index.
|
|
|
|
Value counts = nullptr;
|
|
|
|
if (reduceEnum == torch_upstream::ReductionType::MEAN) {
|
|
|
|
SmallVector<Value> selfShape =
|
|
|
|
getTensorSizes(rewriter, loc, adaptor.getSelf());
|
2023-04-25 23:52:46 +08:00
|
|
|
TypedAttr initAttr;
|
2023-02-22 07:05:55 +08:00
|
|
|
if (llvm::isa<mlir::FloatType>(srcType.getElementType())) {
|
|
|
|
initAttr = rewriter.getFloatAttr(srcType.getElementType(), 1);
|
|
|
|
} else if (llvm::isa<mlir::IntegerType>(srcType.getElementType())) {
|
|
|
|
initAttr = rewriter.getIntegerAttr(srcType.getElementType(), 1);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
Value initElement = rewriter.create<arith::ConstantOp>(loc, initAttr);
|
|
|
|
counts = createInitTensor(rewriter, loc, selfShape,
|
|
|
|
selfType.getElementType(), initElement);
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the original values shouldn't be included, normalize the
|
|
|
|
// input tensor where the scatters take place.
|
|
|
|
if (!includeSelf) {
|
|
|
|
Value normalizationValue;
|
|
|
|
if (reduceEnum == torch_upstream::ReductionType::SUM ||
|
|
|
|
reduceEnum == torch_upstream::ReductionType::MEAN) {
|
|
|
|
// Set the values in the input tensor to '0' so they are not included
|
|
|
|
normalizationValue = rewriter.create<arith::ConstantOp>(
|
|
|
|
loc, rewriter.getZeroAttr(srcType.getElementType()));
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::PROD) {
|
|
|
|
// Set the values in the input tensor to '1' (multiplication identity)
|
|
|
|
if (llvm::isa<mlir::FloatType>(srcType.getElementType())) {
|
|
|
|
normalizationValue = rewriter.create<arith::ConstantOp>(
|
|
|
|
loc, rewriter.getFloatAttr(srcType.getElementType(), 1.0));
|
|
|
|
} else if (llvm::isa<mlir::IntegerType>(srcType.getElementType())) {
|
|
|
|
normalizationValue = rewriter.create<arith::ConstantOp>(
|
|
|
|
loc, rewriter.getIntegerAttr(srcType.getElementType(), 1));
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::MAX) {
|
|
|
|
// Set the values in the input tensor to the smallest element of that
|
|
|
|
// type
|
2023-04-25 23:52:46 +08:00
|
|
|
TypedAttr minAttr = getNumericLimit(rewriter, srcType.getElementType(),
|
2023-02-22 07:05:55 +08:00
|
|
|
/*getMin=*/true);
|
|
|
|
normalizationValue = rewriter.create<arith::ConstantOp>(loc, minAttr);
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::MIN) {
|
|
|
|
// Set the values in the input tensor to the largest element of that
|
|
|
|
// type
|
2023-04-25 23:52:46 +08:00
|
|
|
TypedAttr maxAttr = getNumericLimit(rewriter, srcType.getElementType(),
|
2023-02-22 07:05:55 +08:00
|
|
|
/*getMin=*/false);
|
|
|
|
normalizationValue = rewriter.create<arith::ConstantOp>(loc, maxAttr);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Scatter the normalizations into the input tensor
|
|
|
|
Value indexSize = getTensorSize(rewriter, loc, adaptor.getIndex());
|
|
|
|
indexSize = castIntToIndex(rewriter, loc, indexSize);
|
|
|
|
Value normalizations = createInitTensor(
|
|
|
|
rewriter, loc, SmallVector<Value>({indexSize}),
|
|
|
|
srcType.getElementType(), /*init_element=*/normalizationValue);
|
|
|
|
self = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, normalizations, indices, self,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value update, Value current) {
|
|
|
|
b.create<TMTensor::YieldOp>(loc, update);
|
|
|
|
});
|
|
|
|
if (reduceEnum == torch_upstream::ReductionType::MEAN) {
|
|
|
|
counts = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, normalizations, indices, counts,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value update, Value current) {
|
|
|
|
b.create<TMTensor::YieldOp>(loc, update);
|
|
|
|
});
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create final operation
|
|
|
|
Value scatterOp = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, updates, indices, self,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value update, Value current) {
|
|
|
|
Value result;
|
|
|
|
if (reduceEnum == torch_upstream::ReductionType::SUM ||
|
|
|
|
reduceEnum == torch_upstream::ReductionType::MEAN) {
|
|
|
|
if (update.getType().isa<mlir::IntegerType>()) {
|
|
|
|
result = b.create<arith::AddIOp>(loc, update, current);
|
|
|
|
} else if (update.getType().isa<mlir::FloatType>()) {
|
|
|
|
result = b.create<arith::AddFOp>(loc, update, current);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::PROD) {
|
|
|
|
if (update.getType().isa<mlir::IntegerType>()) {
|
|
|
|
result = b.create<arith::MulIOp>(loc, update, current);
|
|
|
|
} else if (update.getType().isa<mlir::FloatType>()) {
|
|
|
|
result = b.create<arith::MulFOp>(loc, update, current);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::MAX) {
|
|
|
|
if (update.getType().isa<mlir::IntegerType>()) {
|
|
|
|
result = b.create<arith::MaxSIOp>(loc, update, current);
|
|
|
|
} else if (update.getType().isa<mlir::FloatType>()) {
|
|
|
|
result = b.create<arith::MaxFOp>(loc, update, current);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
} else if (reduceEnum == torch_upstream::ReductionType::MIN) {
|
|
|
|
if (update.getType().isa<mlir::IntegerType>()) {
|
|
|
|
result = b.create<arith::MinSIOp>(loc, update, current);
|
|
|
|
} else if (update.getType().isa<mlir::FloatType>()) {
|
|
|
|
result = b.create<arith::MinFOp>(loc, update, current);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
b.create<TMTensor::YieldOp>(loc, result);
|
|
|
|
});
|
|
|
|
|
|
|
|
// Special case for the mean
|
|
|
|
if (reduceEnum == torch_upstream::ReductionType::MEAN) {
|
|
|
|
counts = createTMTensorScatterOp(
|
|
|
|
rewriter, loc, updates, indices, counts,
|
|
|
|
/*uniqueIndices=*/false,
|
|
|
|
[&](OpBuilder &b, Location loc, Value update, Value current) {
|
|
|
|
Value result;
|
|
|
|
if (mlir::IntegerType intType =
|
|
|
|
llvm::dyn_cast<mlir::IntegerType>(current.getType())) {
|
|
|
|
Value constantUpdate = b.create<arith::ConstantOp>(
|
|
|
|
loc, b.getIntegerAttr(intType, 1));
|
|
|
|
result = b.create<arith::AddIOp>(loc, constantUpdate, current);
|
|
|
|
} else if (mlir::FloatType floatType =
|
|
|
|
llvm::dyn_cast<mlir::FloatType>(current.getType())) {
|
|
|
|
Value constantUpdate = b.create<arith::ConstantOp>(
|
|
|
|
loc, b.getFloatAttr(floatType, 1.0));
|
|
|
|
result = b.create<arith::AddFOp>(loc, constantUpdate, current);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
b.create<TMTensor::YieldOp>(loc, result);
|
|
|
|
});
|
|
|
|
|
|
|
|
Value output = rewriter.create<tensor::EmptyOp>(
|
|
|
|
loc, tensor::getMixedSizes(rewriter, loc, self),
|
|
|
|
selfType.getElementType());
|
|
|
|
|
|
|
|
// Finally divide the result
|
|
|
|
scatterOp =
|
|
|
|
rewriter
|
|
|
|
.create<linalg::MapOp>(
|
|
|
|
loc, ValueRange{scatterOp, counts}, output,
|
|
|
|
[&](OpBuilder &b, Location loc, ValueRange args) {
|
|
|
|
Value result;
|
|
|
|
if (llvm::isa<mlir::IntegerType>(args[0].getType())) {
|
|
|
|
result = b.create<arith::DivSIOp>(loc, args[0], args[1]);
|
|
|
|
} else if (llvm::isa<mlir::FloatType>(args[0].getType())) {
|
|
|
|
result = b.create<arith::DivFOp>(loc, args[0], args[1]);
|
|
|
|
} else {
|
|
|
|
llvm_unreachable("Only integer/float types supported!");
|
|
|
|
}
|
|
|
|
b.create<linalg::YieldOp>(loc, result);
|
|
|
|
})
|
|
|
|
.getResult()[0];
|
|
|
|
}
|
|
|
|
auto resultType = getTypeConverter()
|
|
|
|
->convertType(op->getResult(0).getType())
|
|
|
|
.cast<RankedTensorType>();
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::CastOp>(op, resultType, scatterOp);
|
|
|
|
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2023-04-03 18:49:01 +08:00
|
|
|
namespace {
|
|
|
|
class ConvertAtenSortOp : public OpConversionPattern<AtenSortOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenSortOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
|
|
|
|
Location loc = op.getLoc();
|
|
|
|
|
|
|
|
// Step 1. Fetch Input to sort.
|
|
|
|
Value inputTensor = adaptor.getSelf();
|
|
|
|
auto inputType = inputTensor.getType().cast<RankedTensorType>();
|
|
|
|
unsigned inputRank = inputType.getRank();
|
|
|
|
|
|
|
|
// Step 2. Fetch dimension to perform sort in.
|
|
|
|
int64_t dim;
|
|
|
|
if (!matchPattern(op.getDim(), m_TorchConstantInt(&dim)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: only constant dim value is supported");
|
|
|
|
dim = toPositiveDim(dim, inputRank);
|
|
|
|
if (!isValidDim(dim, inputRank)) {
|
|
|
|
return rewriter.notifyMatchFailure(op, "dim is statically invalid");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Step 3. Fetch the order of sorting.
|
|
|
|
bool descending;
|
|
|
|
if (!matchPattern(op.getDescending(), m_TorchConstantBool(&descending)))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: only constant descending value is supported");
|
|
|
|
|
|
|
|
// Step 4. Form a RankedTensorType with same shape as that of the input's
|
|
|
|
// but with elemental type i64.
|
|
|
|
RankedTensorType indicesType =
|
|
|
|
RankedTensorType::get(inputType.getShape(), rewriter.getI64Type());
|
|
|
|
|
|
|
|
// Step 5. Generate indices tensor.
|
|
|
|
SmallVector<Value> dynDims;
|
|
|
|
for (unsigned i = 0; i < inputType.getRank(); i++) {
|
|
|
|
if (inputType.isDynamicDim(i)) {
|
|
|
|
dynDims.push_back(rewriter.create<tensor::DimOp>(loc, inputTensor, i));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Value initEmptyTensor = rewriter.create<tensor::EmptyOp>(
|
|
|
|
loc, inputType.getShape(), rewriter.getI64Type(), dynDims);
|
|
|
|
|
|
|
|
SmallVector<AffineMap> indexingMaps = {
|
|
|
|
AffineMap::getMultiDimIdentityMap(inputRank, op.getContext())};
|
|
|
|
SmallVector<utils::IteratorType> iteratorTypes(
|
|
|
|
inputRank, utils::IteratorType::parallel);
|
|
|
|
Value indicesTensor =
|
|
|
|
rewriter
|
|
|
|
.create<linalg::GenericOp>(
|
|
|
|
loc, initEmptyTensor.getType(), ValueRange{}, initEmptyTensor,
|
|
|
|
indexingMaps, iteratorTypes,
|
|
|
|
[&](OpBuilder &b, Location loc, ValueRange args) {
|
|
|
|
Value index = b.create<linalg::IndexOp>(loc, dim);
|
|
|
|
index = castIndexToInt64(b, loc, index);
|
|
|
|
b.create<linalg::YieldOp>(loc, index);
|
|
|
|
})
|
|
|
|
.getResult(0);
|
|
|
|
|
|
|
|
// Step 6. Create TMTensor::SortOp.
|
|
|
|
SmallVector<Value> operands;
|
|
|
|
operands.push_back(inputTensor);
|
|
|
|
operands.push_back(indicesTensor);
|
|
|
|
SmallVector<Type> elementTypes;
|
|
|
|
elementTypes.push_back(inputType.getElementType());
|
|
|
|
elementTypes.push_back(indicesType.getElementType());
|
|
|
|
|
|
|
|
// The default value for aten.sort op's `stable` parameter is `false`.
|
|
|
|
// Refer: https://pytorch.org/docs/stable/generated/torch.sort.html
|
|
|
|
FailureOr<SmallVector<Value>> sortOpValues =
|
|
|
|
createTMTensorSortOp(rewriter, loc, operands, elementTypes,
|
|
|
|
/*dimension=*/dim, /*isStable=*/false,
|
|
|
|
/*isDescending=*/descending);
|
|
|
|
if (failed(sortOpValues))
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
loc, "Only Integer and Floating element type expected.");
|
|
|
|
|
|
|
|
auto sortOpVal = *sortOpValues;
|
|
|
|
rewriter.replaceOp(op, sortOpVal);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2022-09-09 12:36:40 +08:00
|
|
|
namespace {
|
|
|
|
class ConvertAtenCumsumOp : public OpConversionPattern<AtenCumsumOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenCumsumOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
|
2022-12-08 04:20:41 +08:00
|
|
|
Value input = adaptor.getSelf();
|
2022-09-09 12:36:40 +08:00
|
|
|
auto resultType = input.getType().cast<RankedTensorType>();
|
|
|
|
Type elementType = resultType.getElementType();
|
|
|
|
int64_t inputRank = resultType.getRank();
|
|
|
|
Location loc = op->getLoc();
|
2022-12-08 04:20:41 +08:00
|
|
|
Value dtype = op.getDtype();
|
2022-09-09 12:36:40 +08:00
|
|
|
if (!dtype.getType().isa<Torch::NoneType>())
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unsupported: dtype argument not supported");
|
|
|
|
|
|
|
|
int64_t dim;
|
2022-12-08 04:20:41 +08:00
|
|
|
if (!matchPattern(op.getDim(), m_TorchConstantInt(&dim)))
|
2022-09-09 12:36:40 +08:00
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op, "unimplemented: only constant dim value is supported");
|
|
|
|
dim = toPositiveDim(dim, inputRank);
|
|
|
|
if (!isValidDim(dim, inputRank))
|
|
|
|
return rewriter.notifyMatchFailure(op, "invalid dim");
|
|
|
|
|
|
|
|
SmallVector<Value> sizes = getTensorSizes(rewriter, loc, input);
|
|
|
|
Value output = createZeroInitTensor(rewriter, loc, sizes, elementType);
|
|
|
|
output = rewriter.create<tensor::CastOp>(loc, resultType, output);
|
|
|
|
|
|
|
|
SmallVector<Value> accSizes(sizes);
|
|
|
|
accSizes.erase(accSizes.begin() + dim);
|
2022-11-29 20:33:31 +08:00
|
|
|
SmallVector<int64_t> accStatic(
|
|
|
|
makeShapeTorchCompatible(resultType.getShape()));
|
2022-09-09 12:36:40 +08:00
|
|
|
accStatic.erase(accStatic.begin() + dim);
|
|
|
|
Value acc = createZeroInitTensor(rewriter, loc, accSizes, elementType);
|
2022-11-29 20:33:31 +08:00
|
|
|
Type accType =
|
|
|
|
RankedTensorType::get(makeShapeLLVMCompatible(accStatic), elementType);
|
2022-09-09 12:36:40 +08:00
|
|
|
acc = rewriter.create<tensor::CastOp>(loc, accType, acc);
|
|
|
|
|
|
|
|
Value result = createTMTensorScanOp(
|
|
|
|
rewriter, loc, input, output, acc, dim, /*inclusive=*/true,
|
|
|
|
[](OpBuilder &b, Location loc, Value input, Value acc) {
|
|
|
|
Value sum = (input.getType().isa<mlir::FloatType>()
|
|
|
|
? b.create<arith::AddFOp>(loc, input, acc)
|
|
|
|
: b.create<arith::AddIOp>(loc, input, acc))
|
|
|
|
->getResult(0);
|
|
|
|
b.create<TMTensor::YieldOp>(loc, sum);
|
|
|
|
});
|
|
|
|
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::CastOp>(op, resultType, result);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2023-05-17 03:17:45 +08:00
|
|
|
namespace {
|
|
|
|
class ConvertAtenScaledDotProductAttentionOp
|
|
|
|
: public OpConversionPattern<AtenScaledDotProductAttentionOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern::OpConversionPattern;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenScaledDotProductAttentionOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
Value mask = op.getAttnMask();
|
|
|
|
Value dropoutP = op.getDropoutP();
|
|
|
|
Value isCausal = op.getIsCausal();
|
|
|
|
Value scale = op.getScale();
|
|
|
|
Type elementType =
|
|
|
|
adaptor.getQuery().getType().cast<ShapedType>().getElementType();
|
|
|
|
|
|
|
|
// Verify inputs (only support defaults)
|
|
|
|
if (!mask.getType().isa<Torch::NoneType>())
|
|
|
|
return rewriter.notifyMatchFailure(op.getLoc(),
|
|
|
|
"attention masking not supported");
|
|
|
|
double dropout;
|
|
|
|
if (!matchPattern(dropoutP, m_TorchConstantFloat(&dropout)) ||
|
|
|
|
dropout > 0.0)
|
|
|
|
return rewriter.notifyMatchFailure(op.getLoc(), "dropout not supported");
|
|
|
|
bool causal;
|
|
|
|
if (!matchPattern(isCausal, m_TorchConstantBool(&causal)) || causal)
|
|
|
|
return rewriter.notifyMatchFailure(
|
|
|
|
op.getLoc(), "causal attention masking not supported");
|
|
|
|
if (!scale.getType().isa<Torch::NoneType>()) {
|
|
|
|
double scaleFloat;
|
|
|
|
if (!matchPattern(scale, m_TorchConstantFloat(&scaleFloat)) ||
|
|
|
|
scaleFloat != 1.0)
|
|
|
|
return rewriter.notifyMatchFailure(op.getLoc(),
|
|
|
|
"only default scale supported");
|
|
|
|
}
|
|
|
|
|
|
|
|
SmallVector<int64_t> outSizes(
|
|
|
|
adaptor.getQuery().getType().cast<ShapedType>().getShape());
|
|
|
|
SmallVector<int64_t> valueSizes(
|
|
|
|
adaptor.getValue().getType().cast<ShapedType>().getShape());
|
|
|
|
outSizes[outSizes.size() - 1] = valueSizes[valueSizes.size() - 1];
|
|
|
|
SmallVector<Value> outSizesDynamic(
|
|
|
|
getTensorSizes(rewriter, op.getLoc(), adaptor.getQuery()));
|
|
|
|
outSizesDynamic[outSizesDynamic.size() - 1] = getTensorSizes(
|
|
|
|
rewriter, op.getLoc(), adaptor.getValue())[valueSizes.size() - 1];
|
|
|
|
Type outType = RankedTensorType::get(outSizes, elementType);
|
|
|
|
Value output = createZeroInitTensor(rewriter, op.getLoc(), outSizesDynamic,
|
|
|
|
elementType);
|
|
|
|
|
|
|
|
// Overwrite with tm_tensor::attention
|
|
|
|
auto attention = rewriter.create<AttentionOp>(
|
|
|
|
op.getLoc(), outType,
|
|
|
|
SmallVector<Value>{adaptor.getQuery(), adaptor.getKey(),
|
|
|
|
adaptor.getValue()},
|
|
|
|
SmallVector<Value>{output});
|
|
|
|
|
|
|
|
rewriter.replaceOp(op, attention.getResult());
|
|
|
|
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2022-03-03 00:48:15 +08:00
|
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// The pass
|
|
|
|
// -----------------------------------------------------------------------------
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
class ConvertTorchToTMTensor
|
|
|
|
: public ConvertTorchToTMTensorBase<ConvertTorchToTMTensor> {
|
|
|
|
public:
|
|
|
|
void getDependentDialects(DialectRegistry ®istry) const override {
|
|
|
|
registry.insert<linalg::LinalgDialect>();
|
2022-03-16 18:44:23 +08:00
|
|
|
registry.insert<func::FuncDialect>();
|
2022-03-03 00:48:15 +08:00
|
|
|
registry.insert<tensor::TensorDialect>();
|
2022-10-05 21:28:06 +08:00
|
|
|
registry.insert<arith::ArithDialect>();
|
2022-03-03 00:48:15 +08:00
|
|
|
registry.insert<TMTensorDialect>();
|
|
|
|
TorchConversion::getBackendTypeConversionDependentDialects(registry);
|
|
|
|
}
|
|
|
|
|
|
|
|
void runOnOperation() override {
|
|
|
|
MLIRContext *context = &getContext();
|
|
|
|
ConversionTarget target(*context);
|
2022-03-16 18:44:23 +08:00
|
|
|
target.addLegalDialect<linalg::LinalgDialect, func::FuncDialect,
|
2022-10-05 21:28:06 +08:00
|
|
|
tensor::TensorDialect, arith::ArithDialect,
|
2023-05-17 03:17:45 +08:00
|
|
|
math::MathDialect, Torch::TorchDialect,
|
|
|
|
TMTensorDialect>();
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
TypeConverter typeConverter;
|
|
|
|
typeConverter.addConversion([](Type type) { return type; });
|
|
|
|
TorchConversion::setupBackendTypeConversion(target, typeConverter);
|
|
|
|
|
|
|
|
RewritePatternSet patterns(context);
|
|
|
|
target.addIllegalOp<AtenBincountOp>();
|
|
|
|
patterns.add<ConvertAtenBincountOp>(typeConverter, context);
|
2022-10-28 23:06:11 +08:00
|
|
|
target.addIllegalOp<Aten_IndexPutImplOp>();
|
|
|
|
patterns.add<ConvertAten_IndexPutImplOp>(typeConverter, context);
|
2022-04-14 20:16:39 +08:00
|
|
|
target.addIllegalOp<AtenMaxPool2dWithIndicesBackwardOp>();
|
|
|
|
patterns.add<ConvertAtenMaxPool2dWithIndicesBackwardOp>(typeConverter,
|
|
|
|
context);
|
2023-02-22 07:05:55 +08:00
|
|
|
target.addIllegalOp<AtenScatterReduceTwoOp>();
|
|
|
|
patterns.add<ConvertAtenScatterReduceTwoOp>(typeConverter, context);
|
2023-04-03 18:49:01 +08:00
|
|
|
target.addIllegalOp<AtenSortOp>();
|
|
|
|
patterns.add<ConvertAtenSortOp>(typeConverter, context);
|
2022-09-09 12:36:40 +08:00
|
|
|
target.addIllegalOp<AtenCumsumOp>();
|
|
|
|
patterns.add<ConvertAtenCumsumOp>(typeConverter, context);
|
2023-05-17 03:17:45 +08:00
|
|
|
target.addIllegalOp<AtenScaledDotProductAttentionOp>();
|
|
|
|
patterns.add<ConvertAtenScaledDotProductAttentionOp>(typeConverter,
|
|
|
|
context);
|
2022-03-03 00:48:15 +08:00
|
|
|
|
|
|
|
if (failed(applyPartialConversion(getOperation(), target,
|
|
|
|
std::move(patterns))))
|
|
|
|
return signalPassFailure();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
} // namespace
|
|
|
|
|
2022-04-27 03:27:51 +08:00
|
|
|
std::unique_ptr<OperationPass<func::FuncOp>>
|
2022-03-03 00:48:15 +08:00
|
|
|
mlir::torch::createConvertTorchToTMTensorPass() {
|
|
|
|
return std::make_unique<ConvertTorchToTMTensor>();
|
|
|
|
}
|