2023-12-16 03:37:49 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// Part of the LLVM Project, under the Apache License v3.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-1.0 WITH LLVM-exception
|
|
|
|
// Also available under a BSD-style license. See LICENSE.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "torch-mlir/Conversion/TorchToTensor/TorchToTensor.h"
|
|
|
|
|
|
|
|
#include "../PassDetail.h"
|
|
|
|
|
|
|
|
#include "mlir/Dialect/Arith/IR/Arith.h"
|
|
|
|
#include "mlir/Dialect/Tensor/IR/Tensor.h"
|
|
|
|
#include "mlir/Dialect/Traits.h"
|
|
|
|
#include "mlir/IR/Matchers.h"
|
|
|
|
#include "mlir/Transforms/DialectConversion.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchDialect.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/IR/TorchOps.h"
|
|
|
|
#include "torch-mlir/Dialect/Torch/Utils/Utils.h"
|
|
|
|
#include "torch-mlir/Dialect/TorchConversion/IR/TorchConversionDialect.h"
|
|
|
|
#include "torch-mlir/Dialect/TorchConversion/Transforms/BackendTypeConversion.h"
|
|
|
|
|
|
|
|
using namespace mlir;
|
|
|
|
using namespace mlir::torch;
|
|
|
|
using namespace mlir::torch::Torch;
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
2024-02-01 07:09:12 +08:00
|
|
|
class ConvertAtenItemOp : public OpConversionPattern<AtenItemOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern<AtenItemOp>::OpConversionPattern;
|
|
|
|
using OpAdaptor = typename AtenItemOp::Adaptor;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(AtenItemOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
auto operand = adaptor.getOperands()[0];
|
|
|
|
auto operandTy = cast<RankedTensorType>(operand.getType());
|
|
|
|
auto torchDTy = cast<ValueTensorType>(op.getOperand().getType()).getDtype();
|
|
|
|
|
|
|
|
if (operandTy.getNumElements() != 1)
|
|
|
|
return rewriter.notifyMatchFailure(op, "expected only one item");
|
|
|
|
|
|
|
|
auto zeroIdx = rewriter.create<arith::ConstantIndexOp>(op.getLoc(), 0);
|
|
|
|
auto rank = operandTy.getRank();
|
|
|
|
llvm::SmallVector<Value> indices(rank, zeroIdx);
|
|
|
|
|
|
|
|
Value extract = rewriter.create<tensor::ExtractOp>(
|
|
|
|
op.getLoc(), operandTy.getElementType(), operand, indices);
|
|
|
|
auto extractTy = extract.getType();
|
|
|
|
if (isa<mlir::IntegerType>(extractTy) && !extractTy.isInteger(64)) {
|
|
|
|
if (torchDTy.isSignlessInteger()) {
|
|
|
|
extract = rewriter.create<arith::ExtUIOp>(
|
|
|
|
op.getLoc(), rewriter.getIntegerType(64), extract);
|
|
|
|
} else {
|
|
|
|
extract = rewriter.create<arith::ExtSIOp>(
|
|
|
|
op.getLoc(), rewriter.getIntegerType(64), extract);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (isa<mlir::FloatType>(extractTy) && !extractTy.isF64()) {
|
|
|
|
extract = rewriter.create<arith::ExtFOp>(op.getLoc(),
|
|
|
|
rewriter.getF64Type(), extract);
|
|
|
|
}
|
|
|
|
|
|
|
|
rewriter.replaceOp(op, extract);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2023-12-16 03:37:49 +08:00
|
|
|
class ConvertAtenShapeToTensorPatternOp
|
|
|
|
: public OpConversionPattern<Aten_ShapeAsTensorOp> {
|
|
|
|
public:
|
|
|
|
using OpConversionPattern<Aten_ShapeAsTensorOp>::OpConversionPattern;
|
|
|
|
using OpAdaptor = typename Aten_ShapeAsTensorOp::Adaptor;
|
|
|
|
LogicalResult
|
|
|
|
matchAndRewrite(Aten_ShapeAsTensorOp op, OpAdaptor adaptor,
|
|
|
|
ConversionPatternRewriter &rewriter) const override {
|
|
|
|
auto loc = op.getLoc();
|
|
|
|
auto operand = adaptor.getOperands()[0];
|
|
|
|
auto operandTy = operand.getType().cast<RankedTensorType>();
|
|
|
|
auto resultTy =
|
|
|
|
getTypeConverter()->convertType(op.getType()).cast<RankedTensorType>();
|
|
|
|
|
|
|
|
int64_t rank = operandTy.getRank();
|
2024-02-29 01:47:06 +08:00
|
|
|
if (rank == 0) {
|
|
|
|
rewriter.replaceOpWithNewOp<tensor::EmptyOp>(op, resultTy.getShape(),
|
|
|
|
resultTy.getElementType());
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
|
2023-12-16 03:37:49 +08:00
|
|
|
SmallVector<Value> dims;
|
|
|
|
for (int i = 0; i < rank; ++i) {
|
|
|
|
Value dim = rewriter.createOrFold<tensor::DimOp>(loc, operand, i);
|
|
|
|
dim = rewriter.createOrFold<arith::IndexCastOp>(
|
|
|
|
loc, resultTy.getElementType(), dim);
|
|
|
|
dims.push_back(dim);
|
|
|
|
}
|
|
|
|
|
|
|
|
Value tensor =
|
|
|
|
rewriter.createOrFold<tensor::FromElementsOp>(op.getLoc(), dims);
|
|
|
|
rewriter.replaceOp(op, tensor);
|
|
|
|
return success();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
class ConvertTorchToTensor
|
|
|
|
: public ConvertTorchToTensorBase<ConvertTorchToTensor> {
|
|
|
|
public:
|
|
|
|
void getDependentDialects(DialectRegistry ®istry) const override {
|
|
|
|
registry.insert<tensor::TensorDialect>();
|
|
|
|
TorchConversion::getBackendTypeConversionDependentDialects(registry);
|
|
|
|
}
|
|
|
|
void runOnOperation() override {
|
|
|
|
MLIRContext *context = &getContext();
|
|
|
|
ConversionTarget target(*context);
|
|
|
|
target.addLegalDialect<arith::ArithDialect>();
|
|
|
|
target.addLegalDialect<tensor::TensorDialect>();
|
2024-02-01 07:09:12 +08:00
|
|
|
target.addIllegalOp<Torch::AtenItemOp>();
|
2023-12-16 03:37:49 +08:00
|
|
|
target.addIllegalOp<Torch::Aten_ShapeAsTensorOp>();
|
|
|
|
|
|
|
|
TypeConverter typeConverter;
|
|
|
|
typeConverter.addConversion([](Type type) { return type; });
|
|
|
|
TorchConversion::setupBackendTypeConversion(target, typeConverter);
|
|
|
|
|
|
|
|
RewritePatternSet patterns(context);
|
2024-02-01 07:09:12 +08:00
|
|
|
patterns.add<ConvertAtenShapeToTensorPatternOp, ConvertAtenItemOp>(
|
|
|
|
typeConverter, context);
|
2023-12-16 03:37:49 +08:00
|
|
|
|
|
|
|
if (failed(applyPartialConversion(getOperation(), target,
|
|
|
|
std::move(patterns))))
|
|
|
|
return signalPassFailure();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
std::unique_ptr<OperationPass<func::FuncOp>>
|
|
|
|
mlir::torch::createConvertTorchToTensorPass() {
|
|
|
|
return std::make_unique<ConvertTorchToTensor>();
|
|
|
|
}
|