2021-09-30 00:03:40 +08:00
# Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
# See https://llvm.org/LICENSE.txt for license information.
# SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
# Also available under a BSD-style license. See LICENSE.
2021-07-01 05:13:21 +08:00
# This file describes the sets of tests expected to fail for each config.
# This information is deliberately kept in a side table, rather than
# in-situ on the test, as a deliberate layering decision: tests should
# have unique keys to identify them and enable side tables of various kinds
# (this includes down into lower parts of the stack, where a side table
# might be used to keep more elaborate sets of testing configurations).
2022-04-20 03:35:56 +08:00
from torch_mlir_e2e_test . test_suite import COMMON_TORCH_MLIR_LOWERING_XFAILS
2023-07-13 21:07:54 +08:00
from torch_mlir . _version import torch_version_for_comparison , version
2022-04-20 03:35:56 +08:00
2023-08-18 23:15:54 +08:00
LINALG_XFAIL_SET = COMMON_TORCH_MLIR_LOWERING_XFAILS | {
# Lowering Torch Backend IR -> Linalg-on-Tensors Backend IR failed
# 'linalg.depthwise_conv_2d_nchw_chw' op inferred input/output operand #1 has shape's dimension #0 to be 4, but found 8
2023-08-27 21:56:36 +08:00
" Conv2dWithPaddingDilationStrideStaticModule_depthwise_multiplier " ,
2023-08-18 23:15:54 +08:00
}
2021-08-11 07:10:31 +08:00
2022-11-18 20:21:19 +08:00
TORCHDYNAMO_XFAIL_SET = {
#### General TorchDynamo/PyTorch errors
# RecursionError: maximum recursion depth exceeded
# RuntimeError: Failed running call_function aten.lift_fresh_copy(...
2022-11-24 22:36:13 +08:00
# https://github.com/pytorch/pytorch/issues/89627
2022-11-18 20:21:19 +08:00
" LiftFreshCopyModule_basic " ,
# TypeError: new_empty(): argument 'size' (position 1) must be tuple of ints, but found element of type NoneType at pos 0
# RuntimeError: Failed running call_function aten.convolution_backward(...
2022-11-24 22:36:13 +08:00
# https://github.com/pytorch/pytorch/issues/89629
2022-11-18 20:21:19 +08:00
" ConvolutionBackwardModule2DPadded_basic " ,
" ConvolutionBackwardModule2D_basic " ,
2023-02-07 23:25:59 +08:00
2022-11-18 20:21:19 +08:00
# RuntimeError: Index tensor must have the same number of dimensions as self tensor
# RuntimeError: Failed running call_function aten.nll_loss_backward(...
2022-11-24 22:36:13 +08:00
# https://github.com/pytorch/pytorch/issues/89630
2022-11-18 20:21:19 +08:00
" NllLossModuleBackward1DMeanWeight_basic " ,
" NllLossModuleBackward1DMean_basic " ,
" NllLossModuleBackward1DSumWeight_basic " ,
" NllLossModuleBackward1DSum_basic " ,
" NllLossModuleBackward1DWeight_basic " ,
" NllLossModuleBackward1D_basic " ,
2022-11-29 22:01:42 +08:00
# TypeError: uniform() missing 2 required keyword-only arguments: 'dtype' and 'device'
# RuntimeError: Failed running call_function aten.uniform(...
# https://github.com/pytorch/torchdynamo/issues/1954
" UniformNoCorrelationModule_basic " ,
2022-11-18 20:21:19 +08:00
#### Torch-MLIR internal compiler errors
# These are probably due to slightly different ops being recorded by
# torchdynamo vs. torchscript.
2022-11-29 22:01:42 +08:00
# No upstream decompositions.
2022-11-25 20:33:34 +08:00
# %6:4 = torch.operator "aten._embedding_bag_forward_only"(%1, %3, %5, %false, %int0, %false, %none, %false, %int-1) : (!torch.tensor<*,f32>, !torch.tensor<*,si64>, !torch.tensor<*,si64>, !torch.bool, !torch.int, !torch.bool, !torch.none, !torch.bool, !torch.int) -> (!torch.tensor, !torch.tensor, !torch.tensor, !torch.tensor)
2022-11-29 22:01:42 +08:00
# See also: https://github.com/pytorch/torchdynamo/issues/327
2022-11-18 20:21:19 +08:00
" AtenEmbeddingBagSumExample_basic " ,
2023-05-12 13:46:33 +08:00
2022-11-18 20:21:19 +08:00
# error: failed to legalize operation 'torch.valsem.aten.bernoulli.float' that was explicitly marked illegal
" BernoulliFloatModule_basic " ,
2023-02-16 01:06:29 +08:00
" BernoulliPModule_basic " ,
2022-11-18 20:21:19 +08:00
# error: failed to legalize operation 'torch.aten.view' that was explicitly marked illegal
" ElementwiseFlattenBroadcastModule_basic " ,
" FlattenRank0Module_basic " ,
2022-11-29 22:01:42 +08:00
" UniformModule_basic " ,
2023-06-27 10:55:28 +08:00
" UniformStaticShapeModule_basic " ,
2023-01-01 05:54:25 +08:00
# error: unsupported by backend contract: tensor with unknown rank
# note: see current operation: %1 = "torch.tensor_static_info_cast"(%arg0) : (!torch.vtensor<[5,4,3,2,1],f32>) -> !torch.vtensor<*,f32>
" ElementwisePreluModule_basic " ,
2023-02-07 13:38:44 +08:00
#ERROR: value (Tensor with shape=[2, 3, 6, 10], dtype=torch.float32, min=-1.336e-32, max=+0.9152, mean=+0.4837) is not close to golden value (Tensor with shape=[2, 3, 6, 10], dtype=torch.float32, min=+0.02233, max=+0.9152, mean=+0.4777)
" UpSampleNearest2dDynamicFactor_basic " ,
2023-02-06 19:52:04 +08:00
" ReduceMaxAlongDimUnsignedInt_basic " ,
2023-02-20 15:40:09 +08:00
#ERROR: value (-56) is not equal to golden value (200)
" AtenIntTensorByteDtypeModule_basic " ,
2023-03-07 02:12:58 +08:00
# ERROR: assert isinstance(e, FakeTensor)
" ElementwiseAddScalar_NumToTensorFloat_Module_basic " ,
2023-03-07 09:38:27 +08:00
# ERROR: assert isinstance(e, FakeTensor)
" RsubInt0d_NumToTensor_Module_basic " ,
2022-11-16 13:57:58 +08:00
# ERROR: RuntimeError: Found a custom (non-ATen) operator that either mutates or its inputs: prims::squeeze.. Getting these operators to work with functionalization requires some extra work. For mutable ops you need to register a corresponding out-of-place variant of the op, and you also need to register a Functionalization kernel that performs some boilerplate, telling functionalization to map from the mutable op to the out-of-place op. See a more complete example of how to do this at https://gist.github.com/bdhirsh/7dadbf6296f8f7d1abcf4c482f438aaa.
" PrimsSqueezeModule_basic " ,
" PrimsSqueezeEmptyDimensionsModule_basic " ,
2023-04-05 19:32:52 +08:00
2023-04-10 11:50:26 +08:00
# ERROR: RuntimeError: Found a custom (non-ATen) operator that either mutates or its inputs: prims::view_of.. Getting these operators to work with functionalization requires some extra work. For mutable ops you need to register a corresponding out-of-place variant of the op, and you also need to register a Functionalization kernel that performs some boilerplate, telling functionalization to map from the mutable op to the out-of-place op. See a more complete example of how to do this at https://gist.github.com/bdhirsh/7dadbf6296f8f7d1abcf4c482f438aaa.
" PrimsViewOfModule_basic " ,
" PrimsViewOfZeroRankModule_basic " ,
2023-04-29 07:05:17 +08:00
# See https://github.com/llvm/torch-mlir/pull/2040 and corresponding upstream issue
# https://github.com/pytorch/pytorch/issues/99752.
# torch._dynamo.exc.Unsupported: call_function BuiltinVariable(bool) [TensorVariable()] {}
' TensorToBoolZeroRank_basic ' ,
' TensorToBool_basic ' ,
2023-05-12 13:46:33 +08:00
# START tests failing due to: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(float) [TensorVariable()] {}
2023-04-29 07:05:17 +08:00
' AtenSubFloatModule_basic ' ,
' BoolFloatFalseModule_basic ' ,
' BoolFloatTrueModule_basic ' ,
' CeilFloatModule_basic ' ,
' DivFloatModule_basic ' ,
' GeFloatIntModule_basic ' ,
' GeFloatModule_basic ' ,
' GtFloatIntModule_basic ' ,
' NeFloatIntModule_basic ' ,
' SubFloatModule_basic ' ,
' TensorToFloatZeroRank_basic ' ,
' TensorToFloat_basic ' ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(float) [TensorVariable()] {}
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# START tests failing due to: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(int) [TensorVariable()] {}
2023-04-29 07:05:17 +08:00
' AddIntModule_basic ' ,
' AtenIntTensorCharDtypeModule_basic ' ,
' BoolIntFalseModule_basic ' ,
' BoolIntTrueModule_basic ' ,
' DivIntModule_basic ' ,
' EqIntModule_basic ' ,
' GeIntModule_basic ' ,
' GtIntModule_basic ' ,
' MulIntModule_basic ' ,
' NeIntModule_basic ' ,
' SqrtIntModule_basic ' ,
' SubIntModule_basic ' ,
' TensorToIntZeroRank_basic ' ,
' TensorToInt_basic ' ,
' UnsafeViewCollapseDynamicWithAtenSizeIntModule_basic ' ,
' ViewCollapseDynamicWithAtenSizeIntModule_basic ' ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(int) [TensorVariable()] {}
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: call_method ListVariable() sort [] {'reverse': ConstantVariable(bool)}
2023-04-29 07:05:17 +08:00
' SortIntListReverse_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: call_method ListVariable() sort [] {}
2023-04-29 07:05:17 +08:00
' SortIntList_basic ' ,
2023-05-12 13:46:33 +08:00
# START tests failing due to: torch._dynamo.exc.Unsupported: data dependent operator: aten._local_scalar_dense.default
2023-04-29 07:05:17 +08:00
' AtenFloatScalarModule_basic ' ,
' AtenIntBoolOpModule_basic ' ,
' OneHotModule_basic ' ,
' QuantizedMLP_basic ' ,
' ScalarImplicitFloatModule_basic ' ,
' ScalarImplicitIntModule_basic ' ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: torch._dynamo.exc.Unsupported: data dependent operator: aten._local_scalar_dense.default
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# START tests failing due to: torch._dynamo.exc.Unsupported: dynamic shape operator: aten.bincount.default
2023-04-29 07:05:17 +08:00
' BincountMinlengthModule_basic ' ,
' BincountModule_basic ' ,
' BincountStaticSizeModule_basic ' ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: torch._dynamo.exc.Unsupported: dynamic shape operator: aten.bincount.default
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor bool call_function aten.Bool
2023-04-29 07:05:17 +08:00
' BoolFloatConstantModule_basic ' ,
' BoolIntConstantModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor bool call_function aten.__contains__
2023-04-29 07:05:17 +08:00
' ContainsIntList_False ' ,
' ContainsIntList_True ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor bool call_function aten.all
2023-04-29 07:05:17 +08:00
' AllBoolFalseModule_basic ' ,
' AllBoolTrueModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor bool call_function aten.any
2023-04-29 07:05:17 +08:00
' AnyBoolFalseModule_basic ' ,
' AnyBoolTrueModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor float call_function aten.sqrt
2023-04-29 07:05:17 +08:00
' SqrtIntConstantModule_basic ' ,
2023-07-08 01:01:51 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function aten.size
' BroadcastDynamicDimModule_basic ' ,
2023-05-12 13:46:33 +08:00
# START tests failing due to: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function aten.Int
2023-04-29 07:05:17 +08:00
' AtenIntBoolOpConstFalseModule_basic ' ,
' AtenIntBoolOpConstTrueModule_basic ' ,
' IntFloatModule_basic ' ,
' PowIntFloatModule_basic ' ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function aten.Int
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function aten.len
2023-04-29 07:05:17 +08:00
' LenStrModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function aten.numel
2023-04-29 07:05:17 +08:00
' NumelModule_basic ' ,
' NumelZeroRankModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function prim.max
2023-04-29 07:05:17 +08:00
' PrimMaxIntModule_basic ' ,
2023-05-12 13:46:33 +08:00
# ERROR: torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor int call_function prim.min
2023-04-29 07:05:17 +08:00
' PrimMinIntModule_basic ' ,
2023-05-26 06:40:12 +08:00
' PrimMinIntDynamicModule_basic ' ,
2023-04-29 07:05:17 +08:00
2023-05-12 13:46:33 +08:00
# START tests failing due to: empty graph in dynamo
2023-04-29 07:05:17 +08:00
' IsFloatingPointFloat_True ' ,
' IsFloatingPointInt_False ' ,
' TorchPrimLoopForLikeModule_basic ' ,
' TorchPrimLoopWhileLikeModule_basic ' ,
2023-05-31 14:14:14 +08:00
" ScalarConstantTupleModule_basic " ,
2023-05-12 13:46:33 +08:00
# END tests failing due to: empty graph in dynamo
# ERROR due to: backend never runs because of empty frame
' ConstantBoolParameterModule_basic ' ,
# START tests failing due to: 'torch.aten.mul.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
" AddCDivModule_basic " ,
" ElementwiseMulScalarModule_basic " ,
" ElementwiseMulScalarModule_float " ,
" NativeGroupNormBackwardModule_basic " ,
" UpSampleNearest2dDynamicSize_basic " ,
" UpSampleNearest2dStaticFactor_basic " ,
" UpSampleNearest2dStaticSize_basic " ,
" UpSampleNearest2d_basic " ,
# END tests failing due to: 'torch.aten.mul.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
# START tests failing due to: 'torch.aten.add.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
" BatchNorm1DModule_basic " ,
" BatchNorm1DWith2DInputModule_basic " ,
" BatchNorm2DModule_basic " ,
" BatchNorm3DModule_basic " ,
2023-05-18 00:04:40 +08:00
" BatchNorm1DStaticShapeModule_basic " ,
2023-05-12 13:46:33 +08:00
" ElementwiseAddScalarFloatModule_basic " ,
" ElementwiseAddScalarInt64Module_basic " ,
" ElementwiseAddScalarIntModule_basic " ,
" MobilenetV3Module_basic " ,
" NativeBatchNorm1DModule_basic " ,
" NativeBatchNorm2DModule_basic " ,
" NativeBatchNorm3DModule_basic " ,
" NativeBatchNormNoneWeightModule_basic " ,
" NativeGroupNormModule_basic " ,
" ResNet18Module_basic " ,
" ResNet18StaticModule_basic " ,
# END tests failing due to: 'torch.aten.add.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
# ERROR: 'torch.aten.add.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.int'
" ElementwiseAddScalar_TensorLiteralInt32_Module_basic " ,
" HBC_basic " ,
# ERROR: 'torch.aten.div.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
" ElementwiseDivScalarModule_basic " ,
2023-06-12 17:18:38 +08:00
# ERROR: 'torch.aten.div.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.int'
" ElementwiseAtenDivIntScalarModule_basic " ,
2023-05-12 13:46:33 +08:00
# ERROR: 'torch.aten.mul.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.int'
" ElementwiseMulScalarModule_int " ,
# ERROR: 'torch.aten.sub.Tensor' op operand #1 must be Any Torch tensor type, but got '!torch.float'
" ElementwiseSubScalarFloatModule_basic " ,
" ElementwiseSubScalarIntModule_basic " ,
# ERROR: Exception: Unsupported: missing default value for argument 0 in schema for aten.div.Tensor_mode
" ElementwiseDivRoundingModeFloorModule_basic " ,
" ElementwiseDivRoundingModeTruncModule_basic " ,
# ERROR: Exception: Unsupported op: get_attr
" NumToTensorFloatModule_basic " ,
" NumToTensorIntModule_basic " ,
" TensorFloatModule_basic " ,
" TensorIntModule_basic " ,
# ERROR: Exception: Unsupported: missing default value for argument 0 in schema for aten.randn.generator
" RandnGeneratorF64Module_basic " ,
" RandnGeneratorModule_basic " ,
# START tests failing due to: complex floating point ops
# END tests failing due to: complex floating point ops
2023-05-19 10:07:58 +08:00
# ERROR: Exception: Unsupported: return type List[Tensor] in schema for aten.unbind.int
" UnbindIntListUnpack_Module_basic " ,
" UnbindIntGetItem_Module_basic " ,
2022-10-16 05:46:06 +08:00
# ERROR: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(float) [TensorVariable()] {}
" ScatterValueFloatModule_basic " ,
# ERROR: torch._dynamo.exc.Unsupported: call_function BuiltinVariable(int) [TensorVariable()] {}
" ScatterValueIntModule_basic " ,
2023-05-31 14:14:14 +08:00
2023-07-14 15:26:54 +08:00
# AssertionError: Unregistered operation: torch.aten._unsafe_index_put
" UnsafeIndexPutHackedTwin1DFloatNonAccumulateModule_basic " ,
2023-08-18 23:15:54 +08:00
# Lowering Torch Backend IR -> Linalg-on-Tensors Backend IR failed
# 'linalg.depthwise_conv_2d_nchw_chw' op inferred input/output operand #1 has shape's dimension #0 to be 4, but found 8
" Conv2dWithPaddingDilationStrideStaticModule_depthwise_multiplier " ,
2023-08-25 14:07:30 +08:00
# Exception: Unsupported: node.meta['val'] is not a FakeTensor or list of FakeTensor's: _scaled_dot_product_flash_attention;
" ScaledDotProductAttentionSameModule_basic " ,
" ScaledDotProductAttentionDifferentModule_basic " ,
2023-09-05 21:28:37 +08:00
# AssertionError: Unregistered operation: torch.aten._embedding_bag_forward_only
" AtenEmbeddingBagStaticModule_basic " ,
2023-04-29 07:05:17 +08:00
}
2023-08-25 14:07:30 +08:00
if torch_version_for_comparison ( ) < version . parse ( " 2.1.0.dev " ) :
TORCHDYNAMO_XFAIL_SET - = {
" ScaledDotProductAttentionSameModule_basic " ,
" ScaledDotProductAttentionDifferentModule_basic " ,
}
2023-04-29 07:05:17 +08:00
TORCHDYNAMO_CRASHING_SET = {
2023-05-12 13:46:33 +08:00
# No upstream decompositions.
# %6:4 = torch.operator "aten._embedding_bag_forward_only"(%1, %3, %5, %false, %int0, %false, %none, %false, %int-1) : (!torch.tensor<*,f32>, !torch.tensor<*,si64>, !torch.tensor<*,si64>, !torch.bool, !torch.int, !torch.bool, !torch.none, !torch.bool, !torch.int) -> (!torch.tensor, !torch.tensor, !torch.tensor, !torch.tensor)
# See also: https://github.com/pytorch/torchdynamo/issues/327
" Aten_EmbeddingBagExample_basic " ,
" FullModuleInt3D_basic " ,
" ThresholdBackward1dIntModule_basic " ,
" ThresholdBackward2dIntModule_basic " ,
" ThresholdBackward3dIntModule_basic " ,
# See https://github.com/llvm/torch-mlir/issues/2050
2023-04-29 07:05:17 +08:00
" ElementwiseCloneChannelsLastMemoryFormatModule_basic " ,
2023-05-02 10:35:26 +08:00
" ElementwiseCloneContiguousModule_basic " ,
" ElementwiseCloneModule_basic " ,
2023-04-29 07:05:17 +08:00
" ExpandAsFloatModule_basic " ,
" ExpandAsIntModule_basic " ,
" ExpandModule_basic " ,
" MoveDimIntModule_basic " ,
" MoveDimIntNegativeIndexModule_basic " ,
2023-05-02 10:35:26 +08:00
" NarrowVerticalTest2_basic " ,
" NarrowVerticalTest_basic " ,
2023-04-29 07:05:17 +08:00
" NumpyTRank2Module_basic " ,
" NumpyTRankNDynamicModule_basic " ,
" NumpyTRankNStaticModule_basic " ,
" PermuteModule_basic " ,
" PermuteNegativeIndexModule_basic " ,
" SelectIntNegativeDimAndIndexStaticModule_basic " ,
2023-05-12 13:46:33 +08:00
" TestMultipleTensorAndPrimitiveTypesReturn_basic " ,
2023-05-02 10:35:26 +08:00
" TModuleRank2_basic " ,
" ToCopyModule_basic " ,
" TransposeIntModule_basic " ,
" TransposeIntNegDimsModule_basic " ,
2023-07-18 00:51:24 +08:00
" IndexPutImpl2DNoneIndexStaticModule_basic " ,
2022-11-18 20:21:19 +08:00
}
2023-02-02 21:29:47 +08:00
STABLEHLO_PASS_SET = {
2023-08-04 09:05:34 +08:00
" TileBigDimsSizeModule_basic " ,
" TileSmallDimsSizeModule_basic " ,
2023-07-29 21:55:49 +08:00
" AddIntModule_basic " ,
" AtenIntBoolOpModule_basic " ,
" AtenIntTensorByteDtypeModule_basic " ,
" AtenIntTensorCharDtypeModule_basic " ,
" BoolFloatFalseModule_basic " ,
" BoolFloatTrueModule_basic " ,
" BoolIntFalseModule_basic " ,
" BoolIntTrueModule_basic " ,
" CeilFloatModule_basic " ,
" DivFloatModule_basic " ,
" DivIntModule_basic " ,
" EqIntModule_basic " ,
" GeFloatIntModule_basic " ,
" GeFloatModule_basic " ,
" GeIntModule_basic " ,
" GtFloatIntModule_basic " ,
" GtIntModule_basic " ,
" MulIntModule_basic " ,
" NeFloatIntModule_basic " ,
" NeIntModule_basic " ,
" SqrtIntModule_basic " ,
" SubFloatModule_basic " ,
" SubIntModule_basic " ,
" TensorToBoolZeroRank_basic " ,
" TensorToIntZeroRank_basic " ,
" TensorToFloatZeroRank_basic " ,
2023-08-15 19:36:08 +08:00
" IndexTensorStaticContiguousWithNoneModule_basic " ,
" IndexTensorStaticNonContiguousWithNoneModule_basic " ,
2023-06-21 01:14:09 +08:00
" AliasModule_basic " ,
2023-07-20 16:46:44 +08:00
" TensorIntModule_basic " ,
2023-05-26 22:57:57 +08:00
" AllBoolFalseModule_basic " ,
" AllBoolTrueModule_basic " ,
" AnyBoolFalseModule_basic " ,
" AnyBoolTrueModule_basic " ,
" AtenIntBoolOpConstFalseModule_basic " ,
" AtenIntBoolOpConstTrueModule_basic " ,
2023-09-04 14:04:09 +08:00
" AtenFloatScalarModule_basic " ,
" ScalarImplicitFloatModule_basic " ,
" ScalarImplicitIntModule_basic " ,
2023-05-26 22:57:57 +08:00
" AtenSubFloatModule_basic " ,
" BoolFloatConstantModule_basic " ,
" BoolIntConstantModule_basic " ,
" ContainsIntList_False " ,
" ContainsIntList_True " ,
" IntFloatModule_basic " ,
" IsFloatingPointFloat_True " ,
" IsFloatingPointInt_False " ,
" LenStrModule_basic " ,
" MeanDimAllReduceKeepdimModule_basic " ,
" MeanDimAllReduceModule_basic " ,
" MeanDimDtypeModule_basic " ,
" MeanDimKeepdimModule_basic " ,
" MeanDimModule_basic " ,
" MeanDimNegativeModule_basic " ,
" NumelZeroRankModule_basic " ,
" PowIntFloatModule_basic " ,
" PrimMaxIntModule_basic " ,
" PrimMinIntModule_basic " ,
2023-05-24 01:34:51 +08:00
" PrimMinIntDynamicModule_basic " ,
2023-05-26 22:57:57 +08:00
" SortIntListReverse_basic " ,
" SortIntList_basic " ,
" SqrtIntConstantModule_basic " ,
" StdBiasedModule_basic " ,
" StdDimBiasedModule_basic " ,
" TestMultipleTensorAndPrimitiveTypesReturn_basic " ,
" VarBiasedModule_basic " ,
" VarDimBiasedModule_basic " ,
" VarMeanBiasedModule_basic " ,
" VarMeanDimBiasedModule_basic " ,
2023-04-21 09:38:48 +08:00
" ConstantBoolParameterModule_basic " ,
2023-02-11 05:58:39 +08:00
" MaskedFillScalarIntValueStaticModule_basic " ,
" MaskedFillScalarFloatValueStaticModule_basic " ,
[Torch Dialect] add support for adaptive_avgpool_1d (#2342)
* [MLIR][TORCH] Fix aten.cumsum lowering for int32 input (#2351)
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
update
parent 22e88d523b1970b2e904eb5421d49d987a3d255e
author jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114110 +0800
committer jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114119 +0800
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
add support for adaptive_pool_id
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update:
* update
---------
Co-authored-by: Vivek Khandelwal <vivekkhandelwal1424@gmail.com>
2023-08-05 07:48:09 +08:00
" AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic " ,
2022-12-22 17:12:52 +08:00
" AdaptiveAvgPool2dNonUnitOutputSizeStaticModule_basic " ,
2023-02-02 21:29:47 +08:00
" AddSizeIntModule_basic " ,
" AddSizeIntNegDimModule_basic " ,
2022-09-20 22:31:24 +08:00
" ArangeDtypeFloatModule_basic " ,
" ArangeDtypeIntModule_basic " ,
" ArangeFalsePinMemoryModule_basic " ,
" ArangeFloatModule_basic " ,
" ArangeIntModule_basic " ,
" ArangeNegativeStartFloatModule_basic " ,
" ArangeNegativeStartIntModule_basic " ,
" ArangeStartFloatModule_basic " ,
" ArangeStartIntModule_basic " ,
" ArangeStartNegativeStepFloatModule_basic " ,
" ArangeStartNegativeStepIntModule_basic " ,
" ArangeStartStepFloatModule_basic " ,
" ArangeStartStepIntModule_basic " ,
" ArangeZeroElementOutputModule_basic " ,
2023-02-02 21:29:47 +08:00
" BatchMlpLayerModule_basic " ,
2023-05-18 00:04:40 +08:00
" BatchNorm1DModule_basic " ,
" BatchNorm1DWith2DInputModule_basic " ,
" BatchNorm2DModule_basic " ,
" BatchNorm3DModule_basic " ,
" BatchNorm1DStaticShapeModule_basic " ,
" ResNet18StaticModule_basic " ,
2022-11-21 21:50:35 +08:00
" BmmModule_basic " ,
" BroadcastToModule_basic " ,
" BroadcastToSameRankStaticModule_basic " ,
" BroadcastZeroRankInputStaticModule_basic " ,
2023-06-12 17:17:25 +08:00
" BroadcastListConstructWithMinusOneModule_basic " ,
2023-02-03 10:20:47 +08:00
" BucketizeTensorStaticFloatModule_basic " ,
" BucketizeTensorStaticModule_basic " ,
2023-01-30 13:38:27 +08:00
" CumsumStaticModule_basic " ,
" CumsumStaticNegativeDimModule_basic " ,
2023-04-18 23:59:14 +08:00
" DetachModule_basic " ,
2023-06-07 10:06:27 +08:00
" ElementwiseIsnanModule_basic " ,
2023-01-04 10:11:25 +08:00
" ElementwiseAtenLogicalAndOpPromoteBroadcastStaticShapeModule_basic " ,
" ElementwiseAtenLogicalNotOpModule_basic " ,
" ElementwiseAtenLogicalNotOpPromoteModule_basic " ,
" ElementwiseAtenLogicalOrOpPromoteBroadcastStaticShapeModule_basic " ,
" ElementwiseAtenLogicalXorOpPromoteBroadcastStaticShapeModule_basic " ,
2022-11-24 14:28:34 +08:00
" ElementwiseAtenWhereSelfModule_basic " ,
2023-05-06 06:21:55 +08:00
" ElementwiseWhereScalarOtherStaticModule_basic " ,
" ElementwiseWhereScalarSelfStaticModule_basic " ,
2023-01-12 06:40:03 +08:00
" ElementwiseBitwiseAndStaticShapeModule_basic " ,
" ElementwiseBitwiseNotInt64Module_basic " ,
" ElementwiseBitwiseNotInt32Module_basic " ,
2023-09-06 14:21:51 +08:00
" ElementwiseOrTensorStaticShapeModule_basic " ,
2023-01-12 06:40:03 +08:00
" ElementwiseBitwiseOrStaticShapeModule_basic " ,
" ElementwiseBitwiseXorStaticShapeModule_basic " ,
2022-09-16 15:09:21 +08:00
" ElementwiseClampModule_basic " ,
" ElementwiseClampMinModule_basic " ,
" ElementwiseClampMaxModule_basic " ,
2023-06-11 02:45:35 +08:00
" ElementwiseSignModule_basic " ,
2023-02-17 12:26:46 +08:00
" ElementwisePowModule_basic " ,
2023-05-05 00:55:03 +08:00
" ElementwisePowTensorStaticModule_basic " ,
" ElementwisePowTensorBroadcastStaticModule_basic " ,
2022-09-08 10:15:36 +08:00
" ElementwiseExpModule_basic " ,
2023-02-02 21:29:47 +08:00
" ElementwiseFlattenBroadcastModule_basic " ,
" ElementwiseLeakyReluModule_basic " ,
2023-08-24 00:27:00 +08:00
" ElementwiseEluModule_basic " ,
2023-08-25 22:42:29 +08:00
" ElementwiseEluNonDefaultModule_basic " ,
2022-09-08 10:15:36 +08:00
" ElementwiseLogModule_basic " ,
" ElementwiseNegModule_basic " ,
2022-12-22 10:13:59 +08:00
" ElementwiseRsqrtModule_basic " ,
" ElementwiseSigmoidModule_basic " ,
2022-09-08 10:15:36 +08:00
" ElementwiseSqrtModule_basic " ,
2023-02-07 03:14:26 +08:00
" ElementwiseSinModule_basic " ,
" ElementwiseCosModule_basic " ,
" ElementwiseCeilModule_basic " ,
" ElementwiseFloorModule_basic " ,
2022-09-08 10:15:36 +08:00
" ElementwiseUnaryModule_basic " ,
2023-02-02 21:29:47 +08:00
" ElementwiseUnsqueezeBroadcastModule_basic " ,
2022-09-08 10:15:36 +08:00
" ElementwiseUnsqueezeNegDimsModule_basic " ,
2022-09-23 10:24:36 +08:00
" ElementwiseToDtypeF32ToI64Module_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseAddModule_basic " ,
" ElementwiseAddScalarFloatModule_basic " ,
" ElementwiseAddScalarInt64Module_basic " ,
" ElementwiseAddScalarIntModule_basic " ,
2023-03-07 02:12:58 +08:00
" ElementwiseAddScalar_NumToTensorFloat_Module_basic " ,
" ElementwiseAddScalar_TensorLiteralInt32_Module_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseDivScalarModule_basic " ,
2023-06-12 17:18:38 +08:00
" ElementwiseAtenDivIntScalarModule_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseEqDiffWidthScalarModule_basic " ,
" ElementwiseEqFloatScalarModule_basic " ,
" ElementwiseEqIntScalarModule_basic " ,
2023-06-07 10:06:27 +08:00
" ElementwiseNeFloatScalarModule_basic " ,
" ElementwiseNeFloatTensorStaticModule_basic " ,
" ElementwiseNeIntTensorStaticModule_basic " ,
2023-07-17 15:49:04 +08:00
" ElementwiseEqBoolScalarModule_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseErfModule_basic " ,
" ElementwiseGeluModule_basic " ,
" ElementwiseGtFloatScalarModule_basic " ,
" ElementwiseGtIntScalarModule_basic " ,
" ElementwiseGtMixed2ScalarModule_basic " ,
2022-11-24 14:28:34 +08:00
" ElementwiseGeFloatIntScalarModule_basic " ,
" ElementwiseGeFloatScalarModule_basic " ,
" ElementwiseGeIntScalarModule_basic " ,
" ElementwiseGeMixedIntScalarModule_basic " ,
2023-01-04 00:30:16 +08:00
" ElementwiseLeakyReluStaticModule_basic " ,
2022-11-24 14:28:34 +08:00
" ElementwiseLeFloatIntScalarModule_basic " ,
" ElementwiseLeFloatScalarModule_basic " ,
" ElementwiseLeIntScalarModule_basic " ,
" ElementwiseLeMixedIntScalarModule_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseLtDiffWidthScalarModule_basic " ,
" ElementwiseLtFloatScalarModule_basic " ,
" ElementwiseLtIntScalarModule_basic " ,
" ElementwiseMulScalarModule_basic " ,
" ElementwiseMulScalarModule_float " ,
" ElementwiseMulScalarModule_int " ,
" ElementwiseNeIntScalarModule_basic " ,
" ElementwiseReciprocalModule_basic " ,
" ElementwiseRelu6Module_basic " ,
" ElementwiseReluModule_basic " ,
2022-11-24 14:28:34 +08:00
" ElementwiseRemainderScalarModule_Bool_basic " ,
" ElementwiseRemainderScalarModule_Float_basic " ,
" ElementwiseRemainderScalarModule_Int_Float_basic " ,
" ElementwiseRemainderScalarModule_Int_basic " ,
2022-09-23 20:39:15 +08:00
" ElementwiseSubScalarFloatModule_basic " ,
" ElementwiseSubScalarIntModule_basic " ,
2022-11-24 14:28:34 +08:00
" ElementwiseWhereScalarModule_basic " ,
2023-05-09 13:13:00 +08:00
" ElementwiseAbsModule_basic " ,
2022-11-21 21:50:35 +08:00
" EmbeddingModule1DIndices_basic " ,
" EmbeddingModuleI32Static_basic " ,
" EmbeddingModuleI32_basic " ,
" EmbeddingModuleI64_basic " ,
2022-11-30 19:14:48 +08:00
" EmbeddingModuleF16_basic " ,
2023-07-27 18:35:25 +08:00
" EmptyLikeMemoryFormatModule_basic " ,
" EmptyLikeModule_defaultDtype " ,
" EmptyLikeModule_falsePinMemory " ,
" EmptyLikeModule_float " ,
" EmptyLikeModule_int " ,
2022-09-23 10:24:36 +08:00
" ExpandAsIntModule_basic " ,
2022-09-08 10:15:36 +08:00
" ExpandModule_basic " ,
2023-07-27 18:35:25 +08:00
" Fill_TensorFloat64WithFloat32_basic " ,
" Fill_TensorFloat64WithFloat64_basic " ,
" Fill_TensorFloat64WithInt64_basic " ,
2023-05-12 07:41:46 +08:00
" Fill_TensorFloat64WithFloat32Static_basic " ,
" Fill_TensorFloat64WithInt64Static_basic " ,
2023-06-15 10:27:34 +08:00
" FlipModuleStaticShape_basic " ,
" FlipNegativeIndexModule_basic " ,
2022-09-23 10:24:36 +08:00
" FullLikeModuleDefaultDtype_basic " ,
" FullLikeModuleFalsePinMemory_basic " ,
" FullLikeModuleFloat2D_basic " ,
" FullLikeModuleFloat3DStatic_basic " ,
" FullLikeModuleFloat3D_basic " ,
" FullLikeModuleInt2DStatic_basic " ,
" FullLikeModuleInt2D_basic " ,
" FullLikeModuleInt3D_basic " ,
" FullModuleDefaultDtype_basic " ,
" FullModuleFalsePinMemory_basic " ,
" FullModuleFloat2D_basic " ,
" FullModuleFloat3D_basic " ,
" FullModuleInt2D_basic " ,
" FullModuleInt3D_basic " ,
2022-11-21 21:50:35 +08:00
" GatherStaticModule_basic " ,
" GatherModule_basic " ,
" Gather2DInputModdule_basic " ,
" GatherRandomIndexModule_basic " ,
2023-04-07 19:49:35 +08:00
" GatherNegativeDimModule_basic " ,
2022-12-21 20:09:43 +08:00
" GeluBackwardModule_basic " ,
2023-02-02 21:29:47 +08:00
" HardswishModule_basic " ,
" HardswishRandomModule_basic " ,
2022-11-21 21:50:35 +08:00
" HardTanhIntModule_basic " ,
" HardTanhModule_basic " ,
" HardsigmoidModule_basic " ,
" HardsigmoidRandomModule_basic " ,
" IndexSelectDynamicIndexSizeModule_basic " ,
" IndexSelectSingleIdxModule_basic " ,
" IndexSelectTwoIdxModule_basic " ,
" IndexSelectWholeDimensionModule_basic " ,
" IndexSelectWholeTensorModule_basic " ,
2023-04-07 19:49:35 +08:00
" IndexSelectNegativeDimModule_basic " ,
2023-05-25 02:13:57 +08:00
" IndexTensorStaticModule_basic " ,
" IndexTensorMultiIndexStaticModule_basic " ,
2022-12-22 10:13:59 +08:00
" LayerNormLastDimModule_basic " ,
" LayerNormModule_basic " ,
" LayerNormNormalizeOverAllDimsModule_basic " ,
2023-01-04 00:30:16 +08:00
" LeakyReluBackwardStaticModule_basic " ,
2023-03-21 05:14:27 +08:00
" LinalgVectorNormModule_basic " ,
" LinalgVectorNormKeepDimModule_basic " ,
2022-09-08 10:15:36 +08:00
" MatmulBroadcastBatchDim_basic " ,
" MatmulSingleDynamicBatchDim_basic " ,
" Matmul_3d " ,
" Matmul_4d " ,
2022-11-23 15:02:41 +08:00
" MeanDimEmptyDimModule_basic " ,
2022-09-08 10:15:36 +08:00
" MeanDtypeModule_basic " ,
2022-11-23 15:02:41 +08:00
" MeanDynamicSizesModule_basic " ,
" MeanLargeInputModule_basic " ,
" MeanModule_basic " ,
2023-02-02 21:29:47 +08:00
" Mlp1LayerModule_basic " ,
" Mlp2LayerModule_basic " ,
2022-09-08 10:15:36 +08:00
" MmTanhModule_basic " ,
2022-11-21 21:50:35 +08:00
" Mv_basic " ,
2022-12-22 10:13:59 +08:00
" NativeLayerNormModule4D_basic " ,
" NativeLayerNormModule_basic " ,
2023-04-11 16:02:28 +08:00
" OneHotModule_basic " ,
2022-11-23 15:02:41 +08:00
" PrimsConvertElementTypeModule_basic " ,
2022-09-08 10:15:36 +08:00
" ReduceFrobeniusNormKeepDimModule_basic " ,
2022-11-23 15:02:41 +08:00
" ReduceSumDimIntListElementTypeBoolModule_basic " ,
" ReduceSumElementTypeBoolModule_basic " ,
" ReduceSumDimIntListEmptyDimModule_basic " ,
2022-09-08 10:15:36 +08:00
" ReduceSumDimIntListDtypeFloatModule_basic " ,
" ReduceSumDimIntListDtypeIntModule_basic " ,
" ReduceSumDimIntListKeepDimFloatModule_basic " ,
" ReduceSumDimIntListKeepDimIntModule_basic " ,
" ReduceSumDtypeFloatModule_basic " ,
" ReduceSumDtypeIntModule_basic " ,
2023-03-21 05:14:27 +08:00
" ReduceL1NormModule_basic " ,
" ReduceL1NormWithDTypeModule_basic " ,
" ReduceL2NormModule_basic " ,
" ReduceL3NormAllDimsModule_basic " ,
" ReduceL3NormKeepDimModule_basic " ,
" ReduceLN3NormModule_basic " ,
" NormScalarOptDimKeepDimModule_basic " ,
" NormScalarOptDimModule_basic " ,
" NormalizeModule_basic " ,
2023-05-31 14:14:14 +08:00
" ScalarConstantTupleModule_basic " ,
2022-09-08 10:15:36 +08:00
" SelectIntModule_basic " ,
2023-01-18 02:14:14 +08:00
" SelectIntNegativeDimAndIndexStaticModule_basic " ,
2022-09-08 10:15:36 +08:00
" SliceSingleIdxModule_basic " ,
" SqueezeDimModule_dynamic " ,
" SqueezeDimModule_negDim " ,
2022-12-22 10:13:59 +08:00
" ToCopyBoolDTypeStaticModule_basic " ,
" ToCopyModule_basic " ,
" ToCopyWithDTypeFalsePinMemoryModule_basic " ,
" ToCopyWithDTypeModule_basic " ,
2022-09-08 10:15:36 +08:00
" ReduceFrobeniusNormModule_basic " ,
2022-08-23 16:47:21 +08:00
" FlattenStaticModule_basic " ,
" FlattenRank0Module_basic " ,
" TensorsConcatNegativeDimModule_basic " ,
2023-03-10 08:17:35 +08:00
" TensorsConcatPromoteDTypeModule_basic " ,
2023-05-10 23:25:58 +08:00
" TensorsConcatStaticModule_basic " ,
" TensorsConcatNegativeDimStaticModule_basic " ,
2023-03-11 09:25:25 +08:00
" TensorsStackModule_basic " ,
" TensorsStackNegativeDimModule_basic " ,
" TensorsStackPromoteDTypeModule_basic " ,
2022-09-23 10:24:36 +08:00
" LiftFreshCopyModule_basic " ,
" Mlp2LayerModuleNoBias_basic " ,
2022-08-23 16:47:21 +08:00
" NumelModule_basic " ,
2023-02-02 21:29:47 +08:00
" SiluModule_basic " ,
" SquareModule_basic " ,
2022-08-23 16:47:21 +08:00
" SqueezeModule_allUnitDim " ,
" SqueezeDimModule_unitDim " ,
2022-09-28 02:08:14 +08:00
" ViewCollapseOnesMiddleModule_basic " ,
2022-11-21 21:50:35 +08:00
" ViewDoubleMergeStaticModule_basic " ,
" ViewExpandDynamicDimModule_basic " ,
" ViewFlattenAndExpandModule_basic " ,
2022-09-28 02:08:14 +08:00
" ViewFiveTestStaticModule_basic " ,
" ViewOffsetTestStaticModule_basic " ,
" ViewTwoFiveThreeStaticModule_basic " ,
" ViewTwoToThreeStaticModule_basic " ,
" ViewExpandOnesMiddleOppModule_basic " ,
" ViewOffsetBackwardTestStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" NumToTensorFloatModule_basic " ,
" AtenToDeviceModule_basic " ,
2023-07-25 14:09:53 +08:00
" AvgPool1dStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" AvgPool2dStaticModule_basic " ,
" Conv2dWithPaddingDilationStrideStaticModule_basic " ,
2023-08-18 23:15:54 +08:00
" Conv2dWithPaddingDilationStrideStaticModule_depthwise " ,
" Conv2dWithPaddingDilationStrideStaticModule_depthwise_multiplier " ,
" Conv2dWithPaddingDilationStrideStaticModule_grouped " ,
" Conv2dWithPaddingDilationStrideStaticModule_grouped_multiplier " ,
2022-08-23 16:47:21 +08:00
" Convolution2DStaticModule_basic " ,
2022-08-27 06:29:07 +08:00
" ConvolutionModule2DTransposeStridedStatic_basic " ,
2022-08-23 16:47:21 +08:00
" ElementwiseCloneContiguousModule_basic " ,
2023-02-01 19:13:59 +08:00
" ElementwiseCloneChannelsLastMemoryFormatModule_basic " ,
2022-08-23 16:47:21 +08:00
" ElementwiseCloneModule_basic " ,
" ElementwiseBinaryStaticShapeModule_basic " ,
" ReturnThreeTensorFloat32_basic " ,
" BoolTensorReturnFalseModule_basic " ,
" BoolTensorReturnTrueModule_basic " ,
" BoolTensorReturnMixedModule_basic " ,
" SqueezeModule_static " ,
" TModuleRank1_basic " ,
" TModuleRank0_basic " ,
" ElementwiseToDtypeIdentityModule_basic " ,
2022-11-24 21:02:59 +08:00
" View1DFoldModule_basic " ,
" UnsafeView1DFoldModule_basic " ,
2022-11-21 21:50:35 +08:00
" RsubFloatModule_basic " ,
" RsubFloatModule_noalpha_basic " ,
" RsubIntModule_basic " ,
" RsubIntModule_noalpha_basic " ,
2023-03-07 09:38:27 +08:00
" RsubInt0d_NumToTensor_Module_basic " ,
2023-06-01 11:38:50 +08:00
" ScalarTensorDefaultDtypeModule_basic " ,
" ScalarTensorFloat32Module_basic " ,
" ScalarTensorInt32Module_basic " ,
" ScalarTensorInt64Module_basic " ,
2023-03-23 04:41:04 +08:00
" SelectScattertModule_basic " ,
" SelectScattertStaticModule_basic " ,
2022-11-21 21:50:35 +08:00
" SliceStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" SliceModule_basic " ,
" SliceNegIdxModule_basic " ,
" SliceOutOfLowerBoundStartIndexModule_basic " ,
" SliceOutOfUpperBoundIndexModule_basic " ,
2023-07-20 15:53:54 +08:00
" SliceOutOfUpperBoundIndexStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" SliceStartEqEndModule_basic " ,
" SliceSizeTwoStepModule_basic " ,
" SliceWholeTensorModule_basic " ,
2023-03-23 04:41:04 +08:00
" SliceScatterModule_basic " ,
" SliceScatterNegativeDimModule_basic " ,
" SliceScatterNegativeEndModule_basic " ,
" SliceScatterStaticModule_basic " ,
" SliceScatterStepVariationModule_basic " ,
" SliceScatterZeroDimModule_basic " ,
2022-11-21 21:50:35 +08:00
" SqueezeDimModule_static " ,
" SqueezeDimModule_identity " ,
" SqueezeModule_broadcast " ,
2022-08-23 16:47:21 +08:00
" ReturnTwoTensorF32I64_basic " ,
" Matmul4dStatic_basic " ,
" Matmul_dot " ,
" Matmul_2d " ,
" Matmul_matvec " ,
" Matmul_vecmat " ,
" MaxPool2dWithIndicesStaticModule_basic " ,
" MmDagModule_basic " ,
" MmModule_basic " ,
" MmModule_chained " ,
" MaxPool2dStaticModule_basic " ,
2023-05-19 10:07:35 +08:00
" EmptyModule_contiguous " ,
" EmptyModule_defaultDtype " ,
" EmptyModule_falsePinMemory " ,
" EmptyModule_int " ,
" EmptyModule_float " ,
" NewEmptyModuleDefaultDtype_basic " ,
" NewEmptyModuleFalsePinMemory_basic " ,
" NewEmptyModuleFloat2D_basic " ,
" NewEmptyModuleFloat3D_basic " ,
" NewEmptyModuleInt2D_basic " ,
" NewEmptyModuleInt3D_basic " ,
" NewEmptyModuleLayoutIntDtype_basic " ,
" NewEmptyModuleNonDefaultFloatDtype_basic " ,
" NewEmptyModuleNonDefaultIntDtype_basic " ,
" NewEmptyStridedModuleDefaultDtype_basic " ,
2022-08-23 16:47:21 +08:00
" PermuteModule_basic " ,
" PermuteNegativeIndexModule_basic " ,
2022-09-23 10:24:36 +08:00
" ReduceSumDimIntListKeepDimNegativeDimStaticModule_basic " ,
2023-07-27 18:35:25 +08:00
" ZeroFloat32Module_basic " ,
" ZeroInt32Module_basic " ,
" ZeroInt64Module_basic " ,
2022-09-23 10:24:36 +08:00
" ZerosLikeModule_defaultDtype " ,
" ZerosLikeModule_falsePinMemory " ,
" ZerosLikeModule_float " ,
" ZerosLikeModule_int " ,
2022-08-23 16:47:21 +08:00
" ZerosModuleDefaultDtype_basic " ,
" ZerosModuleInt2D_basic " ,
" ZerosModuleInt3D_basic " ,
" ZerosModuleFloat2D_basic " ,
" ZerosModuleFloat3D_basic " ,
" ZerosModuleFalsePinMemory_basic " ,
" OnesModuleDefaultDtype_basic " ,
" OnesModuleInt_basic " ,
" OnesModuleFloat_basic " ,
" OnesModuleFalsePinMemory_basic " ,
2022-09-23 10:24:36 +08:00
" OnesLikeModule_defaultDtype " ,
" OnesLikeModule_falsePinMemory " ,
" OnesLikeModule_float " ,
" OnesLikeModule_int " ,
2022-08-23 16:47:21 +08:00
" NewZerosModuleDefaultDtype_basic " ,
" NewZerosModuleInt2D_basic " ,
" NewZerosModuleInt3D_basic " ,
" NewZerosModuleFloat2D_basic " ,
" NewZerosModuleFloat3D_basic " ,
" NewZerosModuleFalsePinMemory_basic " ,
" NewOnesModuleDefaultDtype_basic " ,
" NewOnesModuleInt2D_basic " ,
" NewOnesModuleInt3D_basic " ,
" NewOnesModuleFloat2D_basic " ,
" NewOnesModuleFloat3D_basic " ,
" NewOnesModuleFalsePinMemory_basic " ,
2023-03-30 22:08:20 +08:00
" NewZerosStaticModuleLayoutStrided_basic " ,
2022-08-23 16:47:21 +08:00
" DropoutEvalIntModule_basic " ,
" DropoutEvalFloatModule_basic " ,
2023-06-27 14:19:33 +08:00
" DropoutTrainStaticShapeModule_basic " ,
" NativeDropoutEvalFloatModule_basic " ,
" NativeDropoutTrainStaticShapeModule_basic " ,
2022-08-23 16:47:21 +08:00
" ContiguousModule_basic " ,
" DropoutModule_basic " ,
" ViewCollapseModule_basic " ,
2022-11-24 21:02:59 +08:00
" ViewCollapseInferredDimModule_basic " ,
2022-08-23 16:47:21 +08:00
" ViewDynamicExpandCollapseModule_basic " ,
" ViewDynamicExpandModule_basic " ,
" ViewExpandModule_basic " ,
" ViewExpandOnesModule_basic " ,
" ViewExpandOnesBeforeAndAfterModule_basic " ,
" ViewExpandOnesMiddleModule_basic " ,
" ViewExpandCollapseModule_basic " ,
" ViewExpandCollapseWithOnesModule_basic " ,
2022-11-24 21:02:59 +08:00
" ViewExpandInferredDimModule_basic " ,
2023-01-21 03:56:26 +08:00
" ViewNegativeStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" ViewNoChangeStaticModule_basic " ,
" ViewNoChange1dModule_basic " ,
" ViewNoChange2dModule_basic " ,
" ViewNoChange3dModule_basic " ,
" UnsafeViewExpandModule_basic " ,
" ReduceMaxAllDims_basic " ,
" ReduceMaxFloatModule_basic " ,
" ReduceMaxSignedIntModule_basic " ,
" ReduceMaxUnsignedIntModule_basic " ,
2023-08-30 01:12:41 +08:00
" ReduceMinAllDims_basic " ,
" ReduceMinFloatModule_basic " ,
" ReduceMinSignedIntModule_basic " ,
" ReduceMinUnsignedIntModule_basic " ,
2022-08-23 16:47:21 +08:00
" ReduceSumDimIntListFloatModule_basic " ,
" ReduceSumDimIntListIntModule_basic " ,
" ReduceSumFloatModule_basic " ,
" ReduceSumSignedIntModule_basic " ,
" ReduceSumUnsignedIntModule_basic " ,
" RepeatModule_basic " ,
" ReshapeAliasCollapseModule_basic " ,
" ReshapeAliasExpandModule_basic " ,
" ReshapeExpandModule_basic " ,
2022-08-24 08:36:05 +08:00
" RollModule_basic " ,
2022-08-23 16:47:21 +08:00
" TestMultipleTensorReturn_basic " ,
[Torch Dialect] add support for adaptive_avgpool_1d (#2342)
* [MLIR][TORCH] Fix aten.cumsum lowering for int32 input (#2351)
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
update
parent 22e88d523b1970b2e904eb5421d49d987a3d255e
author jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114110 +0800
committer jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114119 +0800
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
add support for adaptive_pool_id
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update:
* update
---------
Co-authored-by: Vivek Khandelwal <vivekkhandelwal1424@gmail.com>
2023-08-05 07:48:09 +08:00
" AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
" AdaptiveAvgPool2dUnitOutputSizeStaticModule_basic " ,
" BaddbmmStaticModule_basic " ,
" BaddbmmBroadcast1DInputModule_basic " ,
" BaddbmmBroadcast2DInputModule_basic " ,
" NarrowHorizontalTest2_basic " ,
" NarrowHorizontalTest_basic " ,
" NarrowVerticalTest2_basic " ,
" NarrowVerticalTest_basic " ,
2023-07-20 16:46:44 +08:00
" NarrowTensorHorizontalModule_basic " ,
" NarrowTensorVerticalModule_basic " ,
2022-08-23 16:47:21 +08:00
" NumToTensorIntModule_basic " ,
" NumpyTRank0Module_basic " ,
" NumpyTRank1Module_basic " ,
" NumpyTRank2Module_basic " ,
" NumpyTRankNStaticModule_basic " ,
" NumpyTRankNDynamicModule_basic " ,
" TModuleRank2_basic " ,
" TensorLiteralModule_basic " ,
" TensorsConcatModule_basic " ,
" TensorOpaqueLiteralModule_basic " ,
" TransposeIntModule_basic " ,
" TransposeIntNegDimsModule_basic " ,
2022-09-23 10:24:36 +08:00
" ToDtypeBoolLayoutNoneModule_basic " ,
2022-11-21 21:50:35 +08:00
" ToDtypeBoolLayoutNoneStaticModule_basic " ,
2022-09-23 10:24:36 +08:00
" ToDtypeLayoutNoneModule_basic " ,
" ToDtypeLayoutStridedModule_basic " ,
" TypeAsSameModule_basic " ,
2023-07-20 09:51:58 +08:00
" TypeAsDifferentModule_basic " ,
2022-09-23 10:24:36 +08:00
" TypeConversionF32ToF64Module_basic " ,
" TypeConversionF64ToF32Module_basic " ,
" TypeConversionI1ToF32Module_basic " ,
" TypeConversionI1ToF64Module_basic " ,
" TypeConversionI1ToI32Module_basic " ,
" TypeConversionI1ToI64Module_basic " ,
" TypeConversionI32ToI64Module_basic " ,
" TypeConversionI64ToI32Module_basic " ,
2022-11-21 21:50:35 +08:00
" TypePromotionAlphaWiderModule_basic " ,
" TypePromotionSameCategoryZeroRankWider_basic " ,
" TypePromotionZeroRankHigherCategoryModule_basic " ,
2022-08-23 16:47:21 +08:00
" OnesModuleCPUDevice_basic " ,
" Permute0RankModule_basic " ,
" UnsafeViewCollapseModule_basic " ,
" UnsafeViewDynamicExpandModule_basic " ,
2022-10-11 17:52:01 +08:00
" AtenRoundIntModule_basic " ,
2022-12-01 23:46:17 +08:00
" TestF16Return_basic " ,
2023-02-02 21:29:47 +08:00
" _LogSoftmaxModuleStable_basic " ,
2022-11-16 13:57:58 +08:00
" PrimsSqueezeModule_basic " ,
" PrimsSqueezeEmptyDimensionsModule_basic " ,
2023-01-02 22:34:39 +08:00
" MoveDimIntModule_basic " ,
" MoveDimIntNegativeIndexModule_basic " ,
2022-12-08 22:15:31 +08:00
" ConvolutionBackwardModule2DStatic_basic " ,
2023-04-14 02:24:39 +08:00
" ConvolutionBackwardModule2DStrided_basic " ,
2023-04-10 11:50:26 +08:00
" PrimsViewOfModule_basic " ,
" PrimsViewOfZeroRankModule_basic " ,
2023-05-11 20:05:01 +08:00
" AtenComplex64Module_basic " ,
2023-05-24 03:43:33 +08:00
" SplitTensorGetItem_Module_basic " ,
2023-06-07 01:38:04 +08:00
" SplitTensorListUnpackModule_basic " ,
2023-07-20 15:53:54 +08:00
" SplitTensorNegativeDimModule_basic " ,
" SplitTensorLastSmallerModule_basic " ,
2023-09-04 09:59:26 +08:00
" SplitWithSizesListUnpackModule_basic " ,
2023-05-19 10:07:58 +08:00
" UnbindIntListUnpack_Module_basic " ,
" UnbindIntGetItem_Module_basic " ,
2023-05-26 10:05:19 +08:00
" ChunkListUnpack_Module_basic " ,
" ChunkListUnpackUneven_Module_basic " ,
2023-05-25 10:32:55 +08:00
" RandIntDtypeModule_basic " ,
" RandIntLowDtypeModule_basic " ,
" RandIntLowModule_basic " ,
" RandIntModule_basic " ,
" RandIntPinMemoryModule_basic " ,
2023-08-27 21:56:36 +08:00
" RandModule_basic " ,
2023-06-27 10:55:28 +08:00
" UniformStaticShapeModule_basic " ,
2023-05-25 10:32:55 +08:00
" UniformNoCorrelationModule_basic " ,
2023-07-18 22:32:26 +08:00
" TupleModule_basic " ,
2023-09-05 21:28:37 +08:00
" AtenEmbeddingBagStaticModule_basic " ,
2022-08-23 16:47:21 +08:00
}
2023-07-29 21:55:49 +08:00
STABLEHLO_CRASHING_SET = {
# These e2e tests crash because currently mlir-hlo's shape-component-analysis
# only support exact one index in tensor::ExtractOp when it's related with
# some tensors' shape. REF:
# https://github.com/tensorflow/mlir-hlo/blob/master/mhlo/analysis/shape_component_analysis.cc#L586
# FIXME if upstream mlir-hlo fix this.
" ViewCollapseDynamicWithAtenSizeIntModule_basic " ,
" UnsafeViewCollapseDynamicWithAtenSizeIntModule_basic " ,
2023-09-05 21:28:37 +08:00
" Aten_EmbeddingBagExample_basic " ,
" AtenEmbeddingBagSumExample_basic "
2023-07-29 21:55:49 +08:00
}
2021-10-08 10:07:03 +08:00
# Write the TOSA set as a "passing" set as it is very early in development
# and very few tests work yet.
2021-10-26 02:43:21 +08:00
TOSA_PASS_SET = {
2023-08-04 09:05:34 +08:00
" TileBigDimsSizeModule_basic " ,
" TileSmallDimsSizeModule_basic " ,
2023-07-18 00:51:24 +08:00
" IndexPutImpl2DNoneIndexStaticModule_basic " ,
2023-06-21 01:14:09 +08:00
" AliasModule_basic " ,
2023-04-27 23:31:36 +08:00
" MaxPool2dEmptyStrideStaticModule_basic " ,
2023-04-21 09:38:48 +08:00
" ConstantBoolParameterModule_basic " ,
2022-09-21 06:07:46 +08:00
" ElementwiseCloneContiguousModule_basic " ,
2023-02-01 19:13:59 +08:00
" ElementwiseCloneChannelsLastMemoryFormatModule_basic " ,
2022-09-21 06:07:46 +08:00
" ElementwiseCloneModule_basic " ,
2021-10-26 04:43:54 +08:00
" ElementwiseUnaryModule_basic " ,
2021-12-04 15:31:38 +08:00
" ElementwiseBinaryModule_basic " ,
2021-10-29 01:09:12 +08:00
" ElementwiseSigmoidModule_basic " ,
2022-03-24 04:35:43 +08:00
" ElementwiseExpModule_basic " ,
2022-08-31 03:44:00 +08:00
" ElementwiseReluModule_basic " ,
2023-01-11 13:42:07 +08:00
" ElementwiseLeakyReluModule_basic " ,
2023-08-24 00:27:00 +08:00
" ElementwiseEluModule_basic " ,
2023-08-25 22:42:29 +08:00
" ElementwiseEluNonDefaultModule_basic " ,
2021-11-11 11:03:36 +08:00
" ElementwiseFloorModule_basic " ,
" ElementwiseLogModule_basic " ,
2021-12-29 10:11:07 +08:00
" ElementwiseBinaryStaticShapeModule_basic " ,
2022-02-09 04:57:23 +08:00
" ElementwiseMinimumModule_basic " ,
" ElementwiseMinimumIntModule_basic " ,
2023-09-11 17:28:22 +08:00
" ElementwiseMinOtherIntModule_basic " ,
" ElementwiseMinOtherModule_basic " ,
2022-02-09 04:57:23 +08:00
" ElementwiseMaximumModule_basic " ,
" ElementwiseMaximumIntModule_basic " ,
2023-09-05 10:52:32 +08:00
" ElementwiseMaxOtherIntModule_basic " ,
" ElementwiseMaxOtherModule_basic " ,
2022-09-28 02:08:14 +08:00
" ViewDoubleMergeStaticModule_basic " ,
" ViewCollapseOnesMiddleModule_basic " ,
" ViewFiveTestStaticModule_basic " ,
" ViewOffsetTestStaticModule_basic " ,
" ViewTwoFiveThreeStaticModule_basic " ,
" ViewTwoToThreeStaticModule_basic " ,
" ViewExpandOnesMiddleOppModule_basic " ,
" ViewOffsetBackwardTestStaticModule_basic " ,
2021-11-12 08:15:58 +08:00
" TanhBackward_basic " ,
2023-03-07 02:16:37 +08:00
" HardtanhBackward_basic " ,
2021-11-24 14:25:59 +08:00
" ElementwiseAddModule_basic " ,
2021-11-08 23:56:40 +08:00
" ReturnThreeTensorFloat32_basic " ,
2021-11-25 06:01:48 +08:00
" AddCMulModule_basic " ,
" AddCDivModule_basic " ,
2021-12-01 05:48:10 +08:00
" SqueezeModule_broadcast " ,
2021-12-08 22:05:02 +08:00
" BoolTensorReturnFalseModule_basic " ,
" BoolTensorReturnTrueModule_basic " ,
" BoolTensorReturnMixedModule_basic " ,
2022-04-14 01:28:27 +08:00
" BoolTensorHandleSignless_basic " ,
2021-12-15 02:03:58 +08:00
" ElementwiseRsqrtModule_basic " ,
2023-01-18 02:14:14 +08:00
" SelectIntNegativeDimAndIndexStaticModule_basic " ,
2021-12-16 13:40:29 +08:00
" SqueezeModule_static " ,
" SqueezeModule_noUnitDim " ,
" SqueezeModule_allUnitDim " ,
2021-12-17 12:08:07 +08:00
" TModuleRank1_basic " ,
" TModuleRank0_basic " ,
2021-12-23 20:04:29 +08:00
" ElementwiseToDtypeIdentityModule_basic " ,
2022-08-11 07:24:02 +08:00
" AtenToDeviceModule_basic " ,
2022-11-24 21:02:59 +08:00
" View1DFoldModule_basic " ,
" UnsafeView1DFoldModule_basic " ,
2022-01-07 00:31:29 +08:00
" SqueezeDimModule_static " ,
" SqueezeDimModule_identity " ,
" SqueezeDimModule_unitDim " ,
2022-01-10 21:24:08 +08:00
" ReturnTwoTensorF32I64_basic " ,
2023-06-11 02:45:35 +08:00
" ElementwiseSignModule_basic " ,
2022-01-15 03:05:11 +08:00
" ElementwisePowModule_basic " ,
2022-01-20 04:10:04 +08:00
" BmmModule_basic " ,
2022-02-19 01:56:48 +08:00
" MmDagModule_basic " ,
2022-06-16 23:45:10 +08:00
" Matmul4dStatic_basic " ,
2022-01-20 04:10:04 +08:00
" Matmul_dot " ,
" Matmul_3d " ,
2022-06-14 20:31:30 +08:00
" RsubFloatModule_basic " ,
" RsubFloatModule_noalpha_basic " ,
2023-03-07 09:38:27 +08:00
" RsubInt0d_NumToTensor_Module_basic " ,
2023-01-12 06:40:03 +08:00
" ElementwiseBitwiseAndModule_basic " ,
" ElementwiseBitwiseAndStaticShapeModule_basic " ,
" ElementwiseBitwiseNotInt32Module_basic " ,
" ElementwiseBitwiseNotInt64Module_basic " ,
2023-09-06 14:21:51 +08:00
" ElementwiseOrTensorStaticShapeModule_basic " ,
" ElementwiseOrTensorModule_basic " ,
2023-01-12 06:40:03 +08:00
" ElementwiseBitwiseOrModule_basic " ,
" ElementwiseBitwiseOrStaticShapeModule_basic " ,
" ElementwiseBitwiseXorModule_basic " ,
" ElementwiseBitwiseXorStaticShapeModule_basic " ,
2023-06-11 02:45:35 +08:00
" ElementwiseGeFloatIntScalarModule_basic " ,
" ElementwiseGeFloatScalarModule_basic " ,
" ElementwiseGeIntScalarModule_basic " ,
" ElementwiseGeMixedIntScalarModule_basic " ,
2022-01-21 02:58:30 +08:00
" ElementwiseGtFloatScalarModule_basic " ,
" ElementwiseGtIntScalarModule_basic " ,
" ElementwiseGtMixed2ScalarModule_basic " ,
" ElementwiseGtFloatTensorModule_basic " ,
" ElementwiseGtIntTensorModule_basic " ,
" ElementwiseLtFloatScalarModule_basic " ,
" ElementwiseLtIntScalarModule_basic " ,
" ElementwiseLtDiffWidthScalarModule_basic " ,
" ElementwiseLtFloatTensorModule_basic " ,
" ElementwiseLtIntTensorModule_basic " ,
" ElementwiseEqFloatScalarModule_basic " ,
" ElementwiseEqIntScalarModule_basic " ,
2023-07-17 15:49:04 +08:00
" ElementwiseEqBoolScalarModule_basic " ,
2022-01-21 02:58:30 +08:00
" ElementwiseEqDiffWidthScalarModule_basic " ,
" ElementwiseEqFloatTensorModule_basic " ,
" ElementwiseEqIntTensorModule_basic " ,
2023-06-07 10:06:27 +08:00
" ElementwiseNeFloatScalarModule_basic " ,
" ElementwiseNeFloatTensorModule_basic " ,
" ElementwiseNeFloatTensorStaticModule_basic " ,
" ElementwiseNeIntTensorModule_basic " ,
" ElementwiseNeIntTensorStaticModule_basic " ,
2022-01-21 02:58:30 +08:00
" ElementwiseMulScalarModule_int " ,
" ElementwiseMulScalarModule_float " ,
" ElementwiseMulTensorIntModule_basic " ,
" ElementwiseDivScalarModule_basic " ,
2023-06-12 17:18:38 +08:00
" ElementwiseAtenDivIntScalarModule_basic " ,
2022-01-21 02:58:30 +08:00
" ElementwiseSubScalarFloatModule_basic " ,
" ElementwiseAddScalarFloatModule_basic " ,
2023-03-07 02:12:58 +08:00
" ElementwiseAddScalar_TensorLiteralInt32_Module_basic " ,
2022-01-21 02:58:30 +08:00
" ElementwiseMulScalarModule_float " ,
" ElementwiseCeilModule_basic " ,
" ElementwiseReciprocalModule_basic " ,
2023-06-07 10:06:27 +08:00
" ElementwiseIsnanModule_basic " ,
2022-01-21 02:58:30 +08:00
" TypePromotionAlphaWiderModule_basic " ,
2022-01-27 11:16:13 +08:00
" Conv2dWithPaddingDilationStrideStaticModule_basic " ,
2023-08-18 23:15:54 +08:00
" Conv2dWithPaddingDilationStrideStaticModule_depthwise " ,
" Conv2dWithPaddingDilationStrideStaticModule_depthwise_multiplier " ,
2022-02-25 03:41:55 +08:00
" BatchNorm1DModule_basic " ,
2022-07-16 00:35:59 +08:00
" BatchNorm1DWith2DInputModule_basic " ,
2022-02-25 03:41:55 +08:00
" BatchNorm2DModule_basic " ,
" BatchNorm3DModule_basic " ,
2023-05-18 00:04:40 +08:00
" BatchNorm1DStaticShapeModule_basic " ,
2022-01-29 13:38:56 +08:00
" FlattenStaticModule_basic " ,
" FlattenRank0Module_basic " ,
" ElementwiseFlattenBroadcastModule_basic " ,
2022-01-30 01:10:50 +08:00
" SquareModule_basic " ,
2022-02-01 05:34:09 +08:00
" MaxPool2dStaticModule_basic " ,
2023-03-22 01:17:39 +08:00
" MaxPool2dStaticCeilModeTrueModule_basic " ,
2022-08-31 03:44:00 +08:00
" ResNet18StaticModule_basic " ,
2022-11-23 02:37:28 +08:00
" ReduceAmaxKeepDim_basic " ,
2022-02-04 06:08:19 +08:00
" NativeLayerNormModule4D_basic " ,
" LayerNormNormalizeOverAllDimsModule_basic " ,
" PermuteModule_basic " ,
" PermuteNegativeIndexModule_basic " ,
2022-02-12 04:30:02 +08:00
" ElementwiseLog2Module_basic " ,
" Threshold1dIntI32Module_basic " ,
" Threshold1dFloatModule_basic " ,
" Threshold2dFloatModule_basic " ,
" Threshold3dFloatModule_basic " ,
" ElementwiseSubScalarIntModule_basic " ,
" ElementwiseAddScalarIntModule_basic " ,
" ElementwiseMulScalarModule_basic " ,
2022-02-17 01:53:51 +08:00
" ZerosModuleDefaultDtype_basic " ,
" ZerosModuleInt2D_basic " ,
" ZerosModuleInt3D_basic " ,
" ZerosModuleFloat2D_basic " ,
" ZerosModuleFloat3D_basic " ,
" ZerosModuleFalsePinMemory_basic " ,
" OnesModuleDefaultDtype_basic " ,
" OnesModuleInt_basic " ,
" OnesModuleFloat_basic " ,
" OnesModuleFalsePinMemory_basic " ,
2022-08-19 14:53:21 +08:00
" OnesModuleCPUDevice_basic " ,
2022-02-28 14:14:40 +08:00
" NewZerosModuleDefaultDtype_basic " ,
" NewZerosModuleInt2D_basic " ,
" NewZerosModuleInt3D_basic " ,
" NewZerosModuleFloat2D_basic " ,
" NewZerosModuleFloat3D_basic " ,
" NewZerosModuleFalsePinMemory_basic " ,
" NewOnesModuleDefaultDtype_basic " ,
" NewOnesModuleInt2D_basic " ,
" NewOnesModuleInt3D_basic " ,
" NewOnesModuleFloat2D_basic " ,
" NewOnesModuleFloat3D_basic " ,
" NewOnesModuleFalsePinMemory_basic " ,
2022-03-02 01:30:58 +08:00
" SiluModule_basic " ,
2022-02-17 00:34:03 +08:00
" DropoutEvalIntModule_basic " ,
" DropoutEvalFloatModule_basic " ,
2022-03-26 05:15:07 +08:00
" ContiguousModule_basic " ,
" DropoutModule_basic " ,
" ViewExpandModule_basic " ,
2022-06-23 06:58:40 +08:00
" ViewExpandOnesModule_basic " ,
2022-07-01 07:39:25 +08:00
" ViewExpandOnesBeforeAndAfterModule_basic " ,
" ViewExpandOnesMiddleModule_basic " ,
2022-07-21 05:35:51 +08:00
" ViewExpandCollapseModule_basic " ,
" ViewExpandCollapseWithOnesModule_basic " ,
2022-11-24 21:02:59 +08:00
" ViewCollapseInferredDimModule_basic " ,
" ViewExpandInferredDimModule_basic " ,
2023-01-21 03:56:26 +08:00
" ViewNegativeStaticModule_basic " ,
2022-03-26 05:15:07 +08:00
" ViewNoChangeStaticModule_basic " ,
" UnsafeViewExpandModule_basic " ,
" ReshapeCollapseModule_basic " ,
2022-03-31 08:00:55 +08:00
" ElementwiseGeluModule_basic " ,
" GeluBackwardModule_basic " ,
2022-04-04 00:19:01 +08:00
" ElementwiseNeIntScalarModule_basic " ,
2022-07-08 14:44:03 +08:00
" Convolution2DStaticModule_basic " ,
2022-04-16 03:37:57 +08:00
" ElementwiseNegModule_basic " ,
2022-04-19 21:47:47 +08:00
" TestMultipleTensorReturn_basic " ,
2022-09-13 06:59:12 +08:00
" TypeAsSameModule_basic " ,
2023-06-09 19:06:25 +08:00
" AdaptiveAvgPool2dNonUnitOutputSizeStaticModule_basic " ,
2022-05-19 20:35:59 +08:00
" AdaptiveAvgPool2dUnitOutputSizeStaticModule_basic " ,
2022-05-30 16:08:54 +08:00
" BaddbmmDynamicModule_basic " ,
" BaddbmmStaticModule_basic " ,
" BaddbmmWithAlphaBetaModule_basic " ,
" BaddbmmWithAlphaModule_basic " ,
" BaddbmmWithBetaModule_basic " ,
" BaddbmmBroadcast1DInputModule_basic " ,
" BaddbmmBroadcast2DInputModule_basic " ,
2022-06-03 20:38:59 +08:00
" NumpyTRank1Module_basic " ,
" NumpyTRank2Module_basic " ,
" NumpyTRankNStaticModule_basic " ,
" NumpyTRankNDynamicModule_basic " ,
2022-07-01 04:13:52 +08:00
" EmbeddingModuleI32Static_basic " ,
2023-07-21 14:54:19 +08:00
" EmbeddingModule1DIndices_basic " ,
2022-07-08 04:05:33 +08:00
" TModuleRank2_basic " ,
" TransposeIntModule_basic " ,
" TransposeIntNegDimsModule_basic " ,
2022-07-14 01:10:18 +08:00
" ArgmaxModule_keepDim " ,
" ArgmaxModule_with_dim " ,
2022-08-11 08:02:06 +08:00
" _LogSoftmaxModuleStable_basic " ,
2022-10-19 00:39:39 +08:00
" ElementwiseAtenWhereSelfModule_basic " ,
2022-11-15 01:09:15 +08:00
" ElementwiseUnsqueezeBroadcastModule_basic " ,
2023-02-11 05:58:39 +08:00
" MaskedFillScalarIntValueModule_basic " ,
2022-12-22 00:56:07 +08:00
" MaskedFillScalarIntValueStaticModule_basic " ,
" MaskedFillTensorIntValueStaticModule_basic " ,
2022-12-13 04:13:07 +08:00
" ElementwiseAddScalarInt64Module_basic " ,
" TensorLiteralModule_basic " ,
2023-03-30 22:08:20 +08:00
" NewZerosStaticModuleLayoutStrided_basic " ,
2022-12-13 04:13:07 +08:00
" TensorOpaqueLiteralModule_basic " ,
" TypePromotionDifferentCategoryModule_basic " ,
" TypePromotionSameCategoryDifferentWidthModule_basic " ,
2023-02-23 20:56:43 +08:00
" TypePromotionSameCategoryZeroRankWider_basic " ,
2022-12-13 04:13:07 +08:00
" TypePromotionZeroRankHigherCategoryModule_basic " ,
2022-12-22 03:04:07 +08:00
" GatherStaticModule_basic " ,
2023-01-20 13:19:00 +08:00
" IndexTensorStaticModule_basic " ,
2023-02-14 15:07:15 +08:00
" IndexTensorMultiIndexStaticModule_basic " ,
2023-04-19 04:36:57 +08:00
" ElementwiseWhereScalarModule_basic " ,
" FullLikeModuleFloat3DStatic_basic " ,
" FullModuleDefaultDtype_basic " ,
" FullModuleFloat3D_basic " ,
" MaskedFillScalarDefaultModule_basic " ,
" NumToTensorFloatModule_basic " ,
2022-09-21 06:07:46 +08:00
" LiftFreshCopyModule_basic " ,
2022-09-21 02:38:09 +08:00
" ReduceSumDimIntListKeepDimNegativeDimStaticModule_basic " ,
2022-10-18 12:22:53 +08:00
" ReduceSumDimIntListFloatModule_basic " ,
" ReduceSumDimIntListIntModule_basic " ,
" ReduceSumDimIntListKeepDimFloatModule_basic " ,
" ReduceSumDimIntListKeepDimIntModule_basic " ,
" ReduceSumFloatModule_basic " ,
" ReduceSumSignedIntModule_basic " ,
" ReduceSumUnsignedIntModule_basic " ,
2022-09-30 00:40:56 +08:00
" BroadcastToSameRankStaticModule_basic " ,
" BroadcastZeroRankInputStaticModule_basic " ,
2023-06-12 17:17:25 +08:00
" BroadcastListConstructWithMinusOneModule_basic " ,
2022-09-23 16:23:09 +08:00
" SliceStaticModule_basic " ,
2023-07-20 15:53:54 +08:00
" SliceOutOfUpperBoundIndexStaticModule_basic " ,
2022-09-30 22:33:41 +08:00
" ArangeStartStepIntModule_basic " ,
" ArangeDtypeFloatModule_basic " ,
" ArangeIntModule_basic " ,
" ArangeNegativeStartIntModule_basic " ,
" ArangeStartIntModule_basic " ,
" ArangeStartNegativeStepIntModule_basic " ,
" ArangeZeroElementOutputModule_basic " ,
2022-10-04 22:00:16 +08:00
" NumToTensorIntModule_basic " ,
2022-10-04 21:05:59 +08:00
" ToDtypeBoolLayoutNoneStaticModule_basic " ,
" ToCopyBoolDTypeStaticModule_basic " ,
" HardTanhIntModule_basic " ,
2022-10-20 19:02:09 +08:00
" AtenRoundIntModule_basic " ,
" MseLossNoReductionModule_basic " ,
2022-10-26 20:41:52 +08:00
" AddCMul_Module_basic " ,
" AddCDiv_Module_basic " ,
2022-12-01 23:46:17 +08:00
" TestF16Return_basic " ,
2023-01-04 00:30:16 +08:00
" ElementwiseLeakyReluModule_basic " ,
" ElementwiseLeakyReluStaticModule_basic " ,
" LeakyReluBackwardModule_basic " ,
" LeakyReluBackwardStaticModule_basic " ,
" ElementwiseRelu6Module_basic " ,
" HardTanhModule_basic " ,
" HardsigmoidModule_basic " ,
" HardsigmoidRandomModule_basic " ,
" HardswishModule_basic " ,
" HardswishRandomModule_basic " ,
2023-04-14 23:43:14 +08:00
" BucketizeTensorStaticFloatModule_basic " ,
" BucketizeTensorStaticModule_basic " ,
" ElementwiseLeFloatTensorModule_basic " ,
" ElementwiseLeIntTensorModule_basic " ,
2023-02-10 01:15:14 +08:00
" FullLikeModuleInt2DStatic_basic " ,
" FullModuleInt3D_basic " ,
" FullModuleFloat2D_basic " ,
2023-04-14 23:43:39 +08:00
" ElementwiseAbsModule_basic " ,
2023-03-15 23:42:15 +08:00
" RepeatModule_basic " ,
" ConstantPad2dStaticModule_basic " ,
" ConstantPadNdModule_basic " ,
" ConstantPadNdPartialStaticModule_basic " ,
" ConstantPadNdStaticModule_basic " ,
" PadModule_basic " ,
2023-03-24 08:55:58 +08:00
" PadWithNoneValModule_basic " ,
" ElementwiseRemainderScalarModule_Float_basic " ,
" ElementwiseRemainderScalarModule_Int_Float_basic " ,
2022-11-16 13:57:58 +08:00
" ElementwiseRemainderScalarModule_Int_basic " ,
2023-06-09 19:06:25 +08:00
" PrimsSqueezeModule_basic " ,
2022-11-16 13:57:58 +08:00
" PrimsSqueezeEmptyDimensionsModule_basic " ,
2023-01-02 22:34:39 +08:00
" MoveDimIntModule_basic " ,
" MoveDimIntNegativeIndexModule_basic " ,
2022-12-08 22:15:31 +08:00
" ElementwiseRemainderScalarModule_Int_basic " ,
" MoveDimIntModule_basic " ,
2023-04-10 11:50:26 +08:00
" PrimsViewOfModule_basic " ,
" PrimsViewOfZeroRankModule_basic " ,
2023-04-18 23:59:14 +08:00
" DetachModule_basic " ,
2023-06-01 11:38:50 +08:00
" ScalarTensorDefaultDtypeModule_basic " ,
" ScalarTensorFloat32Module_basic " ,
" ScalarTensorInt32Module_basic " ,
" ScalarTensorInt64Module_basic " ,
2023-05-19 10:07:58 +08:00
" UnbindIntListUnpack_Module_basic " ,
" UnbindIntGetItem_Module_basic " ,
2023-05-10 23:25:58 +08:00
" TensorsConcatStaticModule_basic " ,
" TensorsConcatNegativeDimStaticModule_basic " ,
2023-05-11 20:05:01 +08:00
" AtenComplex64Module_basic " ,
2023-07-14 14:23:10 +08:00
" ElementwiseSqrtIntModule_basic " ,
" ElementwiseSqrtModule_basic " ,
2023-05-24 03:43:33 +08:00
" SplitTensorGetItem_Module_basic " ,
2023-06-07 01:38:04 +08:00
" SplitTensorListUnpackModule_basic " ,
2023-07-20 15:53:54 +08:00
" SplitTensorNegativeDimModule_basic " ,
" SplitTensorLastSmallerModule_basic " ,
2023-09-04 09:59:26 +08:00
" SplitWithSizesListUnpackModule_basic " ,
2023-05-26 10:05:19 +08:00
" ChunkListUnpack_Module_basic " ,
" ChunkListUnpackUneven_Module_basic " ,
2023-07-18 22:32:26 +08:00
" TupleModule_basic " ,
2023-07-22 00:06:05 +08:00
" NumpyTRank0Module_basic " ,
" Permute0RankModule_basic " ,
2023-07-26 15:30:13 +08:00
" Add_Module_basic " ,
" SoftmaxIntModule_basic " ,
" SoftmaxIntNegDimModule_basic " ,
" _LogSoftmaxModule_basic " ,
" _SoftmaxModule_basic " ,
2021-10-26 02:43:21 +08:00
}
2022-06-10 03:56:01 +08:00
2023-07-13 21:07:54 +08:00
MAKE_FX_TOSA_PASS_SET = ( TOSA_PASS_SET | {
### Tests additionally passing in make_fx_tosa
" NativeGroupNormBackwardModule_basic " ,
2023-07-20 15:53:54 +08:00
" SliceWholeTensorModule_basic " ,
2023-07-13 21:07:54 +08:00
" TensorFloatModule_basic " ,
" TensorIntModule_basic " ,
[Torch Dialect] add support for adaptive_avgpool_1d (#2342)
* [MLIR][TORCH] Fix aten.cumsum lowering for int32 input (#2351)
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
update
parent 22e88d523b1970b2e904eb5421d49d987a3d255e
author jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114110 +0800
committer jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114119 +0800
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
add support for adaptive_pool_id
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update:
* update
---------
Co-authored-by: Vivek Khandelwal <vivekkhandelwal1424@gmail.com>
2023-08-05 07:48:09 +08:00
" AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic " ,
" AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic " ,
2023-07-13 21:07:54 +08:00
} ) - {
### Test failing in make_fx_tosa but not in tosa
# Dynamic shape, has extra unsupported broadcast ops
" Matmul_3d " ,
# failed to legalize operation 'torch.aten.max_pool2d_with_indices
" MaxPool2dEmptyStrideStaticModule_basic " ,
" MaxPool2dStaticCeilModeTrueModule_basic " ,
" MaxPool2dStaticModule_basic " ,
" ResNet18StaticModule_basic " ,
# Unimplemented operator 'aten._index_put_impl_.hacked_twin'
" IndexPutImpl1DFloatNonAccumulateModule_basic " ,
" IndexPutImpl1DIntNonAccumulateModule_basic " ,
2023-07-26 15:30:13 +08:00
# RuntimeError: The size of tensor a (7) must match the size of tensor b (3) at non-singleton dimension 1
" Add_Module_basic " ,
2023-07-13 21:07:54 +08:00
}
if torch_version_for_comparison ( ) < version . parse ( " 2.1.0.dev " ) :
MAKE_FX_TOSA_PASS_SET - = {
# 'tensor.expand_shape' op expected rank expansion, but found source rank 1 >= result rank 1
" ReshapeCollapseModule_basic " ,
2023-08-25 14:07:30 +08:00
# failed to lower torch.aten.empty.memory_format
" BatchNorm1DModule_basic " ,
" BatchNorm1DWith2DInputModule_basic " ,
" BatchNorm2DModule_basic " ,
" BatchNorm3DModule_basic " ,
" BatchNorm1DStaticShapeModule_basic " ,
2023-07-13 21:07:54 +08:00
}
2023-08-30 18:29:39 +08:00
LTC_CRASHING_SET = {
# TODO: update test to move all inputs to the lazy device. Otherwise test fails with:
# Check failed: lazy_tensor Input tensor is not a lazy tensor: CPUBoolType.
" HBC_basic " ,
}
2022-06-10 03:56:01 +08:00
LTC_XFAIL_SET = {
2022-07-14 01:28:05 +08:00
" _Convolution2DAllFalseModule_basic " ,
" _Convolution2DBenchmarkModule_basic " ,
" _Convolution2DCudnnModule_basic " ,
" _Convolution2DDeterministicModule_basic " ,
" _Convolution2DTF32Module_basic " ,
2022-08-22 13:44:44 +08:00
" _ConvolutionDeprecated2DAllFalseModule_basic " ,
" _ConvolutionDeprecated2DBenchmarkModule_basic " ,
2022-09-27 00:16:49 +08:00
" _ConvolutionDeprecated2DCudnnModule_basic " ,
2022-08-22 13:44:44 +08:00
" _ConvolutionDeprecated2DDeterministicModule_basic " ,
[Torch Dialect] add support for adaptive_avgpool_1d (#2342)
* [MLIR][TORCH] Fix aten.cumsum lowering for int32 input (#2351)
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
update
parent 22e88d523b1970b2e904eb5421d49d987a3d255e
author jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114110 +0800
committer jianzhe.xiao <jianzhe.xiao@bytedance.com> 1691114119 +0800
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op (#2340)
[Stablehlo] Add converter to stablehlo for aten.(Int,Float,Bool).Tensor op and configure crashing e2e sets for stablehlo backend.
update PyTorch version to 2.1.0.dev20230729 (#2354)
- torch version: 2.1.0.dev20230729
- torch commit hash: b638df0afb83572724032c824c64e481bb4499a0
- torchvision version: 0.16.0.dev20230729
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230730 (#2356)
- torch version: 2.1.0.dev20230730
- torch commit hash: 0ff243ff350268cc98fe03fa6364375ee2824742
- torchvision version: 0.16.0.dev20230730
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
update PyTorch version to 2.1.0.dev20230731 (#2359)
- torch version: 2.1.0.dev20230731
- torch commit hash: 6298ac688f8caafe30d71ff2ea2e20fbb32065c7
- torchvision version: 0.16.0.dev20230731
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
LTC->MLIR Debug Info support (#1922)
* LTC->MLIR Debug Info support
* SW-95317 Propagate Lazy->Jit->MLIR scope name.
* Enhance location information based on op names
Currently, the location information attached to the ops just considers
the filename, line number and column number. Attaching operation name
would help identify the type of computation by just looking at the
profile of execution.
* Update locations logic; updated debug-info.py test
* Use {scope}/{op_name} format to track names by default
---------
Co-authored-by: Gleb Kazantaev <gleb.kazantaev@cerebras.net>
Co-authored-by: Mark Browning <mark@cerebras.net>
Co-authored-by: Vimal Patel <vimal@polymagelabs.com>
build: update llvm tag to 41895843
Summary of changes:
- Update tags
llvm: 41895843b5915bb78e9d02aa711fa10f7174db43
mhlo: 4726d31f7025da66de0dea709bd56c462edb83c2
Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>
update PyTorch version to 2.1.0.dev20230802 (#2366)
- torch version: 2.1.0.dev20230802
- torch commit hash: c89b16917755c2abbef7b6420e340baf9ae8089e
- torchvision version: 0.16.0.dev20230802
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Change Python version from 3.10 to 3.11 in installation instructions (#2370)
Add CITATION file (#2371)
Add packaging as an install dependency (#2369)
Needed by `torch_mlir._version`. Resolves #2368.
[Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op (#2358)
* [Torch Dialect] emit aten.masked_scatter and aten.masked_scatter_ op
update PyTorch version to 2.1.0.dev20230803 (#2372)
- torch version: 2.1.0.dev20230803
- torch commit hash: f89c73be3a3e8274d025ac46a33a780853841c9e
- torchvision version: 0.16.0.dev20230803
Co-authored-by: Roll PyTorch Action <torch-mlir@users.noreply.github.com>
Prevent failed stable CI job from cancelling nightly jobs (#2373)
The CI jobs that use stable PyTorch are currently not required to pass
in order for a patch to get merged in `main`. This commit makes sure
that if a CI job for stable PyTorch fails, it does not cancel the
other required jobs.
[Torch Dialect] emit aten.tile op and decompose it into aten.repeat (#2355)
update
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update
update:
add support for adaptive_pool_id
update xfail sets
update xfail_sets
update
fix xfail_sets
update:
update:
* update
---------
Co-authored-by: Vivek Khandelwal <vivekkhandelwal1424@gmail.com>
2023-08-05 07:48:09 +08:00
" AdaptiveAvgPool1dNonUnitOutputSizeDynamicModule_basic " ,
" AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic " ,
2022-06-10 03:56:01 +08:00
" AdaptiveAvgPool2dNonUnitOutputSizeDynamicModule_basic " ,
" AdaptiveAvgPool2dNonUnitOutputSizeStaticModule_basic " ,
" AddIntModule_basic " ,
2023-01-18 02:14:14 +08:00
" AtenIntBoolOpModule_basic " ,
2022-06-10 03:56:01 +08:00
" BernoulliTensorModule_basic " ,
" BincountMinlengthModule_basic " ,
" BincountModule_basic " ,
" BincountStaticSizeModule_basic " ,
" BoolFloatFalseModule_basic " ,
" BoolFloatTrueModule_basic " ,
" BoolIntFalseModule_basic " ,
" BoolIntTrueModule_basic " ,
" CeilFloatModule_basic " ,
" DivFloatModule_basic " ,
" EqIntModule_basic " ,
" GeFloatIntModule_basic " ,
" GeFloatModule_basic " ,
2022-07-30 17:54:40 +08:00
" GeIntModule_basic " ,
2022-06-10 03:56:01 +08:00
" GtFloatIntModule_basic " ,
" GtIntModule_basic " ,
" IndexPutImpl1DFloatAccumulateModule_basic " ,
" IndexPutImpl1DFloatNonAccumulateModule_basic " ,
" IndexPutImpl1DIntAccumulateModule_basic " ,
" IndexPutImpl1DIntNonAccumulateModule_basic " ,
" IndexPutImpl2DFloatAccumulateModule_basic " ,
" IndexPutImpl2DFloatNonAccumulateModule_basic " ,
2022-12-08 13:46:54 +08:00
" IndexPutImpl2DIndexModule_basic " ,
2023-07-18 00:51:24 +08:00
" IndexPutImpl2DNoneIndexStaticModule_basic " ,
2022-06-10 03:56:01 +08:00
" IndexPutImpl3DFloatAccumulateModule_basic " ,
" IndexPutImpl3DFloatNonAccumulateModule_basic " ,
2022-12-08 13:46:54 +08:00
" IndexPutImplIndexWithNoneModule_basic " ,
2022-06-10 03:56:01 +08:00
" Matmul_dot " ,
" MulIntModule_basic " ,
2022-10-06 21:11:52 +08:00
" DivIntModule_basic " ,
2022-06-10 03:56:01 +08:00
" NeFloatIntModule_basic " ,
" NeIntModule_basic " ,
" QuantizedMLP_basic " ,
2022-08-24 08:36:05 +08:00
" RollModule_basic " ,
2022-06-10 03:56:01 +08:00
" ScalarImplicitFloatModule_basic " ,
" ScalarImplicitIntModule_basic " ,
" SliceEndSleStartModule_basic " ,
" SliceOutOfUpperBoundIndexModule_basic " ,
2023-07-20 15:53:54 +08:00
" SliceOutOfUpperBoundIndexStaticModule_basic " ,
2022-06-10 03:56:01 +08:00
" SliceStartEqEndModule_basic " ,
" SqrtIntModule_basic " ,
" SubFloatModule_basic " ,
" SubIntModule_basic " ,
2023-03-11 09:25:25 +08:00
" TensorsStackPromoteDTypeModule_basic " ,
2022-06-10 03:56:01 +08:00
" TensorToBoolZeroRank_basic " ,
" TensorToBool_basic " ,
" TensorToFloatZeroRank_basic " ,
" TensorToFloat_basic " ,
" TensorToIntZeroRank_basic " ,
" TensorToInt_basic " ,
" UniformModule_basic " ,
" UnsafeViewCollapseDynamicWithAtenSizeIntModule_basic " ,
" ViewCollapseDynamicWithAtenSizeIntModule_basic " ,
2022-08-02 06:23:45 +08:00
" AtenEmbeddingBagSumExample_basic " ,
2022-08-09 06:56:49 +08:00
" Aten_EmbeddingBagExample_basic " ,
2022-08-11 08:02:06 +08:00
" ElementwiseRemainderScalarModule_Int_Float_basic " ,
" ElementwiseRemainderScalarModule_Bool_basic " ,
2022-09-20 02:50:51 +08:00
" AtenIntTensorByteDtypeModule_basic " ,
" AtenIntTensorCharDtypeModule_basic " ,
2022-11-01 21:08:04 +08:00
" UpSampleNearest2dBackwardVec_basic " ,
" UpSampleNearest2dBackwardOutputSizeNone_basic " ,
2022-11-04 15:57:29 +08:00
" ConvolutionBackwardModule2D_basic " ,
" ConvolutionBackwardModule2DPadded_basic " ,
2022-11-15 22:39:40 +08:00
" VarMeanCorrectionModule_basic " ,
2022-11-21 16:38:47 +08:00
" VarMeanCorrectionNoneModule_basic " ,
" PrimsConvertElementTypeModule_basic " ,
2022-12-28 11:21:33 +08:00
" ElementwisePreluModule_basic " ,
2022-12-09 23:22:26 +08:00
" VarMeanBiasedModule_basic " ,
" VarMeanUnbiasedModule_basic " ,
2023-01-16 19:40:21 +08:00
" RandnLikeModule_basic " ,
" RandnLikeDtypeModule_basic " ,
2023-02-28 10:32:21 +08:00
" BernoulliFloatModule_basic " ,
" BernoulliModule_basic " ,
" BernoulliPModule_basic " ,
" DropoutTrainModule_basic " ,
2023-06-27 14:19:33 +08:00
" DropoutTrainStaticShapeModule_basic " ,
" NativeDropoutTrainModule_basic " ,
" NativeDropoutTrainStaticShapeModule_basic " ,
2023-02-28 10:32:21 +08:00
" StdCorrectionKeepDimModule_basic " ,
" StdCorrectionNoneModule_basic " ,
" VarCorrectionKeepDimModule_basic " ,
" VarCorrectionNoneModule_basic " ,
2022-11-16 13:57:58 +08:00
" AtenFloatScalarModule_basic " ,
" PrimsSqueezeModule_basic " ,
" PrimsSqueezeEmptyDimensionsModule_basic " ,
2023-04-10 11:50:26 +08:00
" PrimsViewOfModule_basic " ,
" PrimsViewOfZeroRankModule_basic " ,
2023-04-11 16:02:28 +08:00
" OneHotModule_basic " ,
2023-04-26 15:14:06 +08:00
" VarMeanDimModule_basic " ,
" VarMeanDimBiasedModule_basic " ,
2023-04-20 00:22:57 +08:00
" AtenComplexImagModule_basic " ,
" AtenComplexRealModule_basic " ,
2023-05-19 10:07:58 +08:00
" AtenComplexViewModule_basic " ,
2023-09-02 12:12:01 +08:00
" AtenRealView128Module_basic " ,
" AtenRealView64Module_basic " ,
2022-10-16 05:46:06 +08:00
" ScatterValueFloatModule_basic " ,
" ScatterValueIntModule_basic " ,
2023-06-27 10:55:28 +08:00
" UniformStaticShapeModule_basic " ,
2023-09-05 21:28:37 +08:00
" AtenEmbeddingBagStaticModule_basic " ,
2022-06-10 03:56:01 +08:00
}