torch-mlir/test/Conversion/TorchToLinalg
Stella Laurenzo 00efec0b73
[linalg] Implement strict mode lowering for aten.view. (#3319)
* Enables assume_strict_symbolic_shapes on fx_importer imported
programs, indicating strict shape semantics.
* Reworks the view->reshape lowering to take advantage of strict mode
and do one of:
  * Collapse to 0D
  * Flatten/Unflatten when there is an inferred dim.
  * Fallback to tensor.reshape
* Splits some test cases up and adds an attribute to control the old
pattern (so new corners can be tested in strict mode in isolation).
* Dynamic inferred mode needs upstream work to generalize expand_shape
(so that case is suppressed here).
* Deletes the assert from the existing tensor.reshape lowering if strict
shape mode is enabled (since the condition it is dynamically asserting
cannot happen).
2024-05-10 13:45:50 -07:00
..
basic.mlir Drop torch attributes at the end of backend conversion. (#2876) 2024-02-13 14:32:02 -08:00
broadcast.mlir [TorchToLinalg] Improve broadcast lowerings in strict symbolic modes (#2505) 2023-10-05 15:15:26 -04:00
convolution.mlir [MLIR][Torch] Do not convert bias tensor to element type if NoneType (#3072) 2024-04-02 14:19:26 +05:30
elementwise.mlir [torch] Lower `torch.aten.sinh` to `linalg` (#2662) 2023-12-18 09:15:12 -08:00
flatten.mlir Integrate llvm-project at dabdec1001dc368373dd581cf72f37a440873ce3 (#3300) 2024-05-08 14:43:06 -04:00
gridsampler.mlir [onnx] Gridsampler addition of nearest mode (#3320) 2024-05-10 11:42:10 -07:00
pooling.mlir MaxPool1d lowering to linalg (#3295) 2024-05-10 22:05:26 +05:30
resize.mlir OnnxToTorch lowering resize op (#3013) 2024-05-08 21:35:03 +00:00
sparse.mlir [torch-mlir][sparse] recognize to_dense primitive (#3308) 2024-05-08 22:50:17 -07:00
unsqueeze.mlir Integrate llvm-project at dabdec1001dc368373dd581cf72f37a440873ce3 (#3300) 2024-05-08 14:43:06 -04:00
view.mlir [linalg] Implement strict mode lowering for aten.view. (#3319) 2024-05-10 13:45:50 -07:00
view_strict.mlir [linalg] Implement strict mode lowering for aten.view. (#3319) 2024-05-10 13:45:50 -07:00