mirror of https://github.com/llvm/torch-mlir
79928cd2dd
We plumb through e2e a fair number of interesting cases:
- unary, binary, ternary elementwise ops
- ops like `torch.aten.add.Tensor` that also take a scalar parameter
- static size-1 broadcasting
We allow the static size-1 broadcasting case, but emit a runtime error
in the case of dynamic size-1 broadcasting. This seems like a sweet spot
subset of things that can be lowered directly to linalg, while not being
overly constraining to users. This is consistent with what IREE is doing
for CHLO->Linalg lowering as well
([code](
|
||
---|---|---|
.. | ||
GlobalizeObjectGraph | ||
adjust-calling-conventions.mlir | ||
canonicalize.mlir | ||
finalizing-backend-type-conversion.mlir | ||
func-backend-type-conversion.mlir | ||
inline-global-slots.mlir | ||
invalid.mlir | ||
maximize-value-semantics.mlir | ||
ops.mlir | ||
prepare-for-globalize-object-graph.mlir | ||
reduce-op-variants.mlir | ||
refine-public-return.mlir | ||
refine-types.mlir | ||
verify-invariants-before-backend-lowering.mlir |