mirror of https://github.com/llvm/torch-mlir
25738b8c19
Attention often broadcasts a mask across the batch dimension as masking is usually performed the same across attention heads. Added this materialization to the mask dimensions optionally. |
||
---|---|---|
.. | ||
TorchConversionToMLProgram | ||
TorchOnnxToTorch | ||
TorchToArith | ||
TorchToLinalg | ||
TorchToSCF | ||
TorchToStablehlo | ||
TorchToTMTensor | ||
TorchToTensor | ||
TorchToTosa | ||
Utils | ||
CMakeLists.txt | ||
PassDetail.h | ||
Passes.cpp |