[LinalgExt] Remove attention tile and decompose
Depends on: https://github.com/iree-org/iree/pull/17536
Good to have this cleanup, but IIRC @harsh-nod mentioned there are cases where we found regular, non FA faster, so tileAndDecomposeAttention may still be useful there?
Good to have this cleanup, but IIRC @harsh-nod mentioned there are cases where we found regular, non FA faster, so tileAndDecomposeAttention may still be useful there?
In those cases, we don't want to use flash attention decomposition. We need to implement AggregateOpInterface for attention op (we currently have it on online_attention op).
Already landed as part of https://github.com/iree-org/iree/commit/dd3f2a392819d121fa5329a1c591be06ae9e887a