Han-Chung Wang
Han-Chung Wang
Here is the input IR: ```mlir hal.executable public @conv_2d_nchw_fchw_dispatch_1 { hal.executable.variant public @rocm_hsaco_fb target() { hal.executable.export public @conv_2d_nchw_fchw_dispatch_1_batch_matmul_64x968x4x320_f16xf16xf32 ordinal(0) layout(#hal.pipeline.layout) attributes {hal.interface.bindings = [#hal.interface.binding, #hal.interface.binding, #hal.interface.binding]} { ^bb0(%arg0: !hal.device): %x,...
closing the issue because there are no action items.
The two prototypes I have for pack/unpack shape inference are https://github.com/openxla/iree/pull/16629 and https://github.com/openxla/iree/pull/16664
note: the flatten is needed for LHS packing as well
I will use below three cases to drive the optimization work. ```mlir func.func @pack_i8(%source: tensor
> Didn't we already have a pass to make the innermost dimension larger? Yes, we have. The patterns will make the innermost dimension as larger as possible, i.e., it flattens...
https://github.com/iree-org/iree/pull/16456 should address the issue. I'll revisit how to land the PR
@Max191 I think you have some local patches and ideas that are required for mixed-types data-tiling work, could you add them to tasklist accordingly? @bjacob please help update this if...
For small tasks, adding a brief description to tasklist is good enough. For large tasks, it would be good if you can create an issue/epic. It's not necessary to do...
The pack issue is temporarily covered by ukernel. Let's focus on unpack kernel in the issue. There are transpose variants in the unpack op. We need to take it into...