@@ -8,7 +8,7 @@ func.func @test_memref(%offset_i: index, %offset_j: index) {
8
8
// CHECK: %[[C0:.*]] = arith.constant 0 : index
9
9
// CHECK: %[[LOCAL:.*]] = memref.alloc() : memref<64x64xf16, 3>
10
10
// CHECK: %[[MEM:.*]] = memref.alloc() : memref<64x128xf16>
11
- // CHECK: %[[MEM]][%arg0, %arg1], %[[LOCAL]][%[[C0]], %[[C0]]]
11
+ // CHECK: amdgpu.gather_to_lds %[[MEM]][%arg0, %arg1], %[[LOCAL]][%[[C0]], %[[C0]]]
12
12
// CHECK-SAME: vector<8xf16>, memref<64x128xf16>, memref<64x64xf16, 3>
13
13
14
14
%alloc = memref.alloc () : memref <64 x64 xf16 , #gpu_lds_addrspace >
@@ -37,7 +37,7 @@ func.func @subview_folding_offset(%offset_i: index, %offset_j: index) {
37
37
// CHECK: %[[IDX0:.*]] = affine.apply #[[MAP]]()[%[[ARG0]]]
38
38
// CHECK: %[[IDX1:.*]] = affine.apply #[[MAP1]]()[%[[ARG1]]]
39
39
40
- // CHECK: %[[MEM]][%[[IDX0]], %[[IDX1]]], %[[LOCAL]][%[[C0]], %[[C0]]]
40
+ // CHECK: amdgpu.gather_to_lds %[[MEM]][%[[IDX0]], %[[IDX1]]], %[[LOCAL]][%[[C0]], %[[C0]]]
41
41
// CHECK-SAME: vector<8xf16>, memref<64x128xf16>, memref<64x64xf16, 3>
42
42
43
43
%alloc = memref.alloc () : memref <64 x64 xf16 , #gpu_lds_addrspace >
0 commit comments