@@ -22,28 +22,28 @@ func @loop_nest_1d() {
22
22
%B = alloc () : memref <512 x f32 >
23
23
%F = alloc () : memref <256 x f32 , 2 >
24
24
// First DMA buffer.
25
- // CHECK: %{{.*}} = alloc() : memref<256xf32>
26
- // CHECK: %{{.*}} = alloc() : memref<256xf32, 2>
25
+ // CHECK: alloc() : memref<256xf32>
26
+ // CHECK: alloc() : memref<256xf32, 2>
27
27
// Tag for first DMA.
28
- // CHECK: %{{.*}} = alloc() : memref<1xi32>
28
+ // CHECK: alloc() : memref<1xi32>
29
29
// First DMA transfer.
30
30
// CHECK: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32>, memref<256xf32, 2>, memref<1xi32>
31
31
// CHECK: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
32
32
// Second DMA buffer.
33
- // CHECK: %{{.*}} = alloc() : memref<256xf32, 2>
33
+ // CHECK: alloc() : memref<256xf32, 2>
34
34
// Tag for second DMA.
35
- // CHECK: %{{.*}} = alloc() : memref<1xi32>
35
+ // CHECK: alloc() : memref<1xi32>
36
36
// Second DMA transfer.
37
37
// CHECK: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<512xf32>, memref<256xf32, 2>, memref<1xi32>
38
38
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
39
39
// CHECK: affine.for %{{.*}} = 0 to 256 {
40
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}}] : memref<256xf32, 2>
41
- // CHECK: %{{.*}} = affine.apply [[MAP_PLUS_256]](%{{.*}})
40
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}}] : memref<256xf32, 2>
41
+ // CHECK: affine.apply [[MAP_PLUS_256]](%{{.*}})
42
42
// Buffer for '%{{.*}}' in faster memref space is smaller size: 256xf32
43
43
// Affine map for 'affine.load %{{.*}}' is composed: %{{.*}} + 256 - 256 = %{{.*}}.
44
44
// CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}}] : memref<256xf32, 2>
45
45
// Already in faster memory space.
46
- // CHECK: %{{.*}} = affine.load %{{.*}}[%{{.*}}] : memref<256xf32, 2>
46
+ // CHECK: affine.load %{{.*}}[%{{.*}}] : memref<256xf32, 2>
47
47
// CHECK-NEXT: }
48
48
// CHECK-NEXT: dealloc %{{.*}} : memref<1xi32>
49
49
// CHECK-NEXT: dealloc %{{.*}} : memref<256xf32, 2>
@@ -83,20 +83,20 @@ func @loop_nest_1d() {
83
83
// CHECK-NEXT: affine.for %{{.*}} = 0 to 32 {
84
84
// CHECK-NEXT: affine.for %{{.*}} = 0 to 32 {
85
85
// CHECK-NEXT: affine.for %{{.*}} = 0 to 16 {
86
- // CHECK-NEXT: %{{.*}} = affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
86
+ // CHECK-NEXT: affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
87
87
// CHECK-NEXT: %{{.*}} = affine.load [[BUFB]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
88
88
// CHECK-NEXT: "foo"(%{{.*}}) : (f32) -> ()
89
89
// CHECK-NEXT: }
90
90
// CHECK-NEXT: affine.for %{{.*}} = 0 to 16 {
91
- // CHECK-NEXT: %{{.*}} = affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
92
- // CHECK-NEXT: %{{.*}} = affine.load [[BUFA]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
91
+ // CHECK-NEXT: affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
92
+ // CHECK-NEXT: affine.load [[BUFA]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
93
93
// CHECK-NEXT: "bar"(%{{.*}}) : (f32) -> ()
94
94
// CHECK-NEXT: }
95
95
// CHECK-NEXT: affine.for %{{.*}} = 0 to 16 {
96
- // CHECK-NEXT: %{{.*}} = "abc_compute"() : () -> f32
97
- // CHECK-NEXT: %{{.*}} = affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
98
- // CHECK-NEXT: %{{.*}} = affine.load [[BUFC]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
99
- // CHECK-NEXT: %{{.*}} = "addf32"(%{{.*}}, %{{.*}}) : (f32, f32) -> f32
96
+ // CHECK-NEXT: "abc_compute"() : () -> f32
97
+ // CHECK-NEXT: affine.apply #map{{[0-9]+}}(%{{.*}}, %{{.*}})
98
+ // CHECK-NEXT: affine.load [[BUFC]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
99
+ // CHECK-NEXT: "addf32"(%{{.*}}, %{{.*}}) : (f32, f32) -> f32
100
100
// CHECK-NEXT: affine.store %{{.*}}, [[BUFC]][%{{.*}} * 16 + %{{.*}}, %{{.*}}] : memref<512x32xf32, 2>
101
101
// CHECK-NEXT: }
102
102
// CHECK-NEXT: "foobar"() : () -> ()
@@ -153,11 +153,11 @@ func @loop_nest_high_d(%A: memref<512 x 32 x f32>,
153
153
// region within a 256 x 8 memref.
154
154
//
155
155
// CHECK-LABEL: func @loop_nest_modulo() {
156
- // CHECK: %{{.*}} = alloc() : memref<256x8xf32>
156
+ // CHECK: alloc() : memref<256x8xf32>
157
157
// CHECK-NEXT: affine.for %{{.*}} = 0 to 32 step 4 {
158
- // CHECK-NEXT: %{{.*}} = affine.apply #map{{[0-9]+}}(%{{.*}})
159
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1x2xf32, 2>
160
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
158
+ // CHECK-NEXT: affine.apply #map{{[0-9]+}}(%{{.*}})
159
+ // CHECK-NEXT: alloc() : memref<1x2xf32, 2>
160
+ // CHECK-NEXT: alloc() : memref<1xi32>
161
161
// Composition of the affine map for '%{{.*}}' causes '%{{.*}}' to be added as a symbol.
162
162
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}, 0], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256x8xf32>, memref<1x2xf32, 2>, memref<1xi32>
163
163
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
@@ -191,16 +191,16 @@ func @loop_nest_tiled() -> memref<256x1024xf32> {
191
191
%0 = alloc () : memref <256 x1024 xf32 >
192
192
affine.for %i0 = 0 to 256 step 32 {
193
193
affine.for %i1 = 0 to 1024 step 32 {
194
- // CHECK: %{{.*}} = alloc() : memref<32x32xf32, 2>
195
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
194
+ // CHECK: alloc() : memref<32x32xf32, 2>
195
+ // CHECK-NEXT: alloc() : memref<1xi32>
196
196
// Strided DMA here: 32 x 32 tile in a 256 x 1024 memref.
197
197
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}, %{{.*}}, %{{.*}} : memref<256x1024xf32>, memref<32x32xf32, 2>, memref<1xi32>
198
198
// CHECK-NEXT: affine.dma_wait
199
199
// CHECK-NEXT: affine.for %{{.*}} = #map
200
200
// CHECK-NEXT: affine.for %{{.*}} = #map
201
201
affine.for %i2 = affine_map <(d0 ) -> (d0 )>(%i0 ) to affine_map <(d0 ) -> (d0 + 32 )>(%i0 ) {
202
202
affine.for %i3 = affine_map <(d0 ) -> (d0 )>(%i1 ) to affine_map <(d0 ) -> (d0 + 32 )>(%i1 ) {
203
- // CHECK: %{{.*}} = affine.load %{{.*}}[-%{{.*}} + %{{.*}}, -%{{.*}} + %{{.*}}] : memref<32x32xf32, 2>
203
+ // CHECK: affine.load %{{.*}}[-%{{.*}} + %{{.*}}, -%{{.*}} + %{{.*}}] : memref<32x32xf32, 2>
204
204
%1 = affine.load %0 [%i2 , %i3 ] : memref <256 x1024 xf32 >
205
205
} // CHECK-NEXT: }
206
206
}
@@ -215,14 +215,14 @@ func @loop_nest_tiled() -> memref<256x1024xf32> {
215
215
func @dma_constant_dim_access (%A : memref <100 x100 xf32 >) {
216
216
%one = constant 1 : index
217
217
%N = constant 100 : index
218
- // CHECK: %{{.*}} = alloc() : memref<1x100xf32, 2>
219
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
218
+ // CHECK: alloc() : memref<1x100xf32, 2>
219
+ // CHECK-NEXT: alloc() : memref<1xi32>
220
220
// No strided DMA needed here.
221
221
// CHECK: affine.dma_start %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<100x100xf32>, memref<1x100xf32, 2>,
222
222
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
223
223
affine.for %i = 0 to 100 {
224
224
affine.for %j = 0 to affine_map <()[s0 ] -> (s0 )> ()[%N ] {
225
- // CHECK: %{{.*}} = affine.load %{{.*}}[0, %{{.*}}] : memref<1x100xf32, 2>
225
+ // CHECK: affine.load %{{.*}}[0, %{{.*}}] : memref<1x100xf32, 2>
226
226
affine.load %A [%one , %j ] : memref <100 x 100 x f32 >
227
227
}
228
228
}
@@ -243,14 +243,14 @@ func @dma_with_symbolic_accesses(%A : memref<100x100xf32>, %M : index) {
243
243
}
244
244
}
245
245
return
246
- // CHECK: %{{.*}} = alloc() : memref<100x100xf32, 2>
247
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
246
+ // CHECK: alloc() : memref<100x100xf32, 2>
247
+ // CHECK-NEXT: alloc() : memref<1xi32>
248
248
// CHECK-NEXT: affine.dma_start %{{.*}}[0, symbol(%{{.*}}) + 9], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}
249
249
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}}
250
250
// CHECK-NEXT: affine.for %{{.*}} = 0 to 100 {
251
251
// CHECK-NEXT: affine.for %{{.*}} = 0 to 100 {
252
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_SYM_SHIFT]](%{{.*}}, %{{.*}})[%{{.*}}, %{{.*}}]
253
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}}, %{{.*}}] : memref<100x100xf32, 2>
252
+ // CHECK-NEXT: affine.apply [[MAP_SYM_SHIFT]](%{{.*}}, %{{.*}})[%{{.*}}, %{{.*}}]
253
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}}, %{{.*}}] : memref<100x100xf32, 2>
254
254
// CHECK-NEXT: }
255
255
// CHECK-NEXT: }
256
256
// CHECK: return
@@ -263,8 +263,8 @@ func @dma_with_symbolic_loop_bounds(%A : memref<100x100xf32>, %M : index, %N: in
263
263
%K = constant 9 : index
264
264
// The buffer size can't be bound by a constant smaller than the original
265
265
// memref size; so the DMA buffer is the entire 100x100.
266
- // CHECK: %{{.*}} = alloc() : memref<100x100xf32, 2>
267
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
266
+ // CHECK: alloc() : memref<100x100xf32, 2>
267
+ // CHECK-NEXT: alloc() : memref<1xi32>
268
268
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<100x100xf32>, memref<100x100xf32, 2>, memref<1xi32>
269
269
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
270
270
affine.for %i = 0 to 100 {
@@ -286,7 +286,7 @@ func @dma_unknown_size(%arg0: memref<?x?xf32>) {
286
286
affine.for %j = 0 to %N {
287
287
// If this loop nest isn't tiled, the access requires a non-constant DMA
288
288
// size -- not yet implemented.
289
- // CHECK: %{{.*}} = affine.load %{{.*}}[%{{.*}}, %{{.*}}] : memref<?x?xf32>
289
+ // CHECK: affine.load %{{.*}}[%{{.*}}, %{{.*}}] : memref<?x?xf32>
290
290
affine.load %arg0 [%i , %j ] : memref <? x ? x f32 >
291
291
// expected-error@-6 {{copy generation failed for one or more memref's in this block}}
292
292
}
@@ -306,7 +306,7 @@ func @dma_memref_3d(%arg0: memref<1024x1024x1024xf32>) {
306
306
%idz = affine.apply affine_map <(d0 ) -> (d0 mod 128 )>(%k )
307
307
// DMA with nested striding (or emulating with loop around strided DMA)
308
308
// not yet implemented.
309
- // CHECK: %{{.*}} = affine.load %{{.*}}[%{{.*}}, %{{.*}}, %{{.*}}] : memref<1024x1024x1024xf32>
309
+ // CHECK: affine.load %{{.*}}[%{{.*}}, %{{.*}}, %{{.*}}] : memref<1024x1024x1024xf32>
310
310
%v = affine.load %arg0 [%idx , %idy , %idz ] : memref <1024 x 1024 x 1024 x f32 >
311
311
// expected-error@-10 {{copy generation failed for one or more memref's in this block}}
312
312
}
@@ -351,22 +351,22 @@ func @multi_load_store_union() {
351
351
}
352
352
return
353
353
}
354
- // CHECK: %{{.*}} = alloc() : memref<512x512xf32>
355
- // CHECK-NEXT: %{{.*}} = alloc() : memref<382x446xf32, 2>
356
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
354
+ // CHECK: alloc() : memref<512x512xf32>
355
+ // CHECK-NEXT: alloc() : memref<382x446xf32, 2>
356
+ // CHECK-NEXT: alloc() : memref<1xi32>
357
357
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}, %{{.*}}, %{{.*}} : memref<512x512xf32>, memref<382x446xf32, 2>, memref<1xi32>
358
358
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
359
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
359
+ // CHECK-NEXT: alloc() : memref<1xi32>
360
360
// CHECK-NEXT: affine.for %{{.*}} = 0 to 256 {
361
361
// CHECK-NEXT: affine.for %{{.*}} = 0 to 256 {
362
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_64]](%{{.*}})
363
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_128]](%{{.*}})
364
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_2]](%{{.*}})
365
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_2]](%{{.*}})
366
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}}, %{{.*}} + 126] : memref<382x446xf32, 2>
367
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}} + 62, %{{.*}}] : memref<382x446xf32, 2>
368
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_128]](%{{.*}})
369
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_PLUS_192]](%{{.*}})
362
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_64]](%{{.*}})
363
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_128]](%{{.*}})
364
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_2]](%{{.*}})
365
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_2]](%{{.*}})
366
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}}, %{{.*}} + 126] : memref<382x446xf32, 2>
367
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}} + 62, %{{.*}}] : memref<382x446xf32, 2>
368
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_128]](%{{.*}})
369
+ // CHECK-NEXT: affine.apply [[MAP_PLUS_192]](%{{.*}})
370
370
// CHECK-NEXT: affine.store %{{.*}}, %{{.*}}[%{{.*}}, %{{.*}} + 190] : memref<382x446xf32, 2>
371
371
// CHECK-NEXT: affine.store %{{.*}}, %{{.*}}[%{{.*}} + 126, %{{.*}}] : memref<382x446xf32, 2>
372
372
// CHECK-NEXT: }
@@ -396,29 +396,29 @@ func @dma_loop_straightline_interspersed() {
396
396
}
397
397
// There are three regions here - the 'load' preceding the loop, the loop
398
398
// itself, and the operations appearing after the loop.
399
- // CHECK: %{{.*}} = alloc() : memref<256xf32>
400
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xf32, 2>
401
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
399
+ // CHECK: alloc() : memref<256xf32>
400
+ // CHECK-NEXT: alloc() : memref<1xf32, 2>
401
+ // CHECK-NEXT: alloc() : memref<1xi32>
402
402
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32>, memref<1xf32, 2>, memref<1xi32>
403
403
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
404
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[0] : memref<1xf32, 2>
404
+ // CHECK-NEXT: affine.load %{{.*}}[0] : memref<1xf32, 2>
405
405
// CHECK-NEXT: dealloc %{{.*}} : memref<1xi32>
406
406
// CHECK-NEXT: dealloc %{{.*}} : memref<1xf32, 2>
407
- // CHECK-NEXT: %{{.*}} = alloc() : memref<254xf32, 2>
408
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
407
+ // CHECK-NEXT: alloc() : memref<254xf32, 2>
408
+ // CHECK-NEXT: alloc() : memref<1xi32>
409
409
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32>, memref<254xf32, 2>, memref<1xi32>
410
410
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
411
411
// CHECK-NEXT: affine.for %{{.*}} = 1 to 255 {
412
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}} - 1] : memref<254xf32, 2>
412
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}} - 1] : memref<254xf32, 2>
413
413
// CHECK-NEXT: }
414
414
// CHECK-NEXT: dealloc %{{.*}} : memref<1xi32>
415
415
// CHECK-NEXT: dealloc %{{.*}} : memref<254xf32, 2>
416
- // CHECK-NEXT: %{{.*}} = alloc() : memref<256xf32, 2>
417
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
416
+ // CHECK-NEXT: alloc() : memref<256xf32, 2>
417
+ // CHECK-NEXT: alloc() : memref<1xi32>
418
418
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32>, memref<256xf32, 2>, memref<1xi32>
419
419
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
420
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
421
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[255] : memref<256xf32, 2>
420
+ // CHECK-NEXT: alloc() : memref<1xi32>
421
+ // CHECK-NEXT: affine.load %{{.*}}[255] : memref<256xf32, 2>
422
422
// CHECK-NEXT: affine.store %{{.*}}, %{{.*}}[0] : memref<256xf32, 2>
423
423
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32, 2>, memref<256xf32>, memref<1xi32>
424
424
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
@@ -449,9 +449,9 @@ func @dma_mixed_loop_blocks() {
449
449
// CHECK: affine.dma_start [[MEM]][%{{.*}}, %{{.*}}], [[BUF]][%{{.*}}, %{{.*}}], [[TAG]][%{{.*}}], %{{.*}} : memref<256x256xvector<8xf32>>, memref<256x256xvector<8xf32>, 2>, memref<1xi32>
450
450
// CHECK-NEXT: affine.dma_wait [[TAG]][%{{.*}}], %{{.*}} : memref<1xi32>
451
451
// CHECK-NEXT: affine.for %{{.*}} = 0 to 256 {
452
- // CHECK: %{{.*}} = affine.load [[BUF]][0, 0] : memref<256x256xvector<8xf32>, 2>
452
+ // CHECK: affine.load [[BUF]][0, 0] : memref<256x256xvector<8xf32>, 2>
453
453
// CHECK: affine.for %{{.*}} = 0 to 256 {
454
- // CHECK-NEXT: %{{.*}} = affine.load [[BUF]][%{{.*}}, %{{.*}}] : memref<256x256xvector<8xf32>, 2>
454
+ // CHECK-NEXT: affine.load [[BUF]][%{{.*}}, %{{.*}}] : memref<256x256xvector<8xf32>, 2>
455
455
456
456
// -----
457
457
@@ -469,7 +469,7 @@ func @relative_loop_bounds(%arg0: memref<1027xf32>) {
469
469
// CHECK-NEXT: [[MEM:%[0-9]+]] = alloc() : memref<1xi32>
470
470
// CHECK-NEXT: affine.for %{{.*}} = 0 to 1024 {
471
471
// CHECK-NEXT: affine.for %{{.*}} = {{#map[0-9]+}}(%{{.*}}) to {{#map[0-9]+}}(%{{.*}}) {
472
- // CHECK-NEXT: %{{.*}} = constant 0.000000e+00 : f32
472
+ // CHECK-NEXT: constant 0.000000e+00 : f32
473
473
// CHECK-NEXT: affine.store %{{.*}}, [[BUF]][%{{.*}}] : memref<1027xf32, 2>
474
474
// CHECK-NEXT: }
475
475
// CHECK-NEXT: }
@@ -496,16 +496,16 @@ func @test_read_write_region_union() {
496
496
return
497
497
}
498
498
499
- // CHECK: %{{.*}} = alloc() : memref<256xf32>
500
- // CHECK-NEXT: %{{.*}} = alloc() : memref<85xf32, 2>
501
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
499
+ // CHECK: alloc() : memref<256xf32>
500
+ // CHECK-NEXT: alloc() : memref<85xf32, 2>
501
+ // CHECK-NEXT: alloc() : memref<1xi32>
502
502
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<256xf32>, memref<85xf32, 2>, memref<1xi32>
503
503
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
504
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
504
+ // CHECK-NEXT: alloc() : memref<1xi32>
505
505
// CHECK-NEXT: affine.for %{{.*}} = 0 to 10 {
506
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_READ_OFFSET]](%{{.*}})
507
- // CHECK-NEXT: %{{.*}} = affine.apply [[MAP_WRITE_OFFSET]](%{{.*}})
508
- // CHECK-NEXT: %{{.*}} = affine.load %{{.*}}[%{{.*}} + 75] : memref<85xf32, 2>
506
+ // CHECK-NEXT: affine.apply [[MAP_READ_OFFSET]](%{{.*}})
507
+ // CHECK-NEXT: affine.apply [[MAP_WRITE_OFFSET]](%{{.*}})
508
+ // CHECK-NEXT: affine.load %{{.*}}[%{{.*}} + 75] : memref<85xf32, 2>
509
509
// CHECK-NEXT: affine.store %{{.*}}, %{{.*}}[%{{.*}}] : memref<85xf32, 2>
510
510
// CHECK-NEXT: }
511
511
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<85xf32, 2>, memref<256xf32>, memref<1xi32>
@@ -567,8 +567,8 @@ func @test_memref_bounds(%arg0: memref<4x4x16x1xvector<8x128xf32>>, %arg1: memre
567
567
return %arg1 , %arg2 : memref <144 x9 xvector <8 x128 xf32 >>, memref <2 xvector <8 x128 xf32 >>
568
568
}
569
569
570
- // CHECK: %{{.*}} = alloc() : memref<4x4x16x1xvector<8x128xf32>, 2>
571
- // CHECK-NEXT: %{{.*}} = alloc() : memref<1xi32>
570
+ // CHECK: alloc() : memref<4x4x16x1xvector<8x128xf32>, 2>
571
+ // CHECK-NEXT: alloc() : memref<1xi32>
572
572
// CHECK-NEXT: affine.dma_start %{{.*}}[%{{.*}}, %{{.*}}, %{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}, %{{.*}}, %{{.*}}, %{{.*}}], %{{.*}}[%{{.*}}], %{{.*}} : memref<4x4x16x1xvector<8x128xf32>>, memref<4x4x16x1xvector<8x128xf32>, 2>, memref<1xi32>
573
573
// CHECK-NEXT: affine.dma_wait %{{.*}}[%{{.*}}], %{{.*}} : memref<1xi32>
574
574
0 commit comments