@@ -7,47 +7,46 @@ func.func private @test_depth_to_space_dynamic_dims(%arg0 : tensor<1x?x8x?xf32>)
7
7
%0 = " onnx.DepthToSpace" (%arg0 ) {blocksize = 4 : si64 } : (tensor <1 x?x8 x?xf32 >) -> tensor <1 x?x32 x?xf32 >
8
8
" func.return" (%0 ) : (tensor <1 x?x32 x?xf32 >) -> ()
9
9
10
- // CHECK-DAG: [[MAP_1_ :#.+]] = affine_map<()[s0] -> (s0 floordiv 16)>
11
- // CHECK-DAG: [[MAP_2_ :#.+]] = affine_map<()[s0] -> (s0 * 4)>
12
- // CHECK-LABEL: func private @test_depth_to_space_dynamic_dims
10
+ // CHECK-DAG: [[MAP_0_ :#.+]] = affine_map<()[s0] -> (s0 floordiv 16)>
11
+ // CHECK-DAG: [[MAP_1_ :#.+]] = affine_map<()[s0] -> (s0 * 4)>
12
+ // CHECK-LABEL: func.func private @test_depth_to_space_dynamic_dims
13
13
// CHECK-SAME: ([[PARAM_0_:%.+]]: memref<1x?x8x?xf32>) -> memref<1x?x32x?xf32> {
14
- // CHECK-DAG: [[VAR_c3_:%.+]] = arith.constant 3 : index
15
- // CHECK-DAG: [[VAR_c2_:%.+]] = arith.constant 2 : index
16
- // CHECK-DAG: [[VAR_c1_:%.+]] = arith.constant 1 : index
17
- // CHECK-DAG: [[VAR_c0_:%.+]] = arith.constant 0 : index
18
- // CHECK-DAG: [[VAR_c32_:%.+]] = arith.constant 32 : index
19
- // CHECK-DAG: [[VAR_c5_:%.+]] = arith.constant 5 : index
20
- // CHECK-DAG: [[VAR_c4_:%.+]] = arith.constant 4 : index
21
- // CHECK-DAG: [[VAR_c8_:%.+]] = arith.constant 8 : index
22
- // CHECK-DAG: [[VAR_5_:%.+]] = builtin.unrealized_conversion_cast [[PARAM_0_]] : memref<1x?x8x?xf32> to tensor<1x?x8x?xf32>
14
+ // CHECK-DAG: [[CST_5_:%.+]] = arith.constant 5 : index
15
+ // CHECK-DAG: [[CST_2_:%.+]] = arith.constant 2 : index
16
+ // CHECK-DAG: [[CST_0_:%.+]] = arith.constant 0 : index
17
+ // CHECK-DAG: [[CST_32_:%.+]] = arith.constant 32 : index
18
+ // CHECK-DAG: [[CST_4_:%.+]] = arith.constant 4 : index
19
+ // CHECK-DAG: [[CST_3_:%.+]] = arith.constant 3 : index
20
+ // CHECK-DAG: [[CST_8_:%.+]] = arith.constant 8 : index
21
+ // CHECK-DAG: [[CST_1_:%.+]] = arith.constant 1 : index
23
22
// CHECK-NOT: separator of consecutive DAGs
24
- // CHECK-DAG: [[VAR_0_ :%.+]] = memref.dim [[PARAM_0_]], [[VAR_c1_ ]] : memref<1x?x8x?xf32>
25
- // CHECK-DAG: [[VAR_1_ :%.+]] = memref.dim [[PARAM_0_]], [[VAR_c3_ ]] : memref<1x?x8x?xf32>
23
+ // CHECK-DAG: [[VAR_dim_ :%.+]] = memref.dim [[PARAM_0_]], [[CST_1_ ]] : memref<1x?x8x?xf32>
24
+ // CHECK-DAG: [[VAR_dim_0_ :%.+]] = memref.dim [[PARAM_0_]], [[CST_3_ ]] : memref<1x?x8x?xf32>
26
25
// CHECK-NOT: separator of consecutive DAGs
27
- // CHECK-DAG: [[VAR_2_ :%.+]] = affine.apply [[MAP_1_ ]](){{.}}[[VAR_0_ ]]{{.}}
28
- // CHECK-DAG: [[VAR_3_ :%.+]] = affine.apply [[MAP_2_ ]](){{.}}[[VAR_1_ ]]{{.}}
26
+ // CHECK-DAG: [[VAR_0_ :%.+]] = affine.apply [[MAP_0_ ]](){{.}}[[VAR_dim_ ]]{{.}}
27
+ // CHECK-DAG: [[VAR_1_ :%.+]] = affine.apply [[MAP_1_ ]](){{.}}[[VAR_dim_0_ ]]{{.}}
29
28
// CHECK-DAG: [[RES_:%.+]] = memref.alloc() {{.*}}: memref<6xindex>
30
- // CHECK: krnl.store [[VAR_c1_]], [[RES_]]{{.}}[[VAR_c0_]]{{.}} : memref<6xindex>
31
- // CHECK: krnl.store [[VAR_c4_]], [[RES_]]{{.}}[[VAR_c1_]]{{.}} : memref<6xindex>
32
- // CHECK: krnl.store [[VAR_c4_]], [[RES_]]{{.}}[[VAR_c2_]]{{.}} : memref<6xindex>
33
- // CHECK: krnl.store [[VAR_2_]], [[RES_]]{{.}}[[VAR_c3_]]{{.}} : memref<6xindex>
34
- // CHECK: krnl.store [[VAR_c8_]], [[RES_]]{{.}}[[VAR_c4_]]{{.}} : memref<6xindex>
35
- // CHECK: krnl.store [[VAR_1_]], [[RES_]]{{.}}[[VAR_c5_]]{{.}} : memref<6xindex>
36
- // CHECK-DAG: [[VAR_6_:%.+]] = builtin.unrealized_conversion_cast [[RES_]] : memref<6xindex> to tensor<6xi64>
37
- // CHECK: [[VAR_7_:%.+]] = "onnx.Reshape"([[VAR_5_]], [[VAR_6_]]) {allowzero = 0 : si64} : (tensor<1x?x8x?xf32>, tensor<6xi64>) -> tensor<?x?x?x?x?x?xf32>
38
- // CHECK: [[VAR_8_:%.+]] = builtin.unrealized_conversion_cast [[VAR_7_]] : tensor<?x?x?x?x?x?xf32> to memref<?x?x?x?x?x?xf32>
39
- // CHECK: [[VAR_9_:%.+]] = memref.cast [[VAR_8_]] : memref<?x?x?x?x?x?xf32> to memref<1x4x4x?x8x?xf32>
40
- // CHECK: [[VAR_10_:%.+]] = builtin.unrealized_conversion_cast [[VAR_9_]] : memref<1x4x4x?x8x?xf32> to tensor<1x4x4x?x8x?xf32>
41
- // CHECK-DAG: [[VAR_11_:%.+]] = "onnx.Transpose"([[VAR_10_]]) {perm = [0, 3, 4, 1, 5, 2]} : (tensor<1x4x4x?x8x?xf32>) -> tensor<1x?x8x4x?x4xf32>
29
+ // CHECK: krnl.store [[CST_1_]], [[RES_]]{{.}}[[CST_0_]]{{.}} : memref<6xindex>
30
+ // CHECK: krnl.store [[CST_4_]], [[RES_]]{{.}}[[CST_1_]]{{.}} : memref<6xindex>
31
+ // CHECK: krnl.store [[CST_4_]], [[RES_]]{{.}}[[CST_2_]]{{.}} : memref<6xindex>
32
+ // CHECK: krnl.store [[VAR_0_]], [[RES_]]{{.}}[[CST_3_]]{{.}} : memref<6xindex>
33
+ // CHECK: krnl.store [[CST_8_]], [[RES_]]{{.}}[[CST_4_]]{{.}} : memref<6xindex>
34
+ // CHECK: krnl.store [[VAR_dim_0_]], [[RES_]]{{.}}[[CST_5_]]{{.}} : memref<6xindex>
35
+ // CHECK-DAG: [[VAR_2_:%.+]] = builtin.unrealized_conversion_cast [[RES_]] : memref<6xindex> to tensor<6xi64>
36
+ // CHECK-DAG: [[VAR_3_:%.+]] = builtin.unrealized_conversion_cast [[PARAM_0_]] : memref<1x?x8x?xf32> to tensor<1x?x8x?xf32>
37
+ // CHECK: [[VAR_4_:%.+]] = "onnx.Reshape"([[VAR_3_]], [[VAR_2_]]) {allowzero = 0 : si64} : (tensor<1x?x8x?xf32>, tensor<6xi64>) -> tensor<?x?x?x?x?x?xf32>
38
+ // CHECK: [[VAR_5_:%.+]] = builtin.unrealized_conversion_cast [[VAR_4_]] : tensor<?x?x?x?x?x?xf32> to memref<?x?x?x?x?x?xf32>
39
+ // CHECK: [[VAR_cast_:%.+]] = memref.cast [[VAR_5_]] : memref<?x?x?x?x?x?xf32> to memref<1x4x4x?x8x?xf32>
40
+ // CHECK: [[VAR_6_:%.+]] = builtin.unrealized_conversion_cast [[VAR_cast_]] : memref<1x4x4x?x8x?xf32> to tensor<1x4x4x?x8x?xf32>
41
+ // CHECK-DAG: [[VAR_7_:%.+]] = "onnx.Transpose"([[VAR_6_]]) {perm = [0, 3, 4, 1, 5, 2]} : (tensor<1x4x4x?x8x?xf32>) -> tensor<1x?x8x4x?x4xf32>
42
42
// CHECK-DAG: [[RES_1_:%.+]] = memref.alloc() {{.*}}: memref<4xindex>
43
- // CHECK: krnl.store [[VAR_c1_]], [[RES_1_]]{{.}}[[VAR_c0_]]{{.}} : memref<4xindex>
44
- // CHECK: krnl.store [[VAR_2_]], [[RES_1_]]{{.}}[[VAR_c1_]]{{.}} : memref<4xindex>
45
- // CHECK: krnl.store [[VAR_c32_]], [[RES_1_]]{{.}}[[VAR_c2_]]{{.}} : memref<4xindex>
46
- // CHECK: krnl.store [[VAR_3_]], [[RES_1_]]{{.}}[[VAR_c3_]]{{.}} : memref<4xindex>
47
- // CHECK: [[VAR_13_:%.+]] = builtin.unrealized_conversion_cast [[RES_1_]] : memref<4xindex> to tensor<4xi64>
48
- // CHECK: [[VAR_14_:%.+]] = "onnx.Reshape"([[VAR_11_]], [[VAR_13_]]) {allowzero = 0 : si64} : (tensor<1x?x8x4x?x4xf32>, tensor<4xi64>) -> tensor<?x?x?x?xf32>
49
- // CHECK: [[VAR_15_:%.+]] = builtin.unrealized_conversion_cast [[VAR_14_]] : tensor<?x?x?x?xf32> to memref<?x?x?x?xf32>
50
- // CHECK: [[VAR_16_:%.+]] = memref.cast [[VAR_15_]] : memref<?x?x?x?xf32> to memref<1x?x32x?xf32>
51
- // CHECK: return [[VAR_16_]] : memref<1x?x32x?xf32>
52
- // CHECK: }
43
+ // CHECK: krnl.store [[CST_1_]], [[RES_1_]]{{.}}[[CST_0_]]{{.}} : memref<4xindex>
44
+ // CHECK: krnl.store [[VAR_0_]], [[RES_1_]]{{.}}[[CST_1_]]{{.}} : memref<4xindex>
45
+ // CHECK: krnl.store [[CST_32_]], [[RES_1_]]{{.}}[[CST_2_]]{{.}} : memref<4xindex>
46
+ // CHECK: krnl.store [[VAR_1_]], [[RES_1_]]{{.}}[[CST_3_]]{{.}} : memref<4xindex>
47
+ // CHECK: [[VAR_8_:%.+]] = builtin.unrealized_conversion_cast [[RES_1_]] : memref<4xindex> to tensor<4xi64>
48
+ // CHECK: [[VAR_9_:%.+]] = "onnx.Reshape"([[VAR_7_]], [[VAR_8_]]) {allowzero = 0 : si64} : (tensor<1x?x8x4x?x4xf32>, tensor<4xi64>) -> tensor<?x?x?x?xf32>
49
+ // CHECK: [[VAR_10_:%.+]] = builtin.unrealized_conversion_cast [[VAR_9_]] : tensor<?x?x?x?xf32> to memref<?x?x?x?xf32>
50
+ // CHECK: [[VAR_cast_2_:%.+]] = memref.cast [[VAR_10_]] : memref<?x?x?x?xf32> to memref<1x?x32x?xf32>
51
+ // CHECK: return [[VAR_cast_2_]] : memref<1x?x32x?xf32>
53
52
}
0 commit comments