@@ -59,7 +59,7 @@ func.func @test_elu(%arg0 : tensor<20x40xf32>) -> tensor<20x40xf32> {
59
59
// CHECK-DAG: [[VAR_3_:%.+]] = stablehlo.exponential [[PARAM_0_]] : tensor<20x40xf32>
60
60
// CHECK: [[VAR_4_:%.+]] = stablehlo.subtract [[VAR_3_]], [[VAR_2_]] : tensor<20x40xf32>
61
61
// CHECK-DAG: [[VAR_5_:%.+]] = stablehlo.multiply [[VAR_1_]], [[VAR_4_]] : tensor<20x40xf32>
62
- // CHECK-DAG: [[VAR_6_:%.+]] = stablehlo.compare GE, [[PARAM_0_]], [[VAR_0_]], NOTYPE : (tensor<20x40xf32>, tensor<20x40xf32>) -> tensor<20x40xi1>
62
+ // CHECK-DAG: [[VAR_6_:%.+]] = stablehlo.compare GE, [[PARAM_0_]], [[VAR_0_]] : (tensor<20x40xf32>, tensor<20x40xf32>) -> tensor<20x40xi1>
63
63
// CHECK: [[VAR_7_:%.+]] = stablehlo.select [[VAR_6_]], [[PARAM_0_]], [[VAR_5_]] : tensor<20x40xi1>, tensor<20x40xf32>
64
64
// CHECK: return [[VAR_7_]] : tensor<20x40xf32>
65
65
// CHECK: }
@@ -287,7 +287,7 @@ func.func @test_less(%arg0: tensor<3x4x5xf32>, %arg1: tensor<3x4x5xf32>) -> tens
287
287
// CHECK: [[VAR_0_:%.+]] = shape.const_shape [3, 4, 5] : tensor<3xindex>
288
288
// CHECK-DAG: [[VAR_1_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_0_]], [[VAR_0_]], dims = [0, 1, 2] : (tensor<3x4x5xf32>, tensor<3xindex>) -> tensor<3x4x5xf32>
289
289
// CHECK-DAG: [[VAR_2_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_1_]], [[VAR_0_]], dims = [0, 1, 2] : (tensor<3x4x5xf32>, tensor<3xindex>) -> tensor<3x4x5xf32>
290
- // CHECK: [[VAR_3_:%.+]] = stablehlo.compare LT, [[VAR_1_]], [[VAR_2_]], NOTYPE : (tensor<3x4x5xf32>, tensor<3x4x5xf32>) -> tensor<3x4x5xi1>
290
+ // CHECK: [[VAR_3_:%.+]] = stablehlo.compare LT, [[VAR_1_]], [[VAR_2_]] : (tensor<3x4x5xf32>, tensor<3x4x5xf32>) -> tensor<3x4x5xi1>
291
291
// CHECK: return [[VAR_3_]] : tensor<3x4x5xi1>
292
292
// CHECK: }
293
293
@@ -305,7 +305,7 @@ func.func @test_binary_elementwise_op_template_unknown_dims(%arg0: tensor<?x4x5x
305
305
// CHECK: [[VAR_2_:%.+]] = shape.broadcast [[VAR_0_]], [[VAR_1_]] : tensor<3xindex>, tensor<3xindex> -> tensor<3xindex>
306
306
// CHECK-DAG: [[VAR_3_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_0_]], [[VAR_2_]], dims = [0, 1, 2] : (tensor<?x4x5xf32>, tensor<3xindex>) -> tensor<?x4x5xf32>
307
307
// CHECK-DAG: [[VAR_4_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_1_]], [[VAR_2_]], dims = [0, 1, 2] : (tensor<1x?x1xf32>, tensor<3xindex>) -> tensor<?x4x5xf32>
308
- // CHECK: [[VAR_5_:%.+]] = stablehlo.compare LT, [[VAR_3_]], [[VAR_4_]], NOTYPE : (tensor<?x4x5xf32>, tensor<?x4x5xf32>) -> tensor<?x4x5xi1>
308
+ // CHECK: [[VAR_5_:%.+]] = stablehlo.compare LT, [[VAR_3_]], [[VAR_4_]] : (tensor<?x4x5xf32>, tensor<?x4x5xf32>) -> tensor<?x4x5xi1>
309
309
// CHECK: return [[VAR_5_]] : tensor<?x4x5xi1>
310
310
// CHECK: }
311
311
@@ -323,7 +323,7 @@ func.func @test_less_unknown_dims_2(%arg0: tensor<?x?x5xf32>, %arg1: tensor<?x4x
323
323
// CHECK: [[VAR_2_:%.+]] = shape.broadcast [[VAR_0_]], [[VAR_1_]] : tensor<3xindex>, tensor<3xindex> -> tensor<3xindex>
324
324
// CHECK-DAG: [[VAR_3_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_0_]], [[VAR_2_]], dims = [0, 1, 2] : (tensor<?x?x5xf32>, tensor<3xindex>) -> tensor<?x4x5xf32>
325
325
// CHECK-DAG: [[VAR_4_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[PARAM_1_]], [[VAR_2_]], dims = [0, 1, 2] : (tensor<?x4x5xf32>, tensor<3xindex>) -> tensor<?x4x5xf32>
326
- // CHECK: [[VAR_5_:%.+]] = stablehlo.compare LT, [[VAR_3_]], [[VAR_4_]], NOTYPE : (tensor<?x4x5xf32>, tensor<?x4x5xf32>) -> tensor<?x4x5xi1>
326
+ // CHECK: [[VAR_5_:%.+]] = stablehlo.compare LT, [[VAR_3_]], [[VAR_4_]] : (tensor<?x4x5xf32>, tensor<?x4x5xf32>) -> tensor<?x4x5xi1>
327
327
// CHECK: return [[VAR_5_]] : tensor<?x4x5xi1>
328
328
// CHECK: }
329
329
@@ -447,7 +447,7 @@ func.func @test_leakyrelu_dynamic(%arg0 : tensor<?x10xf32>) -> tensor<?x10xf32>
447
447
// CHECK-DAG: [[VAR_4_:%.+]] = stablehlo.multiply [[PARAM_0_]], [[VAR_3_]] : tensor<?x10xf32>
448
448
// CHECK-DAG: [[VAR_5_:%.+]] = shape.shape_of [[PARAM_0_]] : tensor<?x10xf32> -> tensor<2xindex>
449
449
// CHECK: [[VAR_6_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[VAR_0_]], [[VAR_5_]], dims = [] : (tensor<f32>, tensor<2xindex>) -> tensor<?x10xf32>
450
- // CHECK: [[VAR_7_:%.+]] = stablehlo.compare GT, [[PARAM_0_]], [[VAR_6_]], NOTYPE : (tensor<?x10xf32>, tensor<?x10xf32>) -> tensor<?x10xi1>
450
+ // CHECK: [[VAR_7_:%.+]] = stablehlo.compare GT, [[PARAM_0_]], [[VAR_6_]] : (tensor<?x10xf32>, tensor<?x10xf32>) -> tensor<?x10xi1>
451
451
// CHECK: [[VAR_8_:%.+]] = stablehlo.select [[VAR_7_]], [[PARAM_0_]], [[VAR_4_]] : tensor<?x10xi1>, tensor<?x10xf32>
452
452
// CHECK: return [[VAR_8_]] : tensor<?x10xf32>
453
453
// CHECK: }
@@ -469,7 +469,7 @@ func.func @test_prelu_dynamic(%arg0 : tensor<?x10x12x12xf32>, %arg1: tensor<10x1
469
469
// CHECK-DAG: [[VAR_6_:%.+]] = stablehlo.multiply [[VAR_4_]], [[VAR_5_]] : tensor<?x10x12x12xf32>
470
470
// CHECK-DAG: [[VAR_7_:%.+]] = shape.shape_of [[VAR_4_]] : tensor<?x10x12x12xf32> -> tensor<4xindex>
471
471
// CHECK: [[VAR_8_:%.+]] = stablehlo.dynamic_broadcast_in_dim [[VAR_0_]], [[VAR_7_]], dims = [] : (tensor<f32>, tensor<4xindex>) -> tensor<?x10x12x12xf32>
472
- // CHECK: [[VAR_9_:%.+]] = stablehlo.compare GT, [[VAR_4_]], [[VAR_8_]], NOTYPE : (tensor<?x10x12x12xf32>, tensor<?x10x12x12xf32>) -> tensor<?x10x12x12xi1>
472
+ // CHECK: [[VAR_9_:%.+]] = stablehlo.compare GT, [[VAR_4_]], [[VAR_8_]] : (tensor<?x10x12x12xf32>, tensor<?x10x12x12xf32>) -> tensor<?x10x12x12xi1>
473
473
// CHECK: [[VAR_10_:%.+]] = stablehlo.select [[VAR_9_]], [[VAR_4_]], [[VAR_6_]] : tensor<?x10x12x12xi1>, tensor<?x10x12x12xf32>
474
474
// CHECK: return [[VAR_10_]] : tensor<?x10x12x12xf32>
475
475
// CHECK: }
0 commit comments