@@ -551,9 +551,9 @@ func.func @extract_load_scalar_dyn_off(%arg0: memref<?xf32>, %arg1: index, %arg2
551
551
return %1 : f32
552
552
}
553
553
554
- // CHECK-LABEL: @extract_load_vec
554
+ // CHECK-LABEL: @extract_load_vec_non_zero_off
555
555
// CHECK-SAME: (%[[ARG0:.*]]: memref<?x?xf32>, %[[ARG1:.*]]: index, %[[ARG2:.*]]: index)
556
- func.func @extract_load_vec (%arg0: memref <?x?xf32 >, %arg1: index , %arg2: index ) -> vector <4 xf32 > {
556
+ func.func @extract_load_vec_non_zero_off (%arg0: memref <?x?xf32 >, %arg1: index , %arg2: index ) -> vector <4 xf32 > {
557
557
// CHECK: %[[C1:.*]] = arith.constant 1 : index
558
558
// CHECK: %[[OFF:.*]] = arith.addi %[[ARG1]], %[[C1]] overflow<nsw> : index
559
559
// CHECK: %[[RES:.*]] = vector.load %[[ARG0]][%[[OFF]], %[[ARG2]]] : memref<?x?xf32>, vector<4xf32>
@@ -563,9 +563,9 @@ func.func @extract_load_vec(%arg0: memref<?x?xf32>, %arg1: index, %arg2: index)
563
563
return %1 : vector <4 xf32 >
564
564
}
565
565
566
- // CHECK-LABEL: @extract_load_scalar_high_rank
566
+ // CHECK-LABEL: @extract_load_scalar_non_zero_off_2d_src_memref
567
567
// CHECK-SAME: (%[[ARG0:.*]]: memref<?x?xf32>, %[[ARG1:.*]]: index, %[[ARG2:.*]]: index)
568
- func.func @extract_load_scalar_high_rank (%arg0: memref <?x?xf32 >, %arg1: index , %arg2: index ) -> f32 {
568
+ func.func @extract_load_scalar_non_zero_off_2d_src_memref (%arg0: memref <?x?xf32 >, %arg1: index , %arg2: index ) -> f32 {
569
569
// CHECK: %[[C1:.*]] = arith.constant 1 : index
570
570
// CHECK: %[[OFF:.*]] = arith.addi %[[ARG2]], %[[C1]] overflow<nsw> : index
571
571
// CHECK: %[[RES:.*]] = memref.load %[[ARG0]][%[[ARG1]], %[[OFF]]] : memref<?x?xf32>
@@ -587,9 +587,9 @@ func.func @extract_load_vec_high_rank(%arg0: memref<?x?x?xf32>, %arg1: index, %a
587
587
return %1 : vector <4 xf32 >
588
588
}
589
589
590
- // CHECK-LABEL: @negative_extract_load_scalar_from_vec_memref
590
+ // CHECK-LABEL: @negative_extract_load_scalar_from_memref_of_vec
591
591
// CHECK-SAME: (%[[ARG0:.*]]: memref<?xvector<4xf32>>, %[[ARG1:.*]]: index)
592
- func.func @negative_extract_load_scalar_from_vec_memref (%arg0: memref <?xvector <4 xf32 >>, %arg1: index ) -> f32 {
592
+ func.func @negative_extract_load_scalar_from_memref_of_vec (%arg0: memref <?xvector <4 xf32 >>, %arg1: index ) -> f32 {
593
593
// CHECK: %[[RES:.*]] = vector.load %[[ARG0]][%[[ARG1]]] : memref<?xvector<4xf32>>, vector<4xf32>
594
594
// CHECK: %[[EXT:.*]] = vector.extract %[[RES]][0] : f32 from vector<4xf32>
595
595
// CHECK: return %[[EXT]] : f32
@@ -621,7 +621,7 @@ func.func @negative_extract_load_scalable(%arg0: memref<?xf32>, %arg1: index) ->
621
621
}
622
622
623
623
//-----------------------------------------------------------------------------
624
- // [Pattern: StoreFromSplat ]
624
+ // [Pattern: StoreOpFromSplatOrBroadcast ]
625
625
//-----------------------------------------------------------------------------
626
626
627
627
// CHECK-LABEL: @store_splat
@@ -661,9 +661,9 @@ func.func @negative_store_scalable(%arg0: memref<?xf32>, %arg1: index, %arg2: f3
661
661
return
662
662
}
663
663
664
- // CHECK-LABEL: @negative_store_vec_memref
664
+ // CHECK-LABEL: @negative_store_memref_of_vec
665
665
// CHECK-SAME: (%[[ARG0:.*]]: memref<?xvector<1xf32>>, %[[ARG1:.*]]: index, %[[ARG2:.*]]: f32)
666
- func.func @negative_store_vec_memref (%arg0: memref <?xvector <1 xf32 >>, %arg1: index , %arg2: f32 ) {
666
+ func.func @negative_store_memref_of_vec (%arg0: memref <?xvector <1 xf32 >>, %arg1: index , %arg2: f32 ) {
667
667
// CHECK: %[[RES:.*]] = vector.splat %[[ARG2]] : vector<1xf32>
668
668
// CHECK: vector.store %[[RES]], %[[ARG0]][%[[ARG1]]] : memref<?xvector<1xf32>>, vector<1xf32>
669
669
%0 = vector.splat %arg2 : vector <1 xf32 >
0 commit comments