@@ -20,8 +20,8 @@ func.func @matvec_static(%3: tensor<128x384xf32>, %4: tensor<384xf32>) -> tensor
2020 } -> tensor <128 xf32 >
2121 return %7 : tensor <128 xf32 >
2222}
23- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [32, 0], distribution = [32, 0], vector_common_parallel = [32 , 0], vector_reduction = [0, 16]>
24- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
23+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [32, 0], vector_common_parallel = [16 , 0], vector_reduction = [0, 16]>
24+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
2525// CHECK: func.func @matvec_static(
2626// CHECK-SAME: translation_info = #[[TRANSLATION]]
2727// CHECK: linalg.generic
@@ -36,8 +36,8 @@ func.func @matvec_dynamic(%11: tensor<?xf32>, %12: tensor<?x?xf32>, %13: tensor<
3636 %15 = linalg.matvec ins (%12 , %13 : tensor <?x?xf32 >, tensor <?xf32 >) outs (%14 : tensor <?xf32 >) -> tensor <?xf32 >
3737 return %15 : tensor <?xf32 >
3838}
39- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [64, 0], distribution = [64, 0], vector_common_parallel = [32 , 0], vector_reduction = [0, 16]>
40- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
39+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [64, 0], vector_common_parallel = [16 , 0], vector_reduction = [0, 16]>
40+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
4141// CHECK: func.func @matvec_dynamic(
4242// CHECK-SAME: translation_info = #[[TRANSLATION]]
4343// CHECK: linalg.matvec
@@ -54,7 +54,7 @@ func.func @dot_static(%3: tensor<384xf32>, %4: tensor<384xf32>) -> tensor<f32> a
5454 return %7 : tensor <f32 >
5555}
5656// CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0], vector_reduction = [16]
57- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
57+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
5858// CHECK: func.func @dot_static(
5959// CHECK-SAME: translation_info = #[[TRANSLATION]]
6060// CHECK: linalg.dot
@@ -70,7 +70,7 @@ func.func @dot_dynamic(%5: tensor<f32>, %8: tensor<?xf32>, %9: tensor<?xf32>) ->
7070 return %11 : tensor <f32 >
7171}
7272// CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0], vector_reduction = [16]>
73- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
73+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
7474// CHECK: func.func @dot_dynamic(
7575// CHECK-SAME: translation_info = #[[TRANSLATION]]
7676// CHECK: linalg.dot
@@ -169,8 +169,8 @@ func.func @matmul_partially_peel(%3: tensor<16641x16xf32>, %4: tensor<16x8xf32>)
169169 %7 = linalg.matmul ins (%3 , %4 : tensor <16641 x16 xf32 >, tensor <16 x8 xf32 >) outs (%6 : tensor <16641 x8 xf32 >) -> tensor <16641 x8 xf32 >
170170 return %7 : tensor <16641 x8 xf32 >
171171}
172- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [43, 8, 0], distribution = [43, 8, 0], vector_common_parallel = [8, 32 , 0], vector_reduction = [0, 0, 16]>
173- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
172+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [43, 8, 0], vector_common_parallel = [3, 8 , 0], vector_reduction = [0, 0, 16]>
173+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
174174// CHECK: func.func @matmul_partially_peel(
175175// CHECK-SAME: translation_info = #[[TRANSLATION]]
176176// CHECK: linalg.matmul
@@ -423,8 +423,8 @@ func.func @matmul_static(%3: tensor<384x512xf32>, %4: tensor<512x128xf32>) -> te
423423 %7 = linalg.matmul ins (%3 , %4 : tensor <384 x512 xf32 >, tensor <512 x128 xf32 >) outs (%6 : tensor <384 x128 xf32 >) -> tensor <384 x128 xf32 >
424424 return %7 : tensor <384 x128 xf32 >
425425}
426- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [48, 64, 0], distribution = [48, 64, 0], vector_common_parallel = [8, 32 , 0], vector_reduction = [0, 0, 16]>
427- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
426+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [48, 64, 0], vector_common_parallel = [8, 16 , 0], vector_reduction = [0, 0, 16]>
427+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
428428// CHECK: func.func @matmul_static(
429429// CHECK-SAME: translation_info = #[[TRANSLATION]]
430430// CHECK: linalg.matmul
@@ -486,8 +486,8 @@ func.func @gemm_unit_N(%5: tensor<?x1xf32>, %6: tensor<?x?xf32>, %7: tensor<?x1x
486486 %8 = linalg.matmul ins (%6 , %5 : tensor <?x?xf32 >, tensor <?x1 xf32 >) outs (%7 : tensor <?x1 xf32 >) -> tensor <?x1 xf32 >
487487 return %8 : tensor <?x1 xf32 >
488488}
489- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [64, 0, 0], distribution = [64, 0, 0], vector_common_parallel = [8, 32 , 0], vector_reduction = [0, 0, 16]>
490- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
489+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [64, 0, 0], vector_common_parallel = [8, 1 , 0], vector_reduction = [0, 0, 16]>
490+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
491491// CHECK: func.func @gemm_unit_N(
492492// CHECK-SAME: translation_info = #[[TRANSLATION]]
493493// CHECK: linalg.matmul
@@ -500,8 +500,8 @@ func.func @gemm_unit_M_unit_N(%4: tensor<1x?xf32>, %5: tensor<?x1xf32>, %6: tens
500500 %7 = linalg.matmul ins (%4 , %5 : tensor <1 x?xf32 >, tensor <?x1 xf32 >) outs (%6 : tensor <1 x1 xf32 >) -> tensor <1 x1 xf32 >
501501 return %7 : tensor <1 x1 xf32 >
502502}
503- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0, 0, 0], vector_common_parallel = [1, 32 , 0], vector_reduction = [0, 0, 16]>
504- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
503+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0, 0, 0], vector_common_parallel = [1, 1 , 0], vector_reduction = [0, 0, 16]>
504+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
505505// CHECK: func.func @gemm_unit_M_unit_N(
506506// CHECK-SAME: translation_info = #[[TRANSLATION]]
507507// CHECK: linalg.matmul
@@ -517,8 +517,8 @@ func.func @matmul_odd(%4: tensor<33x16xf32>, %5: tensor<16x49xf32>, %6: tensor<3
517517 %9 = linalg.matmul ins (%4 , %5 : tensor <33 x16 xf32 >, tensor <16 x49 xf32 >) outs (%8 : tensor <33 x49 xf32 >) -> tensor <33 x49 xf32 >
518518 return %9 : tensor <33 x49 xf32 >
519519}
520- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<cache_parallel = [8, 49, 0], distribution = [8, 49, 0], vector_common_parallel = [8, 32 , 0], vector_reduction = [0, 0, 16]>
521- // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {{\{}}enable_loop_peeling} >
520+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [8, 49, 0], vector_common_parallel = [3, 7 , 0], vector_reduction = [0, 0, 16]>
521+ // CHECK-DAG: #[[TRANSLATION:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
522522// CHECK: func.func @matmul_odd(
523523// CHECK-SAME: translation_info = #[[TRANSLATION]]
524524// CHECK: linalg.matmul
@@ -997,7 +997,7 @@ func.func @non_trivial_program(%3: tensor<128x1x128x1xf32>, %4: tensor<128x1xf32
997997 %10 = linalg.matmul ins (%expanded , %4 : tensor <1 x128 xf32 >, tensor <128 x1 xf32 >) outs (%8 : tensor <1 x1 xf32 >) -> tensor <1 x1 xf32 >
998998 return %10 : tensor <1 x1 xf32 >
999999}
1000- // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0, 0, 0], vector_common_parallel = [1, 32 , 0], vector_reduction = [0, 0, 16]>
1000+ // CHECK-DAG: #[[CONFIG:.+]] = #iree_cpu.lowering_config<distribution = [0, 0, 0], vector_common_parallel = [1, 1 , 0], vector_reduction = [0, 0, 16]>
10011001// CHECK-NOT: lowering_config
10021002// CHECK: func.func @non_trivial_program(
10031003// CHECK-SAME: translation_info = #[[TRANSLATION]]
@@ -1259,9 +1259,9 @@ func.func @custom_op(%arg0 : tensor<384x512xf32>, %arg1 : tensor<512x128xf32>,
12591259 return %1 : tensor <384 x128 xf32 >
12601260}
12611261// CHECK-DAG: #[[CONFIG0:.+]] = #iree_codegen.lowering_config<tile_sizes = {{\[}}[48, 64, 0]]>
1262- // CHECK-DAG: #[[CONFIG1:.+]] = #iree_cpu.lowering_config<cache_parallel = [48, 64], vector_common_parallel = [8, 32 ]>
1263- // CHECK-DAG: #[[CONFIG2:.+]] = #iree_cpu.lowering_config<cache_parallel = [48, 64, 0], distribution = [48, 64, 0], vector_common_parallel = [8, 32 , 0], vector_reduction = [0, 0, 16]>
1264- // CHECK-DAG: #[[TRANSLATION_INFO:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert, {enable_loop_peeling} >
1262+ // CHECK-DAG: #[[CONFIG1:.+]] = #iree_cpu.lowering_config<vector_common_parallel = [8, 16 ]>
1263+ // CHECK-DAG: #[[CONFIG2:.+]] = #iree_cpu.lowering_config<distribution = [48, 64, 0], vector_common_parallel = [8, 16 , 0], vector_reduction = [0, 0, 16]>
1264+ // CHECK-DAG: #[[TRANSLATION_INFO:.+]] = #iree_codegen.translation_info<pipeline = CPUDoubleTilingExpert>
12651265// CHECK: func @custom_op(
12661266// CHECK-SAME: translation_info = #translation
12671267// CHECK: iree_linalg_ext.custom_op
0 commit comments