|
1 | 1 | // DEFINE: %{compile} = mlir-opt %s \
|
2 | 2 | // DEFINE: -transform-interpreter -test-transform-dialect-erase-schedule \
|
3 | 3 | // DEFINE: -one-shot-bufferize="bufferize-function-boundaries" \
|
4 |
| -// DEFINE: -buffer-deallocation-pipeline -cse -canonicalize -convert-vector-to-scf -test-lower-to-llvm |
| 4 | +// DEFINE: -buffer-deallocation-pipeline="private-function-dynamic-ownership" \ |
| 5 | +// DEFINE: -cse -canonicalize -test-lower-to-llvm |
5 | 6 | // DEFINE: %{entry_point} = main
|
6 | 7 | // DEFINE: %{run} = mlir-cpu-runner -e %{entry_point} -entry-point-result=void \
|
7 | 8 | // DEFINE: -shared-libs=%mlir_runner_utils,%mlir_c_runner_utils
|
@@ -70,14 +71,14 @@ func.func @main() {
|
70 | 71 | return
|
71 | 72 | }
|
72 | 73 |
|
73 |
| -func.func @matmul(%A: tensor<7x16xi32>, %B: tensor<16x13xi32>, %C: tensor<7x13xi32>) -> tensor<7x13xi32> { |
| 74 | +func.func private @matmul(%A: tensor<7x16xi32>, %B: tensor<16x13xi32>, %C: tensor<7x13xi32>) -> tensor<7x13xi32> { |
74 | 75 | %C_matmul = linalg.matmul ins(%A, %B: tensor<7x16xi32>, tensor<16x13xi32>)
|
75 | 76 | outs(%C: tensor<7x13xi32>) -> tensor<7x13xi32>
|
76 | 77 |
|
77 | 78 | return %C_matmul : tensor<7x13xi32>
|
78 | 79 | }
|
79 | 80 |
|
80 |
| -func.func @mmt4d(%A: tensor<7x16xi32>, %B: tensor<16x13xi32>, %C: tensor<7x13xi32>) -> tensor<7x13xi32> { |
| 81 | +func.func private @mmt4d(%A: tensor<7x16xi32>, %B: tensor<16x13xi32>, %C: tensor<7x13xi32>) -> tensor<7x13xi32> { |
81 | 82 | %zero = arith.constant 0 : i32
|
82 | 83 |
|
83 | 84 | %A_pack_empty = tensor.empty() : tensor<2x16x8x1xi32>
|
|
0 commit comments