Skip to content

Commit 28b0625

Browse files
yifan_shen3pytorchbot
authored andcommitted
Core ML Has Added Index_Put Support, No Need to Skip Anymore (#2975)
Summary: It was a workaround to skip `aten.index_put` op in Core ML delegation, at the cost of partitioning the Llama model into 13 pieces. For better performance, we prefer to delegate the whole model to Core ML. Since Core ML has added the [necessary support](apple/coremltools#2190), it is time to revert this workaround Pull Request resolved: #2975 Reviewed By: kirklandsign Differential Revision: D56002979 Pulled By: cccclai fbshipit-source-id: e7a7c8c43706cb57eba3e6f720b3d713bec5065b (cherry picked from commit 7d4bafc)
1 parent 27e1a62 commit 28b0625

File tree

1 file changed

+0
-3
lines changed

1 file changed

+0
-3
lines changed

examples/models/llama2/export_llama_lib.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -663,9 +663,6 @@ def _export_llama(modelname, args) -> str: # noqa: C901
663663
partitioners.append(
664664
# pyre-ignore: Undefined attribute [16]: Module `executorch.backends` has no attribute `apple`
665665
CoreMLPartitioner(
666-
skip_ops_for_coreml_delegation=[
667-
"aten.index_put.default",
668-
],
669666
compile_specs=compile_specs,
670667
)
671668
)

0 commit comments

Comments
 (0)