Skip to content

Commit 691bc59

Browse files
committed
Update on "Docs for lower smaller models to mps/coreml/qnn"
Differential Revision: [D56340028](https://our.internmc.facebook.com/intern/diff/D56340028/) [ghstack-poisoned]
1 parent 696faf3 commit 691bc59

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

examples/models/llama2/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -248,7 +248,7 @@ Currently we supported lowering the stories model to other backends, including,
248248
for each backend ([CoreML](https://pytorch.org/executorch/main/build-run-coreml.html), [MPS](https://pytorch.org/executorch/main/build-run-mps.html), [QNN](https://pytorch.org/executorch/main/build-run-qualcomm.html)) before trying to lower them. After the backend library is installed, the script to export a lowered model is
249249
250250
- Lower to CoreML: `python -m examples.models.llama2.export_llama -kv --coreml -c stories110M.pt -p params.json`
251-
- MPS: `python -m examples.models.llama2.export_llama -kv --MPS -c stories110M.pt -p params.json`
251+
- MPS: `python -m examples.models.llama2.export_llama -kv --mps -c stories110M.pt -p params.json`
252252
- QNN: `python -m examples.models.llama2.export_llama -kv --qnn -c stories110M.pt -p params.json`
253253
254254
The iOS LLAMA app supports the CoreML and MPS model and the Android LLAMA app supports the QNN model. On Android, it also allow to cross compiler the llama runner binary, push to the device and run.

0 commit comments

Comments
 (0)