You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/README.md
+4-3Lines changed: 4 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -9,6 +9,7 @@ ExecuTorch's extensive support spans from simple modules like "Add" to comprehen
9
9
## Directory structure
10
10
```
11
11
examples
12
+
├── llm_manual # A storage place for the files that [LLM Maunal](https://pytorch.org/executorch/main/llm/getting-started.html) needs
12
13
├── models # Contains a set of popular and representative PyTorch models
13
14
├── portable # Contains end-to-end demos for ExecuTorch in portable mode
14
15
├── selective_build # Contains demos of selective build for optimizing the binary size of the ExecuTorch runtime
@@ -20,7 +21,7 @@ examples
20
21
| └── mps # Contains end-to-end demos of MPS backend
21
22
├── arm # Contains demos of the Arm TOSA and Ethos-U NPU flows
22
23
├── qualcomm # Contains demos of Qualcomm QNN backend
23
-
├── xtensa# Contains demos of exporting and running a simple model on Xtensa Hifi4 DSP
24
+
├── cadence # Contains demos of exporting and running a simple model on Xtensa DSPs
24
25
├── third-party # Third-party libraries required for working on the demos
25
26
└── README.md # This file
26
27
```
@@ -30,9 +31,9 @@ examples
30
31
31
32
A user's journey may commence by exploring the demos located in the [`portable/`](./portable) directory. Here, you will gain insights into the fundamental end-to-end workflow to generate a binary file from a ML model in [portable mode](../docs/source/concepts.md##portable-mode-lean-mode) and run it on the ExecuTorch runtime.
32
33
33
-
## Demo of Llama2
34
+
## Demo of Llama 2 and Llama 3
34
35
35
-
[This page](./models/llama2/README.md) demonstrates how to run a Llama 2 7B model on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
36
+
[This page](./models/llama2/README.md) demonstrates how to run Llama 2 7B and Llama 3 8B models on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on Android and iOS mobile phones.
0 commit comments