Skip to content

Typo fixes in native-execution.md #1394

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Nov 26, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions docs/native-execution.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,14 +16,14 @@ The 'llama runner' is a native standalone application capable of
running a model exported and compiled ahead-of-time with either
Executorch (ET) or AOT Inductor (AOTI). Which model format to use
depends on your requirements and preferences. Executorch models are
optimized for portability across a range of decices, including mobile
optimized for portability across a range of devices, including mobile
and edge devices. AOT Inductor models are optimized for a particular
target architecture, which may result in better performance and
efficiency.

Building the runners is straightforward with the included cmake build
files and is covered in the next sections. We will showcase the
runners using ~~stories15M~~ llama2 7B and llama3.
runners using llama2 7B and llama3.

## What can you do with torchchat's llama runner for native execution?

Expand Down Expand Up @@ -160,7 +160,7 @@ and native execution environments, respectively.

After exporting a model, you will want to verify that the model
delivers output of high quality, and works as expected. Both can be
achieved with the Python environment. All torchchat Python comands
achieved with the Python environment. All torchchat Python commands
can work with exported models. Instead of loading the model from a
checkpoint or GGUF file, use the `--dso-path model.so` and
`--pte-path model.pte` for loading both types of exported models. This
Expand Down
Loading