You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/models/llava/README.md
+4-5Lines changed: 4 additions & 5 deletions
Original file line number
Diff line number
Diff line change
@@ -7,14 +7,13 @@ In this example, we initiate the process of running multi modality through Execu
7
7
Note that this folder does not host the pretrained LLava model.
8
8
- To have Llava available, follow the [Install instructions](https://github.com/haotian-liu/LLaVA?tab=readme-ov-file#install) in the LLava github. Follow the licence in the specific repo when using L
9
9
- Since the pytorch model version may not be updated, `cd executorch`, run `./install_requirements.sh`.
10
-
- If there is numpy compatibility issue, run `pip install bitsandbytes -I`.
11
-
- Alternatively, run `examples/models/llava_encoder/install_requirements.sh`, to replace the steps above.
12
-
- Run `python3 -m examples.portable.scripts.export --model_name="llava_encoder"`. The llava_encoder.pte file will be generated.
13
-
- Run `./cmake-out/executor_runner --model_path ./llava_encoder.pte` to verify the exported model with ExecuTorch runtime with portable kernels. Note that the portable kernels are not performance optimized. Please refer to other examples like those in llama2 folder for optimization.
10
+
- Run `examples/models/llava/install_requirements.sh`, to install llava specific deps.
11
+
- Run `python3 -m examples.portable.scripts.export --model_name="llava"`. The llava.pte file will be generated.
12
+
- Run `./cmake-out/executor_runner --model_path ./llava.pte` to verify the exported model with ExecuTorch runtime with portable kernels. Note that the portable kernels are not performance optimized. Please refer to other examples like those in llama2 folder for optimization.
14
13
15
14
## TODO
16
15
- Write the pipeline in cpp
17
16
- Have image and text prompts as inputs.
18
17
- Call image processing functions to preprocess the image tensor.
19
-
- Load the llava_encoder.pte model, run it using the image tensor.
18
+
- Load the llava.pte model, run it using the image tensor.
20
19
- The output of the encoder can be combined with the prompt, as inputs to the llama model. Call functions in llama_runner.cpp to run the llama model and get outputs. The ExecuTorch end to end flow for the llama model is located at `examples/models/llama2`.
0 commit comments