-
Notifications
You must be signed in to change notification settings - Fork 607
[executorch] generation.py with kv cache #3030
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/3030
Note: Links to docs will display an error until the docs builds have been completed. ❌ 2 New Failures, 1 Unrelated FailureAs of commit 9a4c1f6 with merge base 7616d42 ( NEW FAILURES - The following jobs have failed:
BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
00e9159
to
5c98ea6
Compare
@lucylq has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
5c98ea6
to
b48bfd1
Compare
@lucylq has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Lint? Also does this work with our current model? Llama2 and stories
Summary: python e2e generation, using tiktoken tokenizer. using text_completion, haven't tried chat_completion. Test Plan: Imported from GitHub, without a `Test Plan:` line. Command, with prompt "Hello, I am" and seq_len = 10 ``` python -m examples.models.llama2.runner.generation --pte llama_4ckpts_x.pte --tokenizer tokenizer.model --prompt="Hello I am" --temperature=0 --params ../llama-models/llama3/params_less.json --max_gen_len=10 ``` fp32, xnn, kv fp32, xnn same results: ``` Result: [{'generation': ' a 25 year old woman. I am a'}] ``` fp32, xnn, int4 ``` Result: [{'generation': ' interested in the following products: - 1 x'}] ``` fp32, xnn, kv, sdpa (need investigation) ``` Result: [{'generation': 'ฉopteraenthalenthalenthalenthalenthalenthalenthalenthal'}] ``` Reviewed By: larryliu0820 Differential Revision: D56087430 Pulled By: lucylq
b48bfd1
to
9a4c1f6
Compare
This pull request was exported from Phabricator. Differential Revision: D56087430 |
No description provided.