-
Notifications
You must be signed in to change notification settings - Fork 607
Enable W2L XNNPACK delegation #212
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This pull request was exported from Phabricator. Differential Revision: D48872594 |
This pull request was exported from Phabricator. Differential Revision: D48872594 |
3 similar comments
This pull request was exported from Phabricator. Differential Revision: D48872594 |
This pull request was exported from Phabricator. Differential Revision: D48872594 |
This pull request was exported from Phabricator. Differential Revision: D48872594 |
Summary: It's a good idea to allow each model to choose its own runner. This also help unblock #212 until we know the root cause of why it runs OOM. Pull Request resolved: #277 Reviewed By: kirklandsign Differential Revision: D49176509 Pulled By: huydhn fbshipit-source-id: b5e74943fb62839e3b9a7ff132842c8614d0f84b
This pull request was exported from Phabricator. Differential Revision: D48872594 |
1 similar comment
This pull request was exported from Phabricator. Differential Revision: D48872594 |
Summary: [Model](https://pytorch.org/audio/main/_modules/torchaudio/models/wav2letter.html#Wav2Letter) is like conv->relu->conv->relu... -> log_softmax (not delegated at the moment) Set _unlift=True in CaptureConfig because we have some issues with conv1d. Unblock for now and deploy on CI to prevent from further regression. Reviewed By: digantdesai Differential Revision: D48872594
This pull request was exported from Phabricator. Differential Revision: D48872594 |
1 similar comment
This pull request was exported from Phabricator. Differential Revision: D48872594 |
This pull request has been merged in c5fecc8. |
* fixed device support for int4, and unit tests * typo * typo * bfloat16 on macos * bfloat16 in mps, redux * typo * remove extraneous use_cuda * device setting in gguf loader
Summary: Model is like conv->relu->conv->relu... -> log_softmax (not delegated at the moment)
Reviewed By: digantdesai
Differential Revision: D48872594