Skip to content

Default to CUDA if available #431

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Apr 25, 2024
Merged

Default to CUDA if available #431

merged 1 commit into from
Apr 25, 2024

Conversation

malfet
Copy link
Contributor

@malfet malfet commented Apr 23, 2024

Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU

Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Apr 23, 2024
@malfet
Copy link
Contributor Author

malfet commented Apr 23, 2024

Not sure if this will work though, need to run some local tests...

@mikekgfb mikekgfb self-requested a review April 25, 2024 02:58
@mikekgfb mikekgfb merged commit 7c011e4 into main Apr 25, 2024
@mikekgfb mikekgfb deleted the malfet-patch-2 branch April 25, 2024 07:32
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
malfet added a commit that referenced this pull request Jul 17, 2024
Unless user specifies device, default it to CUDA if it's available on the platform.

As for all models, other than TinyLlama inference on GPU will be faster than on CPU
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Meta Open Source bot.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants