-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Add messages api compatibility docs #1478
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, thanks a lot !
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let some comments
TGI can be deployed on various cloud providers for scalable and robust text generation. One such provider is Amazon SageMaker, which has recently added support for TGI. Here's how you can deploy TGI on Amazon SageMaker: | ||
|
||
## Amazon SageMaker | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should add a note with the min required version for the support of the messages API. I might confuse people
|
||
_Messages API is compatible to OpenAI Chat Completion API_ | ||
|
||
Text Generation Inference (TGI) now supports the Message API which is fully compatible with the OpenAI Chat Completion API. This means you can use OpenAI's client libraries to interact with TGI's Messages API. Below are some examples of how to utilize this compatibility. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sugg:
Text Generation Inference (TGI) now supports the Message API, which is fully compatible with the OpenAI Chat Completion API. You can use OpenAI's client libraries or third-party libraries expecting OpenAI schema to interact with TGI's Messages API. Below are some examples of how to utilize this compatibility.
hub = { | ||
'HF_MODEL_ID':'HuggingFaceH4/zephyr-7b-beta', | ||
'SM_NUM_GPUS': json.dumps(1), | ||
'MESSAGES_API_ENABLED': True |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Fix intent
This PR adds a new page to the docs that describes the Messages API and how to use it. Additionally this page will contain cloud provider specific information for enabling and using this feature. This PR includes a SageMaker example/information.
This PR adds a new page to the docs that describes the Messages API and how to use it.
Additionally this page will contain cloud provider specific information for enabling and using this feature. This PR includes a SageMaker example/information.