You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Text Generation Inference (TGI) now supports the Messages API, which is fully compatible with the OpenAI Chat Completion API. This feature is available starting from version 1.4.0. You can use OpenAI’s client libraries or third-party libraries expecting OpenAI schema to interact with TGI’s Messages API. Below are some examples of how to utilize this compatibility.
So you should be able to use magentic with HuggingFace TGI by setting the base_url param in OpenaiChatModel. See https://magentic.dev/configuration/
Please let me know here if this works for you or if there are any issues. If it works I'd be happy to accept a PR to add a section to the docs about it with useful links and any additional setup steps needed.
HuggingFace has a standard TGI/Docker type container for serving LLM requests.
It would be useful to take advantage of HuggingFace features for TGI generation.
Github: Large Language Model Text Generation Inference
Docs: Text Generation Inference
The text was updated successfully, but these errors were encountered: