-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add the ability to use models deployed on Azure AI Studio #3902
Comments
Thanks for the issue! I think supporting the Azure AI Model Inference API would be great! I think it makes sense to have it as a separate model client that also implements the We'd love if you're interested in helping build this! |
@jackgerrits I'll be happy to implement these changes, can this task be assigned to me as I have already started work on this task in my fork of this repository. I'll raise a draft pull request in a day or so for us to discuss. |
@rohanthacker - this is supported in dotnet now with #3790 |
Following this issue, since this will enable us to use phi, on some internal cases. |
@rohanthacker any update on this one? |
@edirgarcia Depending on the API you are using. If you are using the Core API, you don't need to wait for this feature you can use the |
Hi @ekzhu, I’ve completed the initial implementation of This is a draft, as I wanted to discuss the significant code duplication between this client and the existing OpenAIChatCompletionClient. If the team agrees with this approach, I’ll code/copy the rest of the implementation as needed. Currently, the two clients are nearly identical, with the only differences being the type variations required by each library. Given that Additionally, I was able to get OpenAIChatCompletionClient to work with models deployed on Azure AI Studio by setting the
Both models work fine when connected directly to the OpenAI API. Happy to keep working on this—just looking for team input on code duplication and next steps. |
@edirgarcia Phi-3.5 is working with the |
Thank you I will test this out next week. |
Thanks @rohanthacker! For the Core API, the user can choose any client they want to use. So this is not a blocker. I think there is still benefit of wrapping the |
@rohanthacker, it has been a while. I am closing this issue for now unless you have already finished or close to finish the implementation, in that case, please let @yanivvak and I know under #4683, and submit a PR for your implementation. |
What feature would you like to be added?
I would like the ability to use a model that is deployed on Azure AI Studio and uses the Azure AI Model Inference API.
If needed, I would like to assist in the creation of the feature. However I have a few questions and require some help about what would be the best way to implement this feature.
Questions:
AzureOpenAIChatCompletionClient
?I have already tried to do this however the API produces an invalid URL and responds with a 404 error, as the endpoint created by Azure AI Studio and the client are not the same.
Looking forward to discussing more on this
Why is this needed?
Azure AI Studio provides a large catalog of models along with various deployment options that make it easy for developers to access a wide variety of models. Given the nature of this project, having the ability to integrate this diverse set of models out of the box will allow for more adoption of the project and allow developers to bring their own model in without the need to code a new client for each.
The text was updated successfully, but these errors were encountered: