Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multi-model providers break the universal function calling expectation inside providers #22

Open
swerner opened this issue Apr 29, 2024 · 0 comments
Milestone

Comments

@swerner
Copy link
Contributor

swerner commented Apr 29, 2024

For model providers like Groq, Local, OpenRouter, etc, each different model you might want to use inside these different services could have different function calling/tool calling mechanism and quirks: Specifically I just tried generating something with Groq+llama3 and the xml response came back in a different format.

For now, going to move away from providing direct solutions for things like Groq and make it similar to what we're doing with output_adapters where you can specify a custom AI provider in your code to use and implement the specific quirks of that model. Over time we'll find patterns that work with things like Llama3 or Mistral or Hermes2 that we can provide as bases.

@swerner swerner added this to the 0.1.0 milestone Apr 29, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant