Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Investigate/document how to use with OpenRouter #1099

Closed
ErikBjare opened this issue Apr 2, 2024 · 6 comments
Closed

Investigate/document how to use with OpenRouter #1099

ErikBjare opened this issue Apr 2, 2024 · 6 comments
Assignees
Labels
documentation Improvements or additions to documentation

Comments

@ErikBjare
Copy link
Collaborator

As I mentioned in #1082 (review), OpenRouter is an easy way to run lots of different LLMs through a OpenAI-compatible API (notably recommended by Aider).

We should consider investigating this as a user-friendly way to use open models without needing your own hardware, and document it accordingly.

@zigabrencic did some investigation on this. Opening this issue to track our investigation and potentially documenting to users how to use it.

One notable downside of it, and open models generally, is that it may lead to all kinds of bugs caused by less powerful models. We might want to mention this in docs so that users are aware, and don't open spurious bug reports.

@ErikBjare ErikBjare added the documentation Improvements or additions to documentation label Apr 2, 2024
@zigabrencic
Copy link
Collaborator

@vibor you can assign me to this one. Will start looking into it at towards the end of the week.

@viborc
Copy link
Collaborator

viborc commented Apr 3, 2024

Done, thanks, @zigabrencic and @ErikBjare!

@TheoMcCabe
Copy link
Collaborator

@zigabrencic
Copy link
Collaborator

zigabrencic commented Apr 15, 2024

@TheoMcCabe that's great. Thanks for sharing the post.

The issue I came across with gpte was that specific open model {sizes, types} result into rather poor performance of gpte. Running on the local machine has the same issue as with OpenRouter.

In other words. Making the API calls to open router via langchain is the easy part. Making gpte work with open models, is well harder. Hopefully running new gpte benchmarks with Open Models gives us some clarity here.

I'm waiting for llama3 to come out this week to see if it's any better. If it is I suspect llama3 to be also available on OpenRouter as well.

I'm not against adding a few lines from your post to the docs right away. We just need to warn the users as Erik mentioned above about experimentally of the feature.

@TheoMcCabe
Copy link
Collaborator

TheoMcCabe commented Apr 16, 2024

Yeah i saw the same thing with mynt. The only things that could handle the logic required to contruct the requests properly were anthropic or open ai models. Even google gemini sucked!

There are other benefits to using open router other than just getting access to additional models which work well with gpte... like being visible on the open router app rankings. And having access to new models early (which may in future work well with gpte)

I'd say the models not working very well shouldnt be a blocker on us completing this integration

@TheoMcCabe
Copy link
Collaborator

I think this can be closed now as you PR is merged @zigabrencic ?

@viborc viborc closed this as completed May 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
Status: Done
Development

No branches or pull requests

4 participants