Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Add support for Gemma-7B model from Groq #2488

Closed
haseeb-heaven opened this issue Mar 13, 2024 · 4 comments
Closed

[Feature]: Add support for Gemma-7B model from Groq #2488

haseeb-heaven opened this issue Mar 13, 2024 · 4 comments
Labels
enhancement New feature or request

Comments

@haseeb-heaven
Copy link
Contributor

The Feature

Groq now supports Gemma-7b model and the model name is gemma-7b-it please add the support for this.
And this is much more faster like 600 Tokens/Second

Here is sample code from their Docs.

from groq import Groq

client = Groq()
completion = client.chat.completions.create(
    model="gemma-7b-it",
    messages=[
        {
            "role": "user",
            "content": ""
        }
    ],
    temperature=0.1,
    max_tokens=8192,
    top_p=1,
    stream=True,
    stop=None,
)

for chunk in completion:
    print(chunk.choices[0].delta.content or "", end="")

Motivation, pitch

Much faster models than other Groq models and also smaller in size.

Twitter / LinkedIn details

No response

@haseeb-heaven haseeb-heaven added the enhancement New feature or request label Mar 13, 2024
@ishaan-jaff
Copy link
Contributor

We already support this https://docs.litellm.ai/docs/providers/groq

Screenshot 2024-03-13 at 7 39 57 AM

@ishaan-jaff
Copy link
Contributor

@haseeb-heaven
Copy link
Contributor Author

We already support this https://docs.litellm.ai/docs/providers/groq

Screenshot 2024-03-13 at 7 39 57 AM

But Gemma is not working from Groq

@haseeb-heaven
Copy link
Contributor Author

Working now thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants