Conversation
|
You can actually use the current OpenAI implementation :) Taken from here: https://github.com/openai/openai-python?tab=readme-ov-file#module-level-client |
There isn't a single instance of the term "groq" in the link you attached: And Google doesn't return any decent results for "https://api.groq.com/openai/v1/chat/completions" So, no you cannot use the current OpenAI implementation. Not to mention that OpenAI itself recommends not using the module level client you link in application code Thanks ;) |
|
Oh, looks like I needed to update myself a little bit on the lib and such, here is the thing: Here (https://console.groq.com/docs/quickstart) you can find the endpoint I mentioned:
And here is how to use it: Note: Groq is OpenAI compatible. Means that you can use any implementation of that API. Here's the output: Please let me know if it works for you! |
|
there has been a rate limit on groq api, does anyone know how to bypass it & use it in free tier? |




Groq is delivering the world's fastest LLM inference with very generous free API credits. The default usage is with llama3-70b as it is currently Groq's most advanced free model, though paying users have preview access to llama3.1-405B. To use llama3.1-405B, replace all instances of
llama3-70b-8192byllama-3.1-405b-reasoning