Content
@
0 reply
0 recast
0 reaction
Giuliano Giacaglia
@giu
Groq enables much faster throughput at the lowest cost possible! Almost 500 tokens per second for lower cost!
4 replies
5 recasts
41 reactions
sean 🔹
@swabbie.eth
but what’s the quality? it’s not on any of the quality lists
0 reply
0 recast
2 reactions
pbnather
@pbnather
It's Mixtral 8x7B, so 62
1 reply
0 recast
2 reactions
sean 🔹
@swabbie.eth
i didn't know that. this comparison lists them separately with unknown # of parameters for grok, but it knows that value for Mixtral 8x7B? https://zapier.com/blog/best-llm/
0 reply
0 recast
1 reaction
pbnather
@pbnather
Grok and Groq are different things. Grok is LLM by Twitter/X. Groq mentioned above is startup that has custom GPUs and lets you run different open source models, like Mixtral 8x7B or Llama2. Point is that using Groq API you can have very fast inference with these models vs using competitors like perplexity.
0 reply
0 recast
1 reaction