Content pfp
Content
@
0 reply
0 recast
0 reaction

Giuliano Giacaglia pfp
Giuliano Giacaglia
@giu
Groq enables much faster throughput at the lowest cost possible! Almost 500 tokens per second for lower cost!
4 replies
10 recasts
82 reactions

sean pfp
sean
@swabbie.eth
but what’s the quality? it’s not on any of the quality lists
0 reply
0 recast
2 reactions

pbnather pfp
pbnather
@pbnather
It's Mixtral 8x7B, so 62
1 reply
0 recast
2 reactions

sean pfp
sean
@swabbie.eth
i didn't know that. this comparison lists them separately with unknown # of parameters for grok, but it knows that value for Mixtral 8x7B? https://zapier.com/blog/best-llm/
0 reply
0 recast
1 reaction

pbnather pfp
pbnather
@pbnather
Grok and Groq are different things. Grok is LLM by Twitter/X. Groq mentioned above is startup that has custom GPUs and lets you run different open source models, like Mixtral 8x7B or Llama2. Point is that using Groq API you can have very fast inference with these models vs using competitors like perplexity.
0 reply
0 recast
1 reaction