Content pfp
Content
@
0 reply
0 recast
0 reaction

altmbr pfp
altmbr
@altmbr
Curious how do folks think about picking the right model for their use case? Ie, when use GPT-4o, Llama 3, Gemini 1.5, DeepSeek or other? How does pricing fit into consideration?
5 replies
0 recast
5 reactions

AfroRick pfp
AfroRick
@afrorick
I start with requirements on latency. If I can tolerate high latency I start with large models and work down. If I can't tolerate high latency I start with the small models and work up until I get something that is giving me 85%+ success rate on typical questions.
0 reply
0 recast
0 reaction