Content pfp
Content
@
0 reply
0 recast
0 reaction

Greg pfp
Greg
@greg
I haven't been following the AI space very closely, especially outside of OpenAI stuff. How far behind are open source alternatives? I've heard about stuff like llama2 supposedly being decent for chat but I'm assuming the agents/function calling and whatnot is a different story?
7 replies
1 recast
13 reactions

Warpmaster General pfp
Warpmaster General
@my
It's largely a question of UX more than capability. ChatGPT(+API) is turnkey, cheap, convenient. If you want to roll your own via LLaMA, Langchain, et al, you're looking down the barrel of a week or two just to get it running, let alone fine tuning it to a level where its output is actually usable (not even useful).
1 reply
0 recast
3 reactions

Neokry pfp
Neokry
@neokry
agents and function calling are external to the LLM I think. you can have a library like langchain do things like this for you and pick model to use as the “backend”
1 reply
0 recast
2 reactions

Katsuya pfp
Katsuya
@kn
It depends on how you measure it/use cases, e.g. some OS models claim they are better than GPT-4 in some benchmark. But in general, my intuition is that OpenAI is ~1 year ahead than OS models. My building approach is: Just go with OpenAI if possible then optimize with other options if absolutely need to.
1 reply
0 recast
3 reactions

Gabriel Ayuso ⌁ pfp
Gabriel Ayuso ⌁
@gabrielayuso.eth
API providers (should) spend considerable resources in model quality and tuning so when you use such APIs you get the benefits of all of that. If you just use an open source model you'll need to do more work on your own to get it to do what you want and do a lot of output sanitization yourself.
0 reply
0 recast
2 reactions

m_j_r pfp
m_j_r
@m-j-r.eth
none of it is really production-level. then again, GPT-4-Turbo's 128k context is no surprise, given Yarn-Mistral has tenuous parity already. really a question of capital cost, since there's a wide variety of performance boosts that can get folded into one model within a stack like Autogen + MemGPT + Voyager.
1 reply
0 recast
1 reaction

Jason Goldberg  pfp
Jason Goldberg
@betashop.eth
will let you know very soon from airstack perspective we've been working on a poc with llama2 fine-tuned to our use case, to see if it can outperform gpt4 by end of this week will let you know!
0 reply
0 recast
0 reaction

Minh Do pfp
Minh Do
@minh
This is where benchmarks like huggingface’s leaderboard come in handy: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
0 reply
0 recast
0 reaction