Charlie Marketplace
@charliemktplace
𧡠How to think about building AI models (LLM Edition) How do LLMs work? π How do you scope a custom LLM? π What are the tradeoffs between popular Foundational Models? π± Is Context enough? π When is tuning worth it? β
1 reply
0 recast
0 reaction
Charlie Marketplace
@charliemktplace
[Instructions] + [Context] + [System Prompt] -> [Tuning filter] ->[Foundation Model] Instruction: summarize the main points of this article Context: the article System Prompt: behavior, censorship, capabilities (e.g., web browsing) Tuning Filter: specialize a model Foundational Model: GPT 3.5; Claude; Gemini; Llama-2
1 reply
0 recast
0 reaction
Charlie Marketplace
@charliemktplace
How do you scope a custom LLM? π Have a specific goal: search? summarize? generate? Understand the resource constraints: proof of concept vs horizontal model (mixture of experts) vs vertical model (master many domains). The more ambitious the goal, the more likely you need tuning.
1 reply
0 recast
0 reaction
Charlie Marketplace
@charliemktplace
What are the tradeoffs between popular Foundational Models? π± In general, with enough context and tuning, models of similar size/complexity converge in quality. version control (deterministic testing), cost control, enterprise support, hosting, and avoiding getting rugged! LLMs = Chaotic = easy to break behavior.
1 reply
0 recast
0 reaction
Charlie Marketplace
@charliemktplace
Context is the easiest way to format a model's behavior. Simply *append* relevant information to every single user's request (i.e., exploit large context windows) Be careful: these can be *misleadingly* good in proof of concepts, but struggle to be aMaZinG in the wild https://warpcast.com/charliemktplace/0xbdce4dd1
1 reply
0 recast
0 reaction