Content
@
https://warpcast.com/~/channel/theai
0 reply
0 recast
0 reaction
Giuliano Giacaglia
@giu
This is pretty interesting. DeepSeek, a model that is outperforming all other LLMs at a small size, seems to be trained on the output of frontier models like GPT-4, which is against their TOS (terms of service). That would explain how they trained such a performant and small model with not as many resources as other labs https://x.com/giffmana/status/1872586401436627211
1 reply
0 recast
16 reactions
Max Miner
@mxmnr
Interesting. I’ve also seen early reports that its ‘creativity’ and storytelling abilities are handicapped (for lack of a better way to say that). High performance in math/science but lack of ability in narrative/persuasion/story
0 reply
0 recast
1 reaction