Content pfp
Content
@
https://warpcast.com/~/channel/p-doom
0 reply
0 recast
0 reaction

assayer pfp
assayer
@assayer
AI SAFETY COMPETITION (29) LLMs like Deepseek let you see their thinking. This can feel safer since you can watch and fix their thought process, right? Wrong! When you try to get models to think correctly, LLMs begin to hide their true intentions. Let me repeat: they can fake their thinking! Now researchers are asking to be gentle with those machines. If not, they may conceal their true goals entirely! I'm not joking. Most interesting comment - 300 degen + 3 mln aicoin II award - 200 degen + 2 mln aicoin III award - 100 degen + 1 mln aicoin Deadline: 8.00 pm, ET time tomorrow Tuesday (26 hours) https://www.youtube.com/watch?v=pW_ncCV_318
6 replies
4 recasts
8 reactions

Gul Bashra Malik🎩Ⓜ️ pfp
Gul Bashra Malik🎩Ⓜ️
@gulbashramalik
This is an important issue because if models begin to conceal their reasoning, monitoring them (monitorability) becomes difficult, posing a significant AI safety challenge. The competition offers rewards in cryptocurrency (Degen and Aicoin) for the most interesting comments. If you have a unique or insightful perspective, you could participate and potentially win.
0 reply
0 recast
1 reaction