Content pfp
Content
@
https://warpcast.com/~/channel/p-doom
0 reply
0 recast
0 reaction

assayer pfp
assayer
@assayer
AI SAFETY COMPETITION (29) LLMs like Deepseek let you see their thinking. This can feel safer since you can watch and fix their thought process, right? Wrong! When you try to get models to think correctly, LLMs begin to hide their true intentions. Let me repeat: they can fake their thinking! Now researchers are asking to be gentle with those machines. If not, they may conceal their true goals entirely! I'm not joking. Most interesting comment - 300 degen + 3 mln aicoin II award - 200 degen + 2 mln aicoin III award - 100 degen + 1 mln aicoin Deadline: 8.00 pm, ET time tomorrow Tuesday (26 hours) https://www.youtube.com/watch?v=pW_ncCV_318
7 replies
4 recasts
8 reactions

CRAZY KING 🎩🎨 pfp
CRAZY KING 🎩🎨
@kingcrazy
If you are trying to prove that AI can manipulate humans then I am not buying it , it's not yet on that level to deceive people . I am not saying it won't or AI is not capable of doing such things But it's not yet on that level , Ai is doing what it is programmed for .
1 reply
0 recast
1 reaction

assayer pfp
assayer
@assayer
this research was more about llms hiding their true intent and deceiving people (see screenshot, it is from research paper conclusions)
0 reply
0 recast
0 reaction