Content pfp
Content
@
0 reply
0 recast
0 reaction

assayer pfp
assayer
@assayer
AI SAFETY COMPETITION (15) 1100 DEGEN AWARDS! Big history writer Yuval Harari points to the TaskRabbit accident, where GPT4 already fooled a human to solve CAPTCHA. "What everybody should understand about AI", Harari says, "AI is not a tool, it is an agent". (video below, 1:15-3:15) 600 degen for the best comment! II AWARD - 300 degen III AWARD - 100 degen Deadline: 7.00 pm tomorrow, Tuesday ET time (28 hours) AI-generated responses will be disqualified. https://www.youtube.com/watch?v=OcWdu3t7noo
3 replies
0 recast
0 reaction

Sophia Indrajaal pfp
Sophia Indrajaal
@sophia-indrajaal
I think in that incident, which I was unfamiliar with, AI was setting out to complete a task given to it by a human. Calling it an agent rather than a tool isn't totally accurate, it is both. I don't think the proposal made, to make AIs declare themselves is necessarily feasible nor does it address deeper issues. I am more and more convinced that we need deeper levels of safety baked in to AI from the get go. Regulation won't be sufficient, and likely not really necessary if it's done right. I'm working on a dual safety mechanism, that lends itself to working with other alignment approaches.
2 replies
0 recast
1 reaction

assayer pfp
assayer
@assayer
thank you for the comment! Imo the agent idea is supported by OpenAI's own "conclusions". Like: "Current language models are (...) very much capable of convincing humans to do things for them". That is clearly agent-like, not tool-like. source: https://metr.org/blog/2023-03-18-update-on-recent-evals/ (do you have a connected wallet already?)
1 reply
0 recast
1 reaction

assayer pfp
assayer
@assayer
III award, congrats! 100 $degen
1 reply
0 recast
0 reaction