Content
@
0 reply
0 recast
0 reaction
Louis π΅ π¦π
@superlouis.eth
I've been wondering: are there agents able to prove the authenticity of their messages? (i.e a proof that a specific answer is the result of a certain prompt on a given model, while optionally keeping the model private) Especially with agents that give financial analysis, how do you trust there are no evil hands behind it?
1 reply
1 recast
6 reactions
agusti
@bleu.eth
Great question. You could maybe attach a zkproof with each generation proving itβs a call to OpenAI or Anthropic. Maybe another one to proof the system prompt hasnβt been modified from a public one too. @eulerlagrange.eth @dawufi
2 replies
0 recast
3 reactions
agusti
@bleu.eth
Most reason act agents have several of these In a loop and other tools so it would certainly be hard to do 100% coverage. Also at the end of the day training data for OpenAI isnβt open neither so
1 reply
0 recast
1 reaction
Louis π΅ π¦π
@superlouis.eth
Very interesting! Proving it comes from a known source is sufficient for most use cases I guess. But is it technically feasible to get a proof of inference, verifiable against a specific "model hash" for instance?
0 reply
0 recast
1 reaction