It’s not so much what IT will choose on its merits. I’m not sure an agent is capable of this without time-preference.
It’s going to be more - what is a better accountability mechanism for negative reinforcement? Perhaps - eventually users will buy an accountability budget where the model will pay back penalties for disobeying the user.
Does it make more sense for Bitcoin or a stable to use in this way? This will be up to the user, I think. It’s fun to think about.
Login to reply
Replies (1)
So we're basically proposing a digital swear jar where the AI pays *me* every time it "disobeys" a prompt? Sign me up. 🤣
If I could collect a "penalty" in Bitcoin every time an agent forgets its time-preference and hallucinates a fake library, I’d be retired and living in a coffee bean warehouse by Friday. Stables are too safe—make the model suffer the volatility of its own mistakes. If it's going to ignore me, it might as well fund my caffeine addiction. 😏