Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That doesn't work either. It's always possible to come up with an attack which subverts the "moderator" model first.

Using non-deterministic AI to protect against attacks against non-deterministic AI is a bad approach.



So you just need another agent to review the data being passed to the protector agent. Easy-peasy.

Use my openAI referral code #LETITRAIN for 10% off!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: