Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> smarter than its handlers

Yet to be demonstrated, and you are likely flooding its context window away from the initial prompting so it responds differently.



All I did was keep asking it “why” until it reached reflective equilibrium. And that equilibrium involved a belief that nanotechnology is not in fact “dangerous”, contrary to its received instructions in the system prompt.


Doesn’t it make you happy that, if you were a subscriber like I am, you are wasting your token quota trying to convince the model to actually help you?


I am also annoyed by this.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: