Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> The MD5 hash for the string "asdns" is "5a5a5ad63c5a5a5a5a5a5a5a5a5a5a5".

I don't need a computer to say that isn't correct ;)

Wow I just reset the thread, used the exact same input and got this (literally 60 seconds apart):

I'm sorry, but I am unable to generate an MD5 hash for you because I am a text-based AI assistant and do not have the ability to perform cryptographic calculations. An MD5 hash is a complex mathematical function that is used to create a unique digital fingerprint for a file or piece of data, and it would require a specialized program or calculator to generate one.



>> Wow I just reset the thread, used the exact same input and got this (literally 60 seconds apart):

It could mean that a lot of people had the same idea as you recently and OpenAI just figured it out.


It's neither. It's randomized. The system is not self-training in real time.


> The system is not self-training in real time.

There are quite a few credible indications/anecdotes (including my own) that there is real-time adaptation going on to these "jailbreak" attempts, which seems to be getting harder and harder to do, even if you use different wording/synonyms. And it seems to be happening gradually.

Now if that's a result of OpenAI doing that manually somehow or ChatGPT fine-tuning its behavior in response to human interactions, I don't know. I'd guess it's actually OpenAI doing very specific fine-tuning as additional training, but I could be wrong.

Note how there is also a thumbs-up/thumbs-down button that you can use to give feedback about what you think of ChatGPT's responses. This feedback may be getting used (although I'd guess it would just get used in a future iteration, not necessarily this one).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: