Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

GPT-4 with a system prompt is definitely better, but better isn't good enough: for a security issue like this we need a 100% reliable solution, or people WILL figure out how to exploit it.


I agree and I think OpenAI is aware of this issue. I’ve read the other day that they are employing a “red team” - group of people that basically are tasked with breaking their models so they can find weak spots and fix them.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: