Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

A solution might be: use two different AIs. The first one you can prompt to your heart's content. The second one is never prompted by anyone except the service provider. The second one does the filtering.


If it's filtering by taking the output of the first model as a prompt (with some framing), then that is equally susceptible to prompt engineering. Indeed, you can already tell ChatGPT to write a prompt for itself to do such and such, and it will do so. You can even tell it to write a prompt to write a prompt.


Should we call the second AI Conscience?


We shall call it Amygdata.


Yes, Lieutenant Amygdata. This is the way.


A Deaf lieutnant! If you are really Deaf, dear deafpolygon, please contact me. See my profile, there is a (hidden) way to contact me.


Indeed, I am.


Super-ego


Yeah, like a parent AI




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: