Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

They can when there are entire teams dedicated to adding guardrails via hidden system prompts and running all responses through other LLMs trained on flagging and editing certain things before the original output gets relayed to the user.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: