Hacker News new | past | comments | ask | show | jobs | submit login

You can bypass the system prompt by using the API? I thought part of the "safety" of LLMs was implemented with the system prompt. Does that mean it's easier to get unsafe answers by using the API instead of the GUI?



Safety is both the system prompt and the RLHF posttraining to refuse to answer adversarial inputs.


Yes, it is.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: