Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think this is the key difference between current LLMs and humans: an LLM will act based on the given prompt, while a human being may have “principles” that cannot betray even if they are being pointed with gun to their heads.

I think the LLM simply correlated the given prompt to the most common pattern in its training: blackmailing.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: