I think this is the key difference between current LLMs and humans: an LLM will act based on the given prompt, while a human being may have “principles” that cannot betray even if they are being pointed with gun to their heads.
I think the LLM simply correlated the given prompt to the most common pattern in its training: blackmailing.
I think the LLM simply correlated the given prompt to the most common pattern in its training: blackmailing.