Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This is not truly solvable. There is an extremely strong outer loop of optimization operating here: we want it.

We will use models that make us feel good over models that don't make us feel good.

This one was a little too ham-fisted (at least, for the sensibilities of people in our media bubble; though I suspect there is also an enormous mass of people for whom it was not), so they turned it down a bit. Later iterations will be subtler, and better at picking up the exact level and type of sycophancy that makes whoever it's talking to unsuspiciously feel good (feel right, feel smart, feel understood, etc).

It'll eventually disappear, to you, as it's dialed in, to you.

This may be the medium-term fate of both LLMs and humans, only resolved when the humans wither away.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: