Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It was trained to model the game. LLMs are trained to model language. Neither are trained to take over the world.


It was not trained to model the game. It was trained to predict the next token based on a sequence of previous tokens, which it wasn't even told are moves in a game, much less how to parse them. And it came up with an internal model of the game based on that that's accurate enough to include the board state. You could say that it "understands" the game at that point, even though it wasn't specifically trained to do that.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: