Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm not convinced that thinking tokens - which sort of have to serve a specific chain of thought purpose - are interchangeable with input tokens during which give the model compute without having it add new text.

For a very imperfect human analogy, it feels like saying "a student can spend as much time thinking about the text as they want, so the textbook can be extremely terse".

Definitely just gut feelings though - not well tested or anything. I could be wrong.





We could definitely use RL to add blank, invisible "<thinking>" tokens whenever the model thinks it should. Or just allow it to say "Hmm." lol.

Easy to test from a technical perspective is all I'm saying, and not a bad idea.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: