Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

counterpoint: thinking models are good since they give similar quality at smaller RAM sizes. if a 16b thinking model is as good as a 60b one shot model, you can use more compute without as much RAM bottleneck


Counter-counterpoint: RAM costs are coming down fast this year. Compute, not so much.

I still agree, though.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: