Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Wait a few month and they will have a distilled model with the same performance and 1% of the run cost.


100X efficiency improvement (doubtful) still means that costs grow 200X faster than benchmark performance.


Even assuming that past rates of inference cost scaling hold up, we would only expect a 2 OoM decrease after about a year or so. And 1% of 3.5b is still a very large number.


And to your point "past performance is not indicative of future results". The extrapolate to infinity approach is the mindfever of this field.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: