Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

There is zero evidence that GPT-4 is a trillion parameter model and some evidence that it is the same parameter size as GPT-3.

Newer research and models out of DeepMind etc. show that GPT-4 performance is attainable at 90% the size of GPT-3. It's well known by now that GPT-3 was hugely over parameterized.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: