Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

In general, most giant LLMs are extremely undertrained at this time. Consider that most of the gains in RoBerta vs bert were from just continuing to train.


Cases of undertraining can be observed whenever the output is repeating gibberish or loops. Happened a lot in GPT2 ai dungeon days


So can we continue training RoBERTa to get it to, say, GPT3 Ada level




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: