Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Aren’t most of the models at the top not suitable for text generation? That’s what makes gpt different from Bert


What are the models at the top used for? Excuse my ignorance.


Mostly mask fill, but Transformers can be fine tuned to downstream tasks relatively easily (T5 was built for translation but is used for autocomplete in many cases)


would you mind sharing some references (or even just googleable terms) for this process of fine tuning?




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: