Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I am actually not quite sure what would be best to recommend, as the pandemic has seen me lag behind the zeitgeist somewhat. A recent favourite would be the Toolformer paper [1], that I intend to read in detail later today. If an LLM would be able to use external tools efficiently, it could be rather powerful and perhaps allow us to scale down the parameter sizes, which somewhat fascinates me.

[1]: https://arxiv.org/abs/2302.04761

Other research questions, but without concrete papers to reference to that current keeps me up at night: 1.) to which degree can we train substantially smaller LLMs for specific tasks that could be run in-house, 2.) it seems like these new breakthroughs may need a different mode of evaluation compared to what we have used since the 80s in the field and I am not sure what that would look like (maybe along the lines of HELM [2]?), and 3.) can AI academia continue to operate in the way it currently does with small research teams or is a change towards what we see in the physical sciences necessary?

[2]: https://arxiv.org/abs/2211.09110



Toolformer seems to already have been productized in ChatGPT Plugins fwiw

> 1.) to which degree can we train substantially smaller LLMs for specific tasks that could be run in-house 2.) it seems like these new breakthroughs may need a different mode of evaluation compared to what we have used since the 80s in the field and I am not sure what that would look like (maybe along the lines of HELM [2]?)

so you are proposing a set of benchmarks for domain specific tasks? by definition they wont be shared benchmarks...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: