Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

As an example, INT8 support in WebGPU would enable running quantized models, allowing larger LLMs to run locally in the browser.

See Limitations section here: https://fleetwood.dev/posts/running-llms-in-the-browser



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: