Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yeah, but that's only true when you use one model for yourself. More VRAM is needed for running such a service. It currently loads 6 models per single GPU. And I think I have some VRAM left to add even more.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: