Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

"Running" and "acceptable inference speed and quality" are two different constraints, particularly at scale/production.


I don't understand what you're trying to say ?

From what I've read 4090 should blow A100 away if you can fit within 22GB VRAM, which a 7B model should comfortably.

And the latency (along with variability and availability) on OpenAI API is terrible because of the load they are getting.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: