Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
alienthrowaway
10 months ago
|
parent
|
context
|
favorite
| on:
Google will let companies run Gemini models in the...
What are the pros of using CUDA-enabled devices for inference?
bitexploder
10 months ago
[–]
My limited understanding is that CUDA wins on smaller batches and jobs but TPU wins on larger jobs. It is just easier to use and better at typical small workloads. At some point for bigger ML loads and inference TPU starts making sense.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: