Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Good lord, hopefully latency isn't 2.5 seconds!


Latency in training literally does not matter. You care about throughput. In serving, where latency matters, most DL frameworks allow you to serve the model from a highly optimized C++ binary, no python needed.

The poster you are replying to is 100% correct.


quote is 'data spend 2500 ms on the wire'. That's not latency. For a nice 10GbE connection, that's optimistically 3 GB or so worth of data. Do you have 3GB of training data? Then it will spend 2500 ms on the wire to distribute to all of your nodes as part of startup.


I can’t even. How could you ever get 2500 msec on transit? That’s like circling the globe ten times.


Maybe a bunch of SSL cert exchanges through some very low bandwidth connections? ;)

Still, it's more likely a figure used for exaggeration, for effect.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: