Distributed TensorFlow training (Google I/O '18)

To efficiently train machine learning models, you will often need to scale your training to multiple GPUs, or even multiple machines. TensorFlow now offers rich functionality to achieve this with just a few lines of code. Join this session to learn how to set this up.

Rate this session by signing-in on the I/O website here → https://goo.gl/sBZMEm

Distribution Strategy API:
https://goo.gl/F9vXqQ
https://goo.gl/Zq2xvJ

ResNet50 Model Garden example with MirroredStrategy API:
https://goo.gl/3UWhj8

Performance Guides:
https://goo.gl/doqGE7
https://goo.gl/NCnrCn

Commands to set up a GCE instance and run distributed training:
https://goo.gl/xzwN4C

Multi-machine distributed training with train_and_evaluate:
https://goo.gl/kyikAC

Watch more TensorFlow sessions from I/O ’18 here → https://goo.gl/GaAnBR
See all the sessions from Google I/O ’18 here → https://goo.gl/q1Tr8x

Subscribe to the TensorFlow channel → https://goo.gl/ht3WGe

#io18 event: Google I/O 2018; re_ty: Publish; product: TensorFlow – General; fullname: Priya Gupta, Anjali Sridhar; event: Google I/O 2018;

Source of this TensorFlow AI Video

AI video(s) you might be interested in …