TF Dev Summit ‘19 | Mesh-TensorFlow: Model Parallelism for Supercomputers
TF Dev Summit ‘19 | Mesh-TensorFlow: Model Parallelism for Supercomputers Batch-splitting (data-parallelism) is the dominant distributed Deep Neural Network (DNN) training strategy, due to its universal applicability and its amenability…
Share