When is TensorFlow's ParameterServerStrategy preferable to its MultiWorkerMirroredStrategy?
问题 When training a neural network across multiple servers and GPUs, I can't think of a scenario where the ParameterServerStrategy would be preferable to the MultiWorkerMirroredStrategy . What are the ParameterServerStrategy 's main use cases and why would it be better than using MultiWorkerMirroredStrategy ? 回答1: MultiWorkerMirroredStrategy is intended for synchronous distributed training across multiple workers, each of which can have multiple GPUs ParameterServerStrategy : Supports parameter