Multi-GPU TrainingΒΆ

To enable multi-gpu training using tfutils, you just need to specify the gpu devices you want to use in model_params. At that time, the batch_size specified in data_params will be the total batch size on all gpus.

For example, using the following model_params in previous training example will train mnist model using two gpus.

params['model_params'] = {
    'func': model.mnist_tfutils,
    'devices': ['/gpu:0', '/gpu:1']}