| __init__(self, module, device_ids=None, output_device=None, dim=0) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| device_ids (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| dim (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| forward(self, inputs, kwargs) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| gather(self, outputs, output_device) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| module (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| output_device (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| parallel_apply(self, replicas, inputs, kwargs) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| replicate(self, module, device_ids) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| scatter(self, inputs, kwargs, device_ids) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
| src_device_obj (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |