__init__(self, module, device_ids=None, output_device=None, dim=0) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
device_ids (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
dim (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
forward(self, inputs, kwargs) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
gather(self, outputs, output_device) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
module (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
output_device (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
parallel_apply(self, replicas, inputs, kwargs) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
replicate(self, module, device_ids) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
scatter(self, inputs, kwargs, device_ids) (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |
src_device_obj (defined in torch.nn.parallel.data_parallel.DataParallel) | torch.nn.parallel.data_parallel.DataParallel | |