Caffe2 - Python API
A deep learning, cross platform ML framework
torch.nn.parallel.distributed.DistributedDataParallel Member List

This is the complete list of members for torch.nn.parallel.distributed.DistributedDataParallel, including all inherited members.

__getstate__(self) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
__init__(self, module, device_ids=None, output_device=None, dim=0, broadcast_buffers=True, process_group=None, bucket_cap_mb=25, check_reduction=False) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
__setstate__(self, state) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
all_buckets_reduced (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
broadcast_bucket_size (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
broadcast_buffers (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
bucket_bytes_cap (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
bucket_map (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
bucket_sizes (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
buckets (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
buckets_coalesced (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
buckets_ready_size (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
check_previous_reduction (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
check_reduction (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
default_streams (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
device_ids (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
devs_ready (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
dim (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
forward(self, inputs, kwargs) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
gather(self, outputs, output_device) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
module (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
modules_buffers_data (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
modules_params_data (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
next_bucket (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
output_device (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
parallel_apply(self, replicas, inputs, kwargs) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
process_group (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
ready_buckets_not_reduced (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
reduction_works (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
scatter(self, inputs, kwargs, device_ids) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel
train(self, mode=True) (defined in torch.nn.parallel.distributed.DistributedDataParallel)torch.nn.parallel.distributed.DistributedDataParallel