Shortcuts

DistributedReadingService

class torchdata.dataloader2.DistributedReadingService(timeout: int = 1800)

DistributedReadingSerivce handles distributed sharding on the graph of DataPipe and guarantee the randomness by sharing the same seed across the distributed processes.

Parameters:

timeout – Timeout for operations executed against the process group in seconds. Default value equals 30 minutes.

finalize() None

Clean up the distributed process group.

initialize(datapipe: Union[IterDataPipe, MapDataPipe]) Union[IterDataPipe, MapDataPipe]

Launches the gloo-backend distributed process group. Carries out distributed sharding on the graph of DataPipe and returnes the graph attached with a FullSyncIterDataPipe at the end.

initialize_iteration() None

Shares the same seed from rank 0 to other ranks across the distributed processes and apply the random seed to the DataPipe graph.

Docs

Access comprehensive developer documentation for PyTorch

View Docs

Tutorials

Get in-depth tutorials for beginners and advanced developers

View Tutorials

Resources

Find development resources and get your questions answered

View Resources