🚀 The feature
The followings are the requirements for DistributedReadingService
- Automatic sharding
- Shard based on Rank, WorldSize, num_workers
- Determinism
- Generate the same seed deterministically for all
Shuffler per epoch.
- Generate different seeds for process-local RNGs to perform different random transformation.
- FullSyncDataPipe
Motivation, pitch
Make DataLoader2 working in distributed training with all syntax sugar.
Alternatives
No response
Additional context
No response