| import torch |
| import math |
| import numpy as np |
|
|
| class SequentialDistributedSampler(torch.utils.data.sampler.Sampler): |
| """ |
| Distributed Sampler that subsamples indicies sequentially, |
| making it easier to collate all results at the end. |
| Even though we only use this sampler for eval and predict (no training), |
| which means that the model params won't have to be synced (i.e. will not hang |
| for synchronization even if varied number of forward passes), we still add extra |
| samples to the sampler to make it evenly divisible (like in `DistributedSampler`) |
| to make it easy to `gather` or `reduce` resulting tensors at the end of the loop. |
| """ |
|
|
| def __init__(self, dataset, batch_size, rank=None, num_replicas=None): |
| if num_replicas is None: |
| if not torch.distributed.is_available(): |
| raise RuntimeError("Requires distributed package to be available") |
| num_replicas = torch.distributed.get_world_size() |
| if rank is None: |
| if not torch.distributed.is_available(): |
| raise RuntimeError("Requires distributed package to be available") |
| rank = torch.distributed.get_rank() |
| self.dataset = dataset |
| self.num_replicas = num_replicas |
| self.rank = rank |
| self.batch_size = batch_size |
| self.num_samples = int(math.ceil(len(self.dataset) * 1.0 / self.batch_size / self.num_replicas)) * self.batch_size |
| self.total_size = self.num_samples * self.num_replicas |
|
|
| def __iter__(self): |
| indices = list(range(len(self.dataset))) |
| |
| indices += [indices[-1]] * (self.total_size - len(indices)) |
| |
| indices = indices[self.rank * self.num_samples : (self.rank + 1) * self.num_samples] |
| return iter(indices) |
|
|
| def __len__(self): |
| return self.num_samples |
|
|
|
|
| def distributed_concat(tensor, num_total_examples): |
| output_tensors = [tensor.clone() for _ in range(torch.distributed.get_world_size())] |
| torch.distributed.all_gather(output_tensors, tensor) |
| concat = torch.cat(output_tensors, dim=0) |
| return concat[:num_total_examples] |