Loading tasks/finetune_utils.py +0 −2 Original line number Diff line number Diff line Loading @@ -90,8 +90,6 @@ def build_data_loader(dataset, micro_batch_size, num_workers, drop_last, sampler = torch.utils.data.distributed.DistributedSampler( dataset, num_replicas=world_size, rank=rank) print_rank_0(len(sampler)) # Data loader. Note that batch size is the per GPU batch size. data_loader = torch.utils.data.DataLoader(dataset, batch_size=micro_batch_size, Loading Loading
tasks/finetune_utils.py +0 −2 Original line number Diff line number Diff line Loading @@ -90,8 +90,6 @@ def build_data_loader(dataset, micro_batch_size, num_workers, drop_last, sampler = torch.utils.data.distributed.DistributedSampler( dataset, num_replicas=world_size, rank=rank) print_rank_0(len(sampler)) # Data loader. Note that batch size is the per GPU batch size. data_loader = torch.utils.data.DataLoader(dataset, batch_size=micro_batch_size, Loading