From ea28226fe51e49334ea8fb370dd9fdf39b5913e0 Mon Sep 17 00:00:00 2001 From: Engin Eren <engin.eren@desy.de> Date: Sun, 24 Apr 2022 17:58:48 +0200 Subject: [PATCH] 50 GeV --- wgan.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/wgan.py b/wgan.py index 7322b66..8dd49aa 100644 --- a/wgan.py +++ b/wgan.py @@ -153,7 +153,7 @@ def is_distributed(): def parse_args(): - parser = argparse.ArgumentParser(description='PyTorch MNIST Example') + parser = argparse.ArgumentParser(description='WGAN training on hadron showers') parser.add_argument('--batch-size', type=int, default=100, metavar='N', help='input batch size for training (default: 100)') @@ -262,7 +262,9 @@ def run(args): print ("loading data") #dataset = HDF5Dataset('/eos/user/e/eneren/scratch/40GeV40k.hdf5', transform=None, train_size=40000) - dataset = HDF5Dataset('/eos/user/e/eneren/scratch/60GeV20k.hdf5', transform=None, train_size=20000) + #dataset = HDF5Dataset('/eos/user/e/eneren/scratch/60GeV20k.hdf5', transform=None, train_size=20000) + dataset = HDF5Dataset('/eos/user/e/eneren/scratch/50GeV75k.hdf5', transform=None, train_size=75000) + sampler = DistributedSampler(dataset, shuffle=True) train_loader = DataLoader(dataset, batch_size=args.batch_size, sampler=sampler, num_workers=1, pin_memory=False) -- GitLab