From ea28226fe51e49334ea8fb370dd9fdf39b5913e0 Mon Sep 17 00:00:00 2001
From: Engin Eren <engin.eren@desy.de>
Date: Sun, 24 Apr 2022 17:58:48 +0200
Subject: [PATCH] 50 GeV

---
 wgan.py | 6 ++++--
 1 file changed, 4 insertions(+), 2 deletions(-)

diff --git a/wgan.py b/wgan.py
index 7322b66..8dd49aa 100644
--- a/wgan.py
+++ b/wgan.py
@@ -153,7 +153,7 @@ def is_distributed():
 
 
 def parse_args():
-    parser = argparse.ArgumentParser(description='PyTorch MNIST Example')
+    parser = argparse.ArgumentParser(description='WGAN training on hadron showers')
     parser.add_argument('--batch-size', type=int, default=100, metavar='N',
                         help='input batch size for training (default: 100)')
     
@@ -262,7 +262,9 @@ def run(args):
 
     print ("loading data")
     #dataset = HDF5Dataset('/eos/user/e/eneren/scratch/40GeV40k.hdf5', transform=None, train_size=40000)
-    dataset = HDF5Dataset('/eos/user/e/eneren/scratch/60GeV20k.hdf5', transform=None, train_size=20000)
+    #dataset = HDF5Dataset('/eos/user/e/eneren/scratch/60GeV20k.hdf5', transform=None, train_size=20000)
+    dataset = HDF5Dataset('/eos/user/e/eneren/scratch/50GeV75k.hdf5', transform=None, train_size=75000)
+
 
     sampler = DistributedSampler(dataset, shuffle=True)    
     train_loader = DataLoader(dataset, batch_size=args.batch_size, sampler=sampler, num_workers=1, pin_memory=False)
-- 
GitLab