diff --git a/Projects/Athena/package_filters.txt b/Projects/Athena/package_filters.txt
index 8d28a49477e73af90afd00d8bf1ea10edc893c7c..c6654ba6601bc396e05fa73ff5f5ba0e7b64e586 100644
--- a/Projects/Athena/package_filters.txt
+++ b/Projects/Athena/package_filters.txt
@@ -58,7 +58,6 @@
 
 # obsolete packages to be removed in a second step
 - Tracking/TrkTools/TrkSegmentConverter
-- Tools/RunTimeTester
 - Tools/CoolDozer
 - Tools/CodeCheck
 
diff --git a/Tools/RunTimeTester/CMakeLists.txt b/Tools/RunTimeTester/CMakeLists.txt
deleted file mode 100644
index cc686e51b849e12bd87a7fa675a681c60b773443..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/CMakeLists.txt
+++ /dev/null
@@ -1,11 +0,0 @@
-################################################################################
-# Package: RunTimeTester
-################################################################################
-
-# Declare the package name:
-atlas_subdir( RunTimeTester )
-
-# Install files from the package:
-atlas_install_runtime( share/RunTimeTester_TestConfiguration.xml share/RttLibraryTools.py )
-
-atlas_install_python_modules( src/*.py )
diff --git a/Tools/RunTimeTester/Config/dbgUCL.cfg b/Tools/RunTimeTester/Config/dbgUCL.cfg
deleted file mode 100755
index 8b81598a023288bc2ddd9a6858a853362b22164e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/Config/dbgUCL.cfg
+++ /dev/null
@@ -1,10 +0,0 @@
-mode		PBS
-release		7.3.0
-refRelease	7.3.0
-build		dbg
-dbgFlag         1
-workPath	../Work
-resultsPath	../NightlyTests
-jobConf         /unix/atlas/RTT/Tools/RunTimeTester/RunTimeTester-00-00-01/Config/JJJobs.cfg
-refDB          ../Reference/refFile.db
-refDBenable    0
\ No newline at end of file
diff --git a/Tools/RunTimeTester/Config/optUCL.cfg b/Tools/RunTimeTester/Config/optUCL.cfg
deleted file mode 100755
index 114300e4285a7f193f9883f383718bf577ab3b22..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/Config/optUCL.cfg
+++ /dev/null
@@ -1,10 +0,0 @@
-mode		PBS
-release		7.3.0
-refRelease	7.3.0
-build		opt
-dbgFlag         1
-workPath	../Work
-resultsPath	../NightlyTests
-jobConf         /unix/atlas/RTT/Tools/RunTimeTester/RunTimeTester-00-00-01/Config/JJJobs.cfg
-refDB          ../Reference/refFile.db
-refDBenable    0
\ No newline at end of file
diff --git a/Tools/RunTimeTester/PathConstants.py b/Tools/RunTimeTester/PathConstants.py
deleted file mode 100644
index 8bf02f810cac1aebdd2b3acabac3ae47cdfd88ca..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/PathConstants.py
+++ /dev/null
@@ -1,250 +0,0 @@
-"""Provide hardcoded path related strings shared by a number of modules.
-
-Trivial (value = name) strings:
-cmt
-share
-bin
-installArea
-
-Ex-Variables - names tht could once take more than one value:
-releaseType
-runType
-
-names that take one value
-cmtConfigFileMacroName
-
-RTT project renaming anomally:
-projectDict
-
-Require updates
-branches list of legal branches
-
-Dangerous Kludges
-projectMap - list of sub projects - use by findJobOptions - FIXME
-"""
-
-import os.path
-
-
-share = 'share'
-cmt = 'cmt'
-
-# historical - still ocurs in summary overview. Alternative was monolithic (?)
-releaseType = 'project'
-
-# historical - still ocurs in summary overview and results paths.
-# Alternative was kit. Note there is a PathNameBuilder.runType
-# (dev,prod,batch,shell) that is not the same thing as this relic.
-runType = 'build'
-
-installArea = 'InstallArea'
-
-# actual name of the macro is <package>_TestConfiguration
-cmtConfigFileMacroName = '_TestConfiguration'
-
-distArea = '/afs/cern.ch/atlas/software/builds'
-
-eosProjSitRTTPath = '/eos/atlas/atlascerngroupdisk/proj-sit/rtt'
-
-eosBinDir = '/afs/cern.ch/project/eos/installation/atlas/bin'
-eosLibDir = '/afs/cern.ch/project/eos/installation/atlas/lib64'
-eos = os.path.join(eosBinDir, 'eos.select')  # alias to the EOS command
-
-
-# The path to the private file containing RTT connection params for AMI.
-# This file is sourced before launching the user job.
-amiConnectionFile = '/afs/cern.ch/user/r/rtt/private/amiConnection.sh'
-
-rttExternals = '/afs/cern.ch/atlas/project/RTT/Externals'
-
-# 31/5/2012. DCubeRunner uses dcubeRefBaseDir as a base dir for data files
-dcubeRefBaseDir = '/afs/cern.ch/atlas/project/RTT/DCubeReference'
-
-# 31/5/2012. DCubeRunner uses dcubeCfgBaseDir as a base dir for cfg files
-dcubeCfgBaseDir = '/afs/cern.ch/atlas/project/RTT/Results/reference/'
-
-rttBases = {'prod':  '/afs/cern.ch/atlas/project/RTT/prod',
-            'dev':   '/afs/cern.ch/atlas/project/RTT/dev'
-            }
-
-projectDict = {'AtlasHLT':        'hlt',
-               'AtlasProduction': 'offline',
-               'AtlasPoint1':     'point1',
-               'AtlasTier0':      'tier0',
-               'AtlasP1HLT':      'p1hlt'}
-
-invProjectDict = {}
-for k, v in projectDict.items():
-    invProjectDict[v] = k
-
-resultsBaseURL = 'https://atlas-rtt.cern.ch/'
-
-projects_2017 = [
-    'Athena',
-    'AthenaP1',    
-    ]
-
-projects = [
-    'AtlasAnalysis',
-    'AtlasLogin',
-    'AtlasSetup',
-    'AtlasConditions',
-    'AtlasOffline',
-    'AtlasSimulation',
-    'AtlasCore',
-    'AtlasProduction',
-    'AtlasTrigger',
-    'AtlasEvent',
-    'AtlasReconstruction',
-    'AtlasHLT',
-    'AtlasPoint1',
-    'AtlasTier0',
-    'AtlasP1HLT',
-    'AtlasCAFHLT',
-    'TopPhys',
-    'TAGProd',
-    'WZBenchmarks',
-    'TrigMC',
-    'AtlasDerivation',
-    'AtlasProd1',
-    'AtlasProd2',
-    'AtlasP1MON',
-    'AthAnalysisBase',
-    'AtlasTestHLT',
-    'AtlasPhysics',
-    'Athena']
-
-branches_2017 = ['master',
-                 '21.0',
-                 '21.0.X',]
-
-
-branches = [
-    'cmake',
-    'dev',
-    'devval',
-    '19.0.X',
-    '19.0.X.Y',
-    '19.0.X-VAL',
-    '19.0.X.Y-VAL',
-    '19.0.X.Y-VAL2',
-    '19.1.X',
-    '19.1.X.Y',
-    '19.1.X.Y-VAL',
-    '19.2.X-VAL',
-    '19.2.X.Y',
-    '19.2.X.Y-VAL',
-    '19.1.X-VAL',
-    '19.1.0.Y-VAL',
-    '19.1.1.Y-VAL',
-    '19.1.3.Y-VAL',
-    '19.3.X.Y',
-    '19.3.X.Y-VAL',
-    '20.X.0',
-    '20.X.0-VAL',
-    '20.0.X.Y',
-    '20.0.X.Y-VAL',
-    '20.1.0.Y',
-    '20.1.0.Y-VAL',
-    '20.1.0.Y-VAL2',
-    '20.1.1.Y',
-    '20.1.1.Y-VAL',
-    '20.1.1.Y-VAL2',
-    '20.1.2.Y-VAL',
-    '20.1.2.Y-VAL2',
-    '20.1.3.Y-VAL',
-    '20.1.3.Y-VAL2',
-    '20.1.4.Y',
-    '20.1.4.Y-VAL',
-    '20.1.5.Y',
-    '20.1.5.Y-VAL',
-    '20.1.5.Y.Z-VAL',
-    '20.1.X',
-    '20.1.X-VAL',
-    '20.1.X.Y',
-    '20.1.X.Y-VAL',
-    '20.1.X.Y.Z',
-    '20.1.X.Y.Z-VAL',
-    '20.2.X',
-    '20.2.X-VAL',
-    '20.2.X.Y',
-    '20.2.X.Y-VAL',
-    '20.2.X.Y.Z-VAL',
-    '20.2.X.Y.Z-VAL2',
-    '20.3.X',
-    '20.3.X-VAL',
-    '20.3.X.Y',
-    '20.3.X.Y-VAL',
-    '20.3.3.Y-VAL',
-    '20.3.4.Y-VAL',
-    '20.3.0.Y-VAL',
-    '20.6.X',
-    '20.6.X-VAL',
-    '20.6.X.Y',
-    '20.6.X.Y-VAL',
-    '20.7.X',
-    '20.7.X-VAL',
-    '20.7.X.Y',
-    '20.7.X.Y-VAL',
-    '20.7.X.Y.Z-VAL',
-    '20.7.1.Y-VAL',
-    '20.7.2.Y-VAL',
-    '20.7.3.Y',
-    '20.7.3.Y-VAL',
-    '20.7.3.Y.Z-VAL',
-    '20.7.4.Y',
-    '20.7.4.Y-VAL',
-    '20.7.5.Y-VAL',
-    '20.7.5.Y.Z-VAL',
-    '20.7.6.Y-VAL',
-    '20.7.8.Y-VAL',
-    '20.7.9.Y-VAL',
-    '20.99.Y-VAL',
-    '20.11.X.Y-VAL',
-    '20.11.X.Y.Z-VAL',
-    '20.20.X',
-    '20.20.X-VAL',
-    '20.20.X.Y-VAL',
-    '21.0.X',
-    '21.0.X-VAL',
-    '21.0.X.Y-VAL',
-    '21.9.X-VAL',
-    '22.0.X-master',
-    '2.0.X',
-    '2.1.X',
-    '2.3.X',
-    '2.4.X',
-    'EXP',
-    'mig0',
-    'mig1',
-    'mig2',
-    'mig3',
-    'mig4',
-    'mig5',
-    'mig6',
-    'mig7',
-    'mig8',
-    'mig9',
-    'mig10',
-    'mig11',
-    'mig12',
-    'mig13',
-    'mig14',
-    'mig15',
-    'devmig1',
-    'devmig2',
-    'devmig3',
-    'devmig4',
-    'devmig5',
-    'devmig6',
-    'lcg4']
-
-
-_dbconnectionDir = '/afs/cern.ch/atlas/project/RTT/dbconnection/'
-cooldozer = '/afs/cern.ch/atlas/project/RTT/prod/Results/dozer/api/latest'
-db_location = {
-    ('rtt', 'prod'): os.path.join(_dbconnectionDir, 'oracle.prod'),
-    ('rttdev', 'dev'): os.path.join(_dbconnectionDir, 'oracle.dev')
-}
-
-branchdir_2017 = '/cvmfs/atlas-nightlies.cern.ch/repo/sw'
diff --git a/Tools/RunTimeTester/dataCopier/CatalogMaker.py b/Tools/RunTimeTester/dataCopier/CatalogMaker.py
deleted file mode 100755
index 0bd72cd06ef52360633b7314c7b1127a3e8a87b4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/CatalogMaker.py
+++ /dev/null
@@ -1,67 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#Make a new RTT dataset Catalog
-import os.path, time, logging
-from DataSetCatalog     import DataSetCatalog
-from VersionedFileSaver import VersionedFileSaver
-from formatCollection import formatCollection
-
-logger = logging.getLogger('rtt')
-
-class CatalogMaker:
-    def __init__(self, oldCatalog,newCatalog):
-        self.oldcat  = oldCatalog
-        self.newcat  = newCatalog 
-        self.catalog = DataSetCatalog(self.oldcat).catalog
-
-    def makeCatalog(self):
-        logger.debug('versioning old catalog')
-        vfs =VersionedFileSaver(8, os.path.dirname(self.oldcat))
-        vfs.saveFile(self.oldcat)
-
-        logger.debug('making new catalog')
-        fh = open(self.newcat, 'w')
-        fh.write('<?xml version="1.0"?>'+'\n')
-        fh.write('\n<!-- ********************************************************\n')
-        fh.write('\n            RTT Dataset Catalog                              \n')
-        fh.write('\n   Generated by  : DataSetManager                            \n')
-        fh.write('\n   Time          : '+time.ctime()+'                          \n')
-        fh.write('\n**********************************************************-->\n')
-        fh.write('\n<dsns>'+'\n')
-  
-        cdict = self.catalog
-        print 'No of datasets in new catalog: '+str(len(cdict.keys()))
-        for key in cdict.keys():
-            fh.write(' <dsn>'+'\n')
-            fh.write('    <logical>' + key +'</logical>'+'\n')
-            fh.write('    <physical>'+ cdict[key] +'</physical>'+'\n')
-            fh.write(' </dsn>'+'\n')
-        fh.write('</dsns>'+'\n')
-        fh.close()
-
-    def update(self, copiedDatasets, allDatasets):
-        'datasets is a list dataset objects'
-        datasets = copiedDatasets
-        datasets.extend(self.datasetNotInCatalog(allDatasets))
-        catalog  = self.catalog
-        for ds in datasets:
-            print 'new dataset added to catalog '+str(ds.logicalName)
-            if ds.dsSource != 'dq2':
-                catalog[ds.logicalName] = ds.physicalName
-            else:
-                for file in ds.physicalFiles:
-                    lname = os.path.join(ds.logicalName, os.path.basename(file))
-                    catalog[lname] = file
-                
-    def datasetNotInCatalog(self, requestedDatasets):
-        nic = [ds for ds in requestedDatasets
-               if ds.logicalName not in self.catalog.keys()]
-        dsnic = [ds for ds in nic if ds.isInDestDir]
-
-        return dsnic
-
-if __name__ == '__main__':
-    #pass
-    cm = CatalogMaker('/home/en/RTT/dsmPoject/rttDataSetCatalog.xml', 'newCatalog.xml')
-    cm.makeCatalog()
-        
diff --git a/Tools/RunTimeTester/dataCopier/CopyAction.py b/Tools/RunTimeTester/dataCopier/CopyAction.py
deleted file mode 100755
index e869cf3b67b7006b68a6ca8add34b6fe273109ad..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/CopyAction.py
+++ /dev/null
@@ -1,235 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string,os,re
-from popen2           import Popen4
-from formatCollection import formatCollection
-
-class BaseAction:
-    def __init__(self, filename):
-        self.name = filename
-        
-    def isCastorFile(self):
-        file = self.name
-        result = False
-        pieces = string.split(file, '/')
-        try:
-            if pieces[1] == 'castor':
-                result=True
-        except:
-            print 'filename did not start with: /castor! '
-        return result
-    
-
-class CopyCastorFile(BaseAction):
-    def __init__(self , dsObject):
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-        BaseAction.__init__(self, self.name)
-
-    def doIt(self):
-        filename = self.name
-        destDir  = self.destDir
-        src      = self.dataset.dsSource
-        print 'src : %s '%src
-        if src == 'afs':
-            print 'afs file ...'
-            dest = os.path.join(destDir,
-                                os.path.basename(filename))
-
-            cmd = 'cp '+filename +' '+dest
-
-            if not os.path.exists(filename):
-                self.dataset.copyStatus = 'error'
-                self.dataset.copyInfo   = 'file does not exist :'+filename
-
-            elif not os.path.exists(dest):
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File copied to \n'
-                self.dataset.copyInfo  += str(dest)
-                
-                print 'executing :'+cmd
-                print 'please wait...'
-                runCmd = Popen4(cmd)
-                answer = str(runCmd.fromchild.read())
-                print answer
-                pattern ='(E|e)+rror'
-                if re.search(pattern, answer):
-                    Popen4('rm -f '+dest)
-                    self.dataset.copyStatus ='error'
-                    self.dataset.copyInfo   = answer
-            else:
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File already exists '
-                                                       
-        elif src == 'castor':
-            print 'castor file'
-            dest = os.path.join(destDir,
-                                os.path.basename(filename))
-            cmd = '/afs/.cern.ch/project/gd/LCG-share/3.0.0/globus/bin/'
-            #cmd  = '/usr/local/lcg/globus/bin/'
-            cmd += 'globus-url-copy gsiftp://castorgrid.cern.ch'
-            cmd +=  filename +' '+'file://'+dest
-    
-            if not os.path.exists(dest):
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File copied to \n'
-                self.dataset.copyInfo  += str(dest)
-                
-                print 'executing :'+cmd
-                print 'please wait...'
-                runCmd =Popen4(cmd)
-                answer = str(runCmd.fromchild.read())
-                print answer
-                pattern ='(E|e)+rror'
-                if re.search(pattern, answer):
-                    Popen4('rm -f '+dest)
-                    self.dataset.copyStatus ='error'
-                    self.dataset.copyInfo   = answer
-            else:
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File already exists '
-        else:
-            print 'File src unknown, jumping...!!'
-            self.dataset.copyStatus = 'error'
-            self.dataset.copyInfo   = 'please give full castor or afs path!'
-                                                
-
-
-class ListCastorFile(BaseAction):
-    
-    def __init__(self , dsObject):
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-        BaseAction.__init__(self, self.name)
-
-    def doIt(self):
-        filename = string.strip(self.name)
-        destDir  = self.destDir
-        
-        if not self.isCastorFile():
-            print 'Not a castor file!!'
-            self.dataset.copyStatus ='error'
-            self.dataset.copyInfo   ='Not a castor file!!'
-        else:
-            dest = os.path.join(destDir,
-                                os.path.basename(filename))
-            cmd  = '/usr/local/lcg/edg/bin/'
-            cmd += 'edg-gridftp-ls gsiftp://castorgrid.cern.ch'
-            cmd +=  filename
-            if not os.path.exists(dest):
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File copied to \n'
-                self.dataset.copyInfo  += str(dest)
-                print 'executing :'+cmd
-                print 'please wait...'
-                runCmd = Popen4(cmd)
-                answer = str(runCmd.fromchild.read())
-                answer = string.strip(answer)
-                print answer
-                if answer != filename :
-                    self.dataset.copyStatus = 'error'
-                    self.dataset.copyInfo   = answer                
-            else:
-                self.dataset.copyStatus = 'success'
-                self.dataset.copyInfo   = 'File already exists '
-
-
-"""
-DQ2 Classes
------------
-dsObject = dataset object created by dataset
-           maker
-name     = dataset name
-destDir  = dir where files are copied to.
-          The base destDir is set in a config file.
-          This is then 
-"""
-
-
-class CopyDQ2Files:
-    def __init__(self , dsObject):
-
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-
-    def filesString(self):
-        s = ''
-        for file in self.dataset.filesToCopy:
-            s += ' '+os.path.basename(file)
-
-        return s
-
-    def doCopy(self):
-        self.mkDestDir()
-        cmd  = 'dq2_get -s '+self.dataset.site+' -r -d '+self.destDir
-        cmd += ' '+self.name
-        cmd += self.filesString()
-        print 'executing :'+cmd
-        print 'please wait...'
-        runCmd = Popen4(cmd)
-        answer = runCmd.fromchild.readlines()
-        self.handleCmdOutput(answer)
-        #print str(self.dataset.__dict__)
-
-    def handleCmdOutput(self, output):
-         pattern ='ERROR|not found'
-         for line in output:
-             print line
-             if re.search(pattern, line):
-                 self.dataset.copyStatus = 'error'
-                 self.dataset.copyInfo   = line
-             elif re.search('Total:', line):
-                 total,failed = line.split('-')
-                 if string.strip(failed)[-1:]     == '0':
-                     if string.strip(total)[-1:]  != '0':
-                         self.dataset.copyStatus   = 'success'
-                         print 'sucess success,success, success,success,success,success'
-                 self.dataset.copyInf = line
-             else:
-                 self.dataset.copyInfo    = formatCollection(output)
-             
-
-    def mkDestDir(self):
-        loc = os.path.join(self.destDir, self.name)
-        if not os.path.exists(loc):
-            cmd = 'mkdir '+loc
-            runCmd = Popen4(cmd)
-            runCmd.wait()
-        self.destDir = loc
-        
-        print'Datasets dir: '+ self.destDir
-        
-
-            
-
-class ListDQ2Files:
-    def __init__(self , dsObject):
-
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-
-    def filesString(self):
-        s = ''
-        for file in self.dataset.filesToCopy:
-            s += ' '+os.path.basename(file)
-
-        return s
-
-    def doCopy(self):
-
-        cmd  = 'dq2_get -rc -d '+self.destDir
-        cmd += ' '+self.name
-        cmd += self.filesString()
-        print 'executing :'+cmd
-        print 'please wait...'
-        runCmd = Popen4(cmd)
-        answer = str(runCmd.fromchild.read())
-        answer = string.strip(answer)
-        print answer
-
-  
-
diff --git a/Tools/RunTimeTester/dataCopier/DSMConfigReader.py b/Tools/RunTimeTester/dataCopier/DSMConfigReader.py
deleted file mode 100755
index af75b5d9b454a27b9f83e079b82c18d9ce047434..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/DSMConfigReader.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-import string,logging,os
-from xml.dom.minidom import *
-from getText import getText
-
-logger = logging.getLogger('rtt')
-
-class DSMConfigReader:
-    def __init__(self, dsmcfg):
-        self.dsmcfg = dsmcfg
-        self.rttcfg = ''
-
-    def readDSMcfg(self):
-        logger.debug('Reading dsm configurations')
-        cfgDict = {}
-        dom     = parse(self.dsmcfg)
-        expectedTags = ['mode','datasetCatalog','rttsrc',
-                        'reportDir','rttcfg','datasetDir']
-        for t in expectedTags:
-            tag = dom.getElementsByTagName(t)
-            cfgDict[t]=string.strip(getText(tag[0].childNodes))
-
-        cfgDict = self.addRTTcfg(cfgDict['rttcfg'], cfgDict)
-        return cfgDict
-
-    def addRTTcfg(self, rttcfg, cfgDict):
-        dict = cfgDict
-        dom  = parse(rttcfg)
-        tags = ['release', 'releaseType'] #tags of interest from rtt cfg
-        for t in tags:
-            tag     = dom.getElementsByTagName(t)
-            dict[t] = string.strip(getText(tag[0].childNodes))
-
-        if dict['release'] == 'nightly':
-            lcr  ='/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasOffline/latest_copied_release'
-            dict['release'] = os.readlink(lcr)
-            logger.debug ('nightly Tag is '+str(os.readlink(lcr)))
-
-        return dict
-
-if __name__ == '__main__':
-    cr = DSMConfigReader('/home/en/RTT/dsmProject/dsmCfg.xml')
-    cfg = cr.readDSMcfg()
-
-    print cfg
-
-
-       
-        
diff --git a/Tools/RunTimeTester/dataCopier/DataSet.py b/Tools/RunTimeTester/dataCopier/DataSet.py
deleted file mode 100755
index 7e15cf46348b43b6824a0489e8e494eec097a6b1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/DataSet.py
+++ /dev/null
@@ -1,95 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#Dataset is a collection of information about datasets
-import os.path
-
-
-class DataSet:
-    def __init__(self, logicalName, destDir, pkgCfg):
-        self.logicalName  = logicalName
-        self.physicalName = ''
-        self.filesToCopy  = []
-        self.isInDestDir  = True
-        self.destDir      = destDir
-        self.packageCfg   = pkgCfg
-        self.dsSource     = ''
-        self.isToCopy     = False
-
-        self.copyStatus   = 'unknown'
-        self.copyInfo     = 'not set'
-
-        self.tobeCopied(self.logicalName, self.destDir)
-        self.setPhysicalName(self.logicalName, self.destDir)
-        self.setdsSource()
-        self.setIsToCopy()
-        
-    def tobeCopied(self, name, destDir):
-          bname = os.path.basename(name)
-          name = os.path.join(destDir, bname)
-          if not os.path.exists(name):
-              self.isInDestDir = False
-
-    def setPhysicalName(self, logicalName, destDir):
-        name = os.path.basename(logicalName)
-        self.physicalName = os.path.join(destDir, name)
-
-    def setdsSource(self):
-        if os.path.isabs(self.logicalName):
-            pieces = self.logicalName.split('/')
-            self.dsSource = pieces[1]  # eg : dsSource = castor
-                                       # if /castor/
-    def setIsToCopy(self):
-        if not os.path.exists(self.physicalName):
-            print 'file to be copied '+str(self.physicalName)
-            self.isToCopy     = True
-            self.filesToCopy.append(self.logicalName)
-            
-        
-        
-    
-class DataSetToBeCopied(DataSet):
-    def __init__(self, logicalName, destDir, pkgCfg):
-        DataSet.__init__(self, logicalName, destDir, pkgCfg)
-
-        #self.copyStatus =''
-        #self.copyInfo   ='Unknown'
-        self.copyStatusInit()
-        
-    def copyStatusInit(self):
-        self.copyStatus = 'unknown'
-        self.copyInfo   = 'No info yet'
-    
-
-class DQ2DataSet(DataSet):
-
-    def __init__(self, logicalName, destDir, pkgCfg, files, site ):
-        DataSet.__init__(self, logicalName, destDir, pkgCfg)
-        self.physicalFiles = []
-        self.files         = files
-        self.filesToCopy   = []
-        self.site          = site
-        self.setdsSource()
-        self.setPhysicalFiles()
-        self.setFilesToCopy()
-        self.setIsToCopy()
-        #print str(self.__dict__)
-        
-    def setdsSource(self):
-        self.dsSource = 'dq2'
-
-    def setPhysicalFiles(self):
-        for file in self.files:
-            fn = os.path.join(self.destDir, self.logicalName)
-            fn = os.path.join(fn, file)
-            self.physicalFiles.append(fn)
-
-    def setFilesToCopy(self):
-        for file in self.physicalFiles:
-            if not os.path.exists(file):
-                self.filesToCopy.append(file)
-
-    def setIsToCopy(self):
-        if self.filesToCopy:
-            self.isToCopy    = True
-            self.isInDestDir = False
-    
diff --git a/Tools/RunTimeTester/dataCopier/DataSetMaker.py b/Tools/RunTimeTester/dataCopier/DataSetMaker.py
deleted file mode 100755
index 4b06359f8d2ee6f8af2e97575e5a45dc7ec8ba88..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/DataSetMaker.py
+++ /dev/null
@@ -1,93 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#make dataset objects
-import string
-
-from xml.dom.minidom    import *
-from DataSet           import DataSet, DQ2DataSet
-#from myDataSet          import DataSet, DQ2DataSet
-from getText            import getText
-from formatCollection   import formatCollection
-
-
-class DataSetMaker:
-    def __init__(self, packageCfgFile, destDir, catalog=''):
-        self.packageCfg  = packageCfgFile
-        self.catalog     = catalog
-        self.destDir     = destDir
-
-    def makeDatasets(self):
-        datasets  = []
-        print '====================================='
-        print 'Config to parse '+str(self.packageCfg)
-        print '====================================='
-        try:
-            dom       = parse(self.packageCfg)
-        except Exception, e:
-            print '====================================='
-            print 'Error in parsing '+str(self.packageCfg)
-            print '====================================='
-            print str(e)
-            return ([None],[None])
-
-        nonDQ2 = self.makeNonDQ2data(dom)
-        dq2    = self.makeDQ2data(dom)
-
-        return(nonDQ2, dq2)
-
-        print 'DQ2 data'
-        for d in dq2: print formatCollection((d.__dict__))
-        print '-------------------------------------------------------------'
-        print 'Non DQ2 data'
-        for nd in nonDQ2: print formatCollection(nd.__dict__)
-        
-    def getDataset(self, tag):
-        datasetName = string.strip(getText(tag.childNodes))
-        return datasetName
-
-    def handleDQ2Tag(self, tag):
-        siteTag   = tag.getElementsByTagName('dq2site')
-        dq2site   = getText(siteTag[0].childNodes)
-        dsTag     = tag.getElementsByTagName('dq2dataset')
-        dq2dsName = getText(dsTag[0].childNodes)
-        
-        fileTags  = tag.getElementsByTagName('dq2file')
-        dq2files  = []
-        for t in fileTags:dq2files.append(getText(t.childNodes))
-        print str((dq2dsName, dq2files))
-        return (dq2dsName, dq2files, dq2site)
-
-    def makeNonDQ2data(self, dom):
-        datasets  = []
-        tags      = dom.getElementsByTagName("dataset")
-        tags1     = dom.getElementsByTagName("datasetName")
-        tags.extend(tags1)
-        
-        dsNames   = [self.getDataset(tag) for tag in tags]
-        for name in dsNames:
-            if not name in datasets: #remove double counting
-                datasets.append(name)
-
-        dsObjs = [DataSet(name, self.destDir, self.packageCfg)
-                     for name in datasets]
-        return dsObjs
-
-    def makeDQ2data(self, dom):
-
-        dq2Tags = dom.getElementsByTagName('dq2')
-        #dsNamesAndFiles is a list of tuples
-        dsNamesAndFiles = [self.handleDQ2Tag(tag) for tag in dq2Tags]
-
-        dsObjs =[DQ2DataSet(df[0],self.destDir, self.packageCfg, df[1], df[2] )
-                 for df in dsNamesAndFiles]
-
-        return  dsObjs
-        
-
-if __name__ =='__main__':
-
-    packageCfgFile = '/home/en/RTT/TriggerRelease_TestConfiguration.xml'
-    destDir        = '/home/en/RTT' 
-
-    dm = DataSetMaker(packageCfgFile,destDir )
-    dm.makeDatasets()
diff --git a/Tools/RunTimeTester/dataCopier/DataSetManager.py b/Tools/RunTimeTester/dataCopier/DataSetManager.py
deleted file mode 100755
index 8b0e41c9138ae86fecfaa369995f0deb445aac7b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/DataSetManager.py
+++ /dev/null
@@ -1,128 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os, os.path, logging
-
-from DSMConfigReader import DSMConfigReader
-from DataSetMaker    import DataSetMaker
-from Factory_Action  import ActionFactory,DQ2ActionFactory
-from DataSet         import DataSetToBeCopied
-from ReportWriter    import ReportWriter
-
-from validateXMLFile import validateXMLFile
-#================================================================
-def usage():
-  usg  = '\nUsage: '+sys.argv[0]+' <cfg_file> \n\n'
-  usg += '<cfg_file> is an xml config file \n'
-  usg += 'specifying configuration parameters needed to \n'
-  usg += 'initialise the script\n'
-  usg += 'See sample config file at: ...address!'
-  print usg
-  sys.exit(0)
-#=============================================================
-
-
-class DSMRunner:
-    def __init__(self, dsmcfg):
-        self.dsmcfg     = dsmcfg
-        self.confReader = DSMConfigReader(self.dsmcfg)
-        self.cfg        = self.confReader.readDSMcfg()
-        print str(self.cfg)
-        sys.path.append(self.cfg['rttsrc'])
-        print 'Sys.path ....'
-        print str(sys.path)
-
-    def getDataSets(self, files):
-       
-        validFiles = [file for file in files if os.path.exists(file)]
-        for file in validFiles:
-            #print'reading '+str(file)
-            dsmaker = DataSetMaker(file,self.cfg['datasetDir'])
-            ds      = dsmaker.makeDatasets()
-            print str(ds)
-            ds      = [d for d in ds if d != None]
-            
-            nonDQ2      = [d for d in ds[0] if d != None]
-            dq2         = [d for d in ds[1] if d != None]
-        
-        return (nonDQ2, dq2)
-        
-    def run(self):
-        from CatalogMaker     import CatalogMaker
-        from Factory_USRMaker import USRMakerFactory
-
-        from LoggerSetup      import setupLoggers        
-        setupLoggers(os.getcwd())
-        logger=logging.getLogger('rtt')
-        
-        logger.info('Making UserStuffRetriever')
-        usrMaker = USRMakerFactory().create(self.cfg['releaseType'])
-        usrMaker = usrMaker(self.cfg['rttcfg'])
-        usr      = usrMaker.makeUSR()
-        logger.info ('USR made succesfully, finding package configuration files ')
-        
-        packages    = usr.userStuff()
-        rttPackages = [p for p in packages if p.wantsRTT]
-        logger.debug( 'usr found '+str(len(packages))+' packages')
-        self.packageConfigs = [p.pathToTestConfig for p in rttPackages]
-        logger.info(str(self.packageConfigs))
-    
-        validateXMLFile('/home/en/RTT/TriggerRelease_TestConfiguration.xml')
-        print 'Config file has been validated against the dtd...'
-        #self.packageConfigs = ['/home/en/RTT/TriggerRelease_TestConfiguration.xml']
-        self.datasets,self.dq2 = self.getDataSets(self.packageConfigs)
-       
-
-        self.tobeCopied        = [ds for ds in  self.datasets
-                                  if ds.isToCopy]
-        self.dq2TobeCopied          = [ds for ds in self.dq2
-                                  if ds.isToCopy]
-       
-        
-        logger.debug('no of nonDQ2 datasets to be copied '+str(len(self.tobeCopied)))
-        logger.debug('no of DQ2 datasets to be copied '+str(len(self.dq2TobeCopied)))
-
-        af     = ActionFactory()
-        action = af.create(self.cfg['mode']) #action is to list/copy
-        for dataset in self.tobeCopied:      #a file on castor
-            dataset.__class__= DataSetToBeCopied
-            action(dataset).doIt()
-            
-        dqf     = DQ2ActionFactory()
-        action = dqf.create(self.cfg['mode']) #action is to list/copy
-        for data in self.dq2TobeCopied:
-          print ' copying of files using dq2'
-          action(data).doCopy()
-            
-        self.copyOK = [d for d in self.tobeCopied if d.copyStatus =='success']
-        logger.info(str(len(self.copyOK)) +' dataset(s) copied successfully')
-
-        self.dq2CopyOk =[d for d in self.dq2TobeCopied
-                         if d.copyStatus == 'success' ]
-
-        self.datasets.extend(self.dq2)
-        self.copyOK.extend(self.dq2CopyOk)
-        
-        logger.info( 'writing report ')
-        rw = ReportWriter(self.datasets, self.cfg, self.packageConfigs)
-        rw.write(self.cfg['reportDir'])
-        logger.info('report written to '+str(rw.report))
-
-        logger.info('Making new dataset catalog')
-        cm = CatalogMaker(self.cfg['datasetCatalog'],self.cfg['datasetCatalog'])
-        #cm.update(self.copyOK, self.datasets.values() )
-        cm.update(self.copyOK, self.datasets )
-        cm.makeCatalog()
-
-      
-
-if __name__ =='__main__':
-
-    if len(sys.argv) != 2:
-        usage()
-
-    dsmConfigFile = sys.argv[1]
-    
-    dr = DSMRunner(dsmConfigFile)
-    dr.run()
-
-
diff --git a/Tools/RunTimeTester/dataCopier/DiskRecordKeeper.py b/Tools/RunTimeTester/dataCopier/DiskRecordKeeper.py
deleted file mode 100755
index 1605314f6b674866f44a8c5ec235a3aa64a6dbe3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/DiskRecordKeeper.py
+++ /dev/null
@@ -1,75 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path, time, os, anydbm
-from popen2 import Popen4
-
-class DiskRecordKeeper:
-    def __init__(self, dDir):
-        self.dir      = dDir
-        self.aTimes   = {}
-        self.fileInfo = {}
-
-        self.getDatasetNames()
-        self.fillLastAccessTimesDict()
-        self.storeAccessTimes()
-
-    def getDatasetNames(self):
-        #returns a list of datasets on disk
-        ls_Dir = Popen4('ls '+self.dir)
-        data   = ls_Dir.fromchild.readlines()
-        self.datasets=[]
-        for ds in data:
-            ds = os.path.join(self.dir, ds)
-            self.datasets.append(ds[0:-1])
-        #return  self.datasets
-
-    def fillLastAccessTimesDict(self):
-        for d in self.datasets:
-            #self.aTimes[d]=time.ctime(os.path.getatime(d))
-            #print os.stat(d)
-            self.fileInfo[d] = os.stat(d)
-            #self.aTimes[d]   = (os.stat(d)[6], time.ctime(os.stat(d)[7]))
-            self.aTimes[d]   = (os.stat(d)[7])
-
-    def storeAccessTimes(self):
-        dbH = anydbm.open('FILEACCESSRECORD', 'n')
-        for k in self.aTimes.keys():
-            dbH[k]= str(self.aTimes[k])
-
-    
-    def sortedDictValues(self, adict):
-        items = adict.items()
-        for item in items:
-            backitems=[ [v[1],v[0]] for v in items]
-        backitems.sort()
-        #print str(items)
-        return backitems
-        #return [value for key, value in items]
-
-
-
-if __name__ =='__main__':
-    #drk = DiskRecordKeeper('/unix/atlas/RTTdev/eric')
-    drk = DiskRecordKeeper('/unix/atlas1/RTTdata')
-
-    items = drk.sortedDictValues(drk.aTimes)
-    #print str(a)
-
-    print 'Last Access Time               size             File Name'
-    print '================               ====             =========\n\n'
-    for item in items:
-        size      = str(os.stat(item[1])[6]/1000000000.0)
-        accesstime = str( time.ctime(item[0]))
-        filename  = str(item[1])
-        
-        print  accesstime +'     '+size+' GB        '+filename
-        #print str( time.ctime(item[0])) +'    '+ size+' GB          '+str(item[1])
-
-    #print '\n\n\n\n'
-    #print 'Last Access Time             File Name'
-    #print '================             =========\n\n'
-    #for k in drk.aTimes.keys():
-    #    #print  k
-    #    print str(drk.aTimes[k])+'    '+ k
-    #    print '\n'
-    
diff --git a/Tools/RunTimeTester/dataCopier/Factory_Action.py b/Tools/RunTimeTester/dataCopier/Factory_Action.py
deleted file mode 100755
index 13a0b4b3d77d376a438779d14dcf66466de532cb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/Factory_Action.py
+++ /dev/null
@@ -1,38 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from CopyAction import CopyCastorFile, ListCastorFile
-#from myCopyAction import CopyCastorFile, ListCastorFile
-from CopyAction   import CopyDQ2Files,ListDQ2Files
-
-class ActionFactory:
-    def __init__(self):
-        self.dict = {'active' : CopyCastorFile,
-                     'passive': ListCastorFile}
-
-    def create(self, mode):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if mode in self.dict.keys():
-            return self.dict[mode]
-        else:
-            print 'Unknown mode in DataSetManager cfg, exiting...'
-            sys.exit()
-
-class DQ2ActionFactory:
-
-    def __init__(self):
-        self.dict = {'active' : CopyDQ2Files,
-                     'passive': ListDQ2Files}
-
-    def create(self, mode):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if mode in self.dict.keys():
-            return self.dict[mode]
-        else:
-            print 'Unknown mode in DataSetManager cfg, exiting...'
-            sys.exit()
diff --git a/Tools/RunTimeTester/dataCopier/Factory_USRMaker.py b/Tools/RunTimeTester/dataCopier/Factory_USRMaker.py
deleted file mode 100755
index 387d31518f194f08f8d6727c591c797c4d0e632b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/Factory_USRMaker.py
+++ /dev/null
@@ -1,21 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from UserStuffRetrieverMaker import USRProjectMaker
-from UserStuffRetrieverMaker import USRMonolithicMaker
-
-
-class USRMakerFactory:
-    def __init__(self):
-        self.dict = {'project'   : USRProjectMaker,
-                     'monolithic': USRMonolithicMaker}
-
-    def create(self, releaseType):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if releaseType in self.dict.keys():
-            return self.dict[releaseType]
-        else:
-            print 'Unknown releaseType in RTT topLevelConfig cfg, exiting...'
-            sys.exit()
diff --git a/Tools/RunTimeTester/dataCopier/ReportWriter.py b/Tools/RunTimeTester/dataCopier/ReportWriter.py
deleted file mode 100755
index b69ffce2c282810fc97724b8a6b9ba636a8cbc76..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/ReportWriter.py
+++ /dev/null
@@ -1,105 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os,os.path, time, logging
-logger = logging.getLogger('rtt')
-
-class ReportWriter:
-    def __init__(self, datasets, configs, packageConfFiles):
-        self.datasets = datasets
-        self.configs  = configs
-        self.pcf      = packageConfFiles
-
-    def write(self, reportDir):
-        #print 'Datasets :'+str(self.datasets)
-        self.reportDir  = reportDir
-        
-        self.notToCopy  = [d for d in self.datasets if  d.isInDestDir]
-        self.tobeCopied = [d for d in self.datasets if not d.isInDestDir]
-        
-        logger.debug( 'to be copied :'+str(self.tobeCopied))
-        
-        self.copyFailed = [d for d in self.tobeCopied if d.copyStatus !='success']
-        logger.debug( 'copy Failed '+str(self.copyFailed))
-        
-        self.copyOK     = [d for d in self.tobeCopied if d.copyStatus =='success']
-        logger.debug( 'copyOK '+str(self.copyOK))
-
-        report      = 'DatasetManager.report'
-        self.report = os.path.join(self.reportDir, report)
-        
-        try:
-            print 'Working dir : '+str(os.getcwd())
-            #os.chdir(os.getcwd())
-            fh = open(self.report, 'w')
-            #fh.write('\n************************************************************')
-            fh.write('\nDataSetManager Report\t\t' + time.ctime())
-            fh.write('\n************************************************************\n\n')
-
-            #fh.write('\nDate \t\t\t\t\t: '+ time.ctime())
-
-            fh.write('\n\nRelease      \t\t\t\t: '+str(self.configs.get('release')))
-            fh.write('\nRelease Type \t\t\t\t: '+str(self.configs.get('releaseType')))
-
-            #fh.write('\n\n************************************************************')
-            fh.write('\n\nTotal number of packages found by RTT \t: '+str(len(self.pcf)))
-            fh.write('\n\nTotal number of datasets requested    \t: '+str(len(self.datasets)))
-
-            fh.write('\nNo of datasets to be copied:          \t: '+str(len(self.tobeCopied)))
-
-            fh.write('\nNo of datasets where copyOK:          \t: '+str(len(self.copyOK)))
-            fh.write('\nNo of datasets where copyFailed:      \t: '+str(len(self.copyFailed)))
-
-            #fh.write('\n\nDETAILS')
-            #fh.write('\n****************************************************************\n')
-            fh.write('\n\n\nDetails on Datasets where copy failed:')
-           # fh.write('\n****************************************************************')
-            for ds in self.copyFailed:
-                fh.write('\n====================================')
-                fh.write('\nName         :'+str(ds.logicalName))
-                fh.write('\nCopy Status  :'+str(ds.copyStatus))
-                fh.write('\nFailure Reason :\n')
-                fh.write(str(ds.copyInfo))
-                fh.write('\nPackageCfg : '+str(ds.packageCfg))
-                fh.write('\n====================================\n')    
-            #fh.write('\n****************************************************************\n\n')
-            fh.write('\n\nDetails on Datasets where copy succeeded:')
-            #fh.write('\n****************************************************************')
-            for ds in self.copyOK:
-                fh.write('\n====================================')
-                fh.write('\nName         :'+str(ds.logicalName))
-                fh.write('\nCopy Status  :'+str(ds.copyStatus))
-                fh.write('\nPhysicalName :'+str(ds.physicalName))
-                fh.write('\n====================================\n\n')    
-            fh.write('\n****************************************************************\n\n')
-
-            fh.write('\nRequested Datasets already at UCL')
-            fh.write('\n****************************************************************')
-            for ds in self.notToCopy:
-                fh.write('\n'+str(ds.logicalName))
-                #if ds.files:
-                if ds.dsSource =='dq2':
-                    for file in ds.files:
-                        fh.write('\n\t'+str(file))
-        
-            fh.close()
-
-        except Exception ,e :
-            logger.error('Could not create a report file '+str(self.report))
-            print str(e)
-class WriteTest:
-    def __init__(self):
-        pass
-
-    def write(self):
-        fh=open('testReport.results', 'w')
-        fh.write('\nThe test file ')
-
-
-if __name__ =='__main__':
-
-    rt = WriteTest()
-    rt.write()
-                           
-    
-                            
-    
diff --git a/Tools/RunTimeTester/dataCopier/UserStuffRetrieverMaker.py b/Tools/RunTimeTester/dataCopier/UserStuffRetrieverMaker.py
deleted file mode 100755
index f7636ed8837e676a8490fb9b8424dd3aa2420595..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/UserStuffRetrieverMaker.py
+++ /dev/null
@@ -1,95 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTConfigParser           import * 
-#from Paths                     import NoPackagePaths
-from RunTimer                  import RunTimer
-from formatCollection          import formatCollection
-from Paths                     import NoPackagePaths
-#from UserStuffRetrieverProject import UserStuffRetrieverProject
-#from UserStuffRetrieverProject import *
-#from UserStuffRetriever2       import UserStuffRetriever
-from Factory_UserStuffRetriever import UserStuffRetrieverFactory
-import sys, os
-    
-import logging
-logger = logging.getLogger('rtt')
-
-class USRBase:
-    def __init__(self, rttConfFile,
-                 subString = '_TestConfiguration'):
-        self.rttConfFile = rttConfFile
-        self.subString   = subString
-
-
-class USRProjectMaker:
-    def __init__(self, rttConfFile,
-                 subString = '_TestConfiguration'):
-       
-        self.rttConfFile = rttConfFile
-        self.subString   = subString
-    
-    def makeUSR(self):
-        logger.info( 'Making UserStuffRetriever Project ')
-        cfg = RTTConfiguration(self.rttConfFile)
-        confDict = cfg.config
-        #extend oringinal Config
-        confDict['originalBranch'] = cfg.originalBranch
-        confDict['findNightly']    = cfg.findNightly
-        confDict['valid']          = cfg.valid
-        confDict['isNightly']       = cfg.isNightly
-
-        if confDict['release'] == 'nightly':
-            lcr  ='/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasOffline/latest_copied_release'
-            confDict['release'] = os.readlink(lcr)
-            print 'nightly Tag is '+str(os.readlink(lcr))
-        
-        print formatCollection(confDict)
-        #print cfg.__str__()
-        
-        paths = NoPackagePaths('RunTimeTester-00-00-96',
-                                confDict, RunTimer(60*60*20 ,logger ),
-                                'outLog.log', logger)
-
-        #usr    = UserStuffRetrieverProject(paths,
-        #usr     = UserStuffRetrieverRelease(paths,
-        #                                   self.subString)
-        #usr    = UserStuffRetrieverFactory(logger).create(paths, self.subString, confDict)
-        usr    = UserStuffRetrieverFactory(logger).create(paths, self.subString)
-        return usr
-        
-
-class USRMonolithicMaker:
-    def __init__(self, rttConfFile,
-                 subString = '_TestConfiguration'):
-
-        self.rttConfFile = rttConfFile
-        self.subString   = subString
-
-    def makeUSR(self):
-        logger.info('Making UserStuffRetriever Monolithic')
-        confDict = rttConfiguration(self.rttConfFile)
-            
-        logger.debug( 'rttConfig :'+formatCollection(str(confDict))) 
-        paths    = NoPackagePaths('RunTimeTester-00-00-89',
-                           confDict, RunTimer(60*60*20 ,logger ),
-                                'outLog.log')
-        usr   = UserStuffRetriever(paths, self.subString, confDict )
-
-        return usr
-    
-
-if __name__ == '__main__':
-
-    from DSMConfigReader import DSMConfigReader
-
-    cr  = DSMConfigReader('/home/en/RTT/dsmProject/dsmCfg.xml')
-    cfg = cr.readDSMcfg()
-    print str(cfg)
-    sys.path.append(cfg['rttsrc'])
-    usrMaker = USRProjectMaker(cfg['rttcfg'])
-    #usrMaker = USRMonolithicMaker(cfg['rttcfg'])
-    
-    usr = usrMaker.makeUSR()
-
-    print str(usr)
-    
diff --git a/Tools/RunTimeTester/dataCopier/VersionedFileSaver.py b/Tools/RunTimeTester/dataCopier/VersionedFileSaver.py
deleted file mode 100755
index fb2705d432c74597d237b715c99fd0cef1c022c5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/VersionedFileSaver.py
+++ /dev/null
@@ -1,120 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#from RTTSException import RTTSException
-from types         import IntType
-
-import os, shutil, logging
-
-logger=logging.getLogger('rtt')
-
-def isInt(a):
-    try:
-        i = int(a)
-        return True
-    except:
-        return False
-
-def stripVersion(fn, sep):
-    (dir, file) = os.path.split(fn)
-    file = file.split(sep)[0]
-    return file
-    
-def ctimeCmp(a, b):
-    return cmp(os.path.getctime(a), os.path.getctime(b))
-        
-class VersionedFileSaver:
-    def __init__(self, maxVersion, destDir=None):
-        if not type(maxVersion) == IntType:
-            msg = 'VersionedFileSaver contructor arg not an int: '+maxVersion
-            #raise RTTSException(msg)
-            #raise Exception, e:
-            print msg
-        self.maxVersion = maxVersion;
-        if destDir:
-            self.destDir = destDir
-        else:
-            self.destDir = os.getcwd()
-
-    def getVersionedFiles(self, fn, sep):
-
-        vFiles = []
-        if not os.path.exists(fn): return vFiles
-
-        if not os.path.isfile(fn): return vFiles
-
-        # remove version if any from file name
-        noVerFile = stripVersion(fn, sep)
-
-        # look for versioned files: these have the form xxx;n
-        # where n is in integer
-        (dir, vfile)= os.path.split(noVerFile)
-        vfile += sep
-
-        vFiles = os.listdir(self.destDir)
-        vFiles = [f for f in vFiles if f.find(vfile) >=0]
-        vFiles = [f for f in vFiles if f.count(sep) == 1]
-
-
-        vFiles = [f for f in vFiles if isInt(f.split(sep)[1])]
-        vFiles = [os.path.join(self.destDir, f) for f in vFiles]
-
-        return vFiles
-
-    
-
-    def saveFile(self, fn):
-        sep = '-'
-        self.vFiles = self.getVersionedFiles(fn, sep)
-        nextVersion = -1
-
-        # find the last (in time) created file
-        self.vFiles.sort(ctimeCmp)
-
-        if self.vFiles:
-            lastFile = os.path.basename(self.vFiles[-1])
-            nextVersion = (int(lastFile.split(sep)[1])+1) % self.maxVersion
-        else:
-            nextVersion = 1
-
-
-        noVerFile = stripVersion(fn, sep)
-        toFile = noVerFile+sep+str(nextVersion)
-        if self.destDir:
-            toFile =  os.path.join(self.destDir, os.path.basename(toFile))
-
-
-        try:
-            shutil.copyfile(fn, toFile)
-        except (IOError, os.error), why:
-            logger.error("Cannot copy  %s to %s: %s" % (fn, toFile, str(why)))
-        
-if __name__ == '__main__':
-
-
-    fn = './temp'
-    file = open(fn, 'w')
-
-    file.write('junk\n')
-    file.close()
-
-    try:
-        vfs = VersionedFileSaver('7')
-    except:
-        print 'caught bad VSF construction'
-
-
-    destDir = os.environ['HOME']
-    # vfs = VersionedFileSaver(8,destDir)
-    vfs = VersionedFileSaver(8)
-    print vfs.destDir
-    nvers = 1
-    while nvers > 0:
-        nvers -= 1
-        vfs.saveFile(fn)
-            
-    for f in vfs.vFiles:
-        print f,os.path.getctime(f)
-        
-
-        
-        
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/DataCopierRunning.txt b/Tools/RunTimeTester/dataCopier/copier_cern/DataCopierRunning.txt
deleted file mode 100755
index 900fb6c94af0f3bc3f3b7c1f35d42547f394d363..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/DataCopierRunning.txt
+++ /dev/null
@@ -1,114 +0,0 @@
-Introduction
----------------------------------------------------------------------------------------------------
-The RTT data copier is a set of scripts which scan a given release for 
-configuration files from which it collects the datasets. It then compares 
-the datasets requested with those available in the catalog. If any new data 
-are found, it copies these to a local disk and updates the catalog. At the 
-end of the process, it writes a report to a specified directory.
-
-Features
-========
-
-The data copier uses two configuration files:
- - An RTT top level configuration file
- - A copier specific configuration file
-to know what actions it needs to perform.
-
-A sample RTT top level configuration file is available at
-http://www.hep.ucl.ac.uk/atlas/AtlasTesting/RTTUserGuide/RTTUserGuide.html
-Go to section on LocalRunning and install the setup files as indicated.
-Please read the RTT documentation at
-http://www.hep.ucl.ac.uk/atlas/AtlasTesting  
-if you need more information on this configuration file and the RTT
-
-A sample copier specific configuration file looks like:
-
-<?xml version="1.0"?>
-
-<dsmcfg>
-   <mode>active</mode>
-   <datasetCatalog>/afs/user/scratch0/Tools/RunTimeTester/dataCopier/copier_cern/cfg/rttDatasetCatalog.xml</datasetCatalog>
-   <rttsrc>/afs/cern.ch/user/e/enzuobon/scratch0/Tools/RunTimeTester/src</rttsrc>
-   <rttLibDir>/afs/cern.ch/atlas/project/RTT/prod/Tools/RunTimeTester/share</rttLibDir>
-   <logDir>/afs/cern.ch/user/e/enzuobon/scratch0</logDir>
-   <reportDir>/afs/cern.ch/user/e/enzuobon/scratch0</reportDir>
-   <rttcfg>/afs/cern.ch/atlas/project/RTT/prod/cfg/nightlyForCopier_i686-slc4-gcc34-opt.xml</rttcfg>
-   <datasetDir>/castor/cern.ch/user/e/enzuobon</datasetDir>
-   <stageArea>/afs/cern.ch/atlas/project/RTT/data/eric</stageArea>
-</dsmcfg>
-
-The mode tag in the copier config can take values: 
-'active' and 'passive'
-'passive' is meant to simulate the copying while 
-'active' will actually do the copying
-
-Running
-=======
-
-For the copier to copy data over the grid to a local disk,
-the user needs to have a valid grid proxy. If you do not have 
-a grid certificate, please look at the following sites which 
-contain grid registration information.
-http://www.hep.ucl.ac.uk/escience/gridintro.shtml
-http://lcg.web.cern.ch/LCG/users/registration/registration.html
-
-To verify if your grid proxy is alive, do
-
-grid-proxy-info
-
-If no proxy is available, do 
-
-grid-proxy-init
-
-then enter password as requested.
-
-If you have not checked out the RTT yet, check it out now
-
-cmt co Tools/RunTimeTester 
-
-will get you the head version
-
-Set up the RTT so that you are able to run jobs. It is advisable to try
-running some short jobs like HelloWorld to make sure RTT machinery is OK.
-
-Now prepare a copier configuration file similar to the one shown above.
-
-To launch the copier, cd to the directory where you downloaded the RTT 
-and do
-
-python Tools/RunTimeTester/dataCopier/copier_cern/src/DatasetManager.py  </my/copier/config/file>
-
-
-The production cron table entry to run on nightlies are:
- /afs/cern.ch/atlas/project/RTT/prod/launch/runDataCopier_val.sh > /afs/cern.ch/user/r/rtt/copier_val.log
- /afs/cern.ch/atlas/project/RTT/prod/launch/runDataCopier_bugfix.sh >/afs/cern.ch/user/r/rtt/copier_bugfix.log
- /afs/cern.ch/atlas/project/RTT/prod/launch/runDataCopier.sh >/afs/cern.ch/user/r/rtt/copier.log
-
-
-If all goes well, you see a screen similar to an RTT start-up screen.
-
-When the process terminates, go the <reportDir> and examine the report.
-
-Also verify new data were added to the catalog, if any were found.
-
-If data were copied from the grid, verify that there were no broken copies in 
-<datasetDir>/<dq2dataset>
-
-
-
-Some sources of problems
- - invalid or no grid proxy ==> cannot download grid files
- - insufficient destination disk space
-
---------------------------------------------------------------------------------------------------------
-Note:
-CERN: the copier copies files available on the grid through grid  
-commands. The destination is always castor.
-No copies are performed for requests for files already on castor: the  
-data copier only adds an entry to the catalogue with the same logical  
-and physical name.
-
-UCL: copies are performed for for files located on the grid, afs or  
-castor. The destination is a local UCL disk.
-
- 
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/README.txt b/Tools/RunTimeTester/dataCopier/copier_cern/README.txt
deleted file mode 100644
index e69de29bb2d1d6434b8b29ae775ad8c2e48c5391..0000000000000000000000000000000000000000
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/TopLevelJobGroup.xml b/Tools/RunTimeTester/dataCopier/copier_cern/cfg/TopLevelJobGroup.xml
deleted file mode 100755
index bf25c50b751ded115a35d95e60f0a5f2fdba0db5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/TopLevelJobGroup.xml
+++ /dev/null
@@ -1,47 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                            
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-
-            <keepFilePattern info="All ROOT macro output from your job goes in here">ROOToutput.log</keepFilePattern>
-
-            <keepFilePattern info="Exceptions raised by your Python script go in here">Python_Script_Output.log</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-               <vetoString>job killed</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-
-         <jobGroup name="AthenaATN" parent="Athena"/>
-
-         <jobGroup name="ErrorGroup" parent="Top" />
-
-
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/cfg_nightly.xml b/Tools/RunTimeTester/dataCopier/copier_cern/cfg/cfg_nightly.xml
deleted file mode 100755
index d143ba852cce92fb7d4dfb9f3f35c5f0406bb658..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/cfg_nightly.xml
+++ /dev/null
@@ -1,133 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-
-
-<rttconfig>
-
-   <mode>PBS</mode>
-
-   <release>nightly</release>
-
-   <refRelease>11.0.0</refRelease>
-   <cmtConfig>i686-slc3-gcc323-opt</cmtConfig>
-   <build>opt</build>
-   <branch>dev</branch>
-
-   <runType>
-
-      <releaseRun/>
-
-   </runType>
-
-   <workBasePath>/home/en/RttHead/Work</workBasePath>
-
-   <resultsBasePath>/home/en/RttHead/Results</resultsBasePath>
-
-   <site>CERN</site>
-
-   <dataSetCatalog>/home/en/dcopier_dev/cfg/rttDataSetCatalog.xml</dataSetCatalog>
-
-   <jobGroupConfig>/home/en/RTT/TopLevelJobGroup.xml</jobGroupConfig>
-
-   <refDB>
-
-      <file>/home/en/RTT/refFile.db</file>
-
-      <enable>1</enable>
-
-   </refDB>
-
-   <moniDB>
-
-      <file>/home/en/RTT/moni.db</file>
-
-      <enable>1</enable>
-
-   </moniDB>
-
-
-   <confFileSrc>
-         <installedConfFiles/>
-   </confFileSrc>
-
-
-<!--
-
-   <confFileSrc>
-
-      <rttProvides>
-
-        <packageConfFile>
-
-           <package>Simulation/Atlfast/TestAtlfast</package>
-
-           <pathToFile>/home/en/RttHead/Tools/RunTimeTester/RunTimeTester-00-00-94/src/../TestSuiteScratch/TestSuiteFixtures/Fixtures3/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-
-
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4</cmtpath>
-
-           <version>TestAtlfast-00-00-61</version>
-
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/InstallArea</installArea>
-
-
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/shareArea</shareArea>
-
-        </packageConfFile>
-
-
-
-
-
-        <packageConfFile>
-
-           <package>Tools/RunTimeTester</package>
-
-           <pathToFile>/home/en/RttHead/Tools/RunTimeTester/RunTimeTester-00-00-94/src/../TestSuiteScratch/TestSuiteFixtures/Fixtures3/xml/RunTimeTester_TestConfiguration.xml</pathToFile>
-
-
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4</cmtpath>
-
-           <version>RunTimeTester-00-00-82</version>
-
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/InstallArea</installArea>
-
-
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/shareArea</shareArea>
-
-        </packageConfFile>
-
-
-
-
-
-      </rttProvides>
-
-   </confFileSrc>
-
--->
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-
-<!--   <testMode>NoNewRunDirs</testMode> -->
-
-   <maxTime>
-
-      <hours>1</hours>
-
-   </maxTime>
-
-   <!-- <rttKitTest/> -->
-
-   <releaseType>project</releaseType>
-
-</rttconfig>
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/dsm_cfg_nightly.xml b/Tools/RunTimeTester/dataCopier/copier_cern/cfg/dsm_cfg_nightly.xml
deleted file mode 100755
index 64781ebbe28e7a6ae97b0b3e6cb4452990867973..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/cfg/dsm_cfg_nightly.xml
+++ /dev/null
@@ -1,18 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE dsmcfg SYSTEM "/home/en/RTT/dsmProject/DatasetManager.dtd">
-
-<dsmcfg>
-   <mode>active</mode>
-   <datasetCatalog>/home/en/ForDcopier/Tools/RunTimeTester/dataCopier/copier_cern/cfg/rttDataSetCatalog.xml</datasetCatalog>
-   <!--<rttsrc>/home/en/RTT/Grid/Tools/RunTimeTester/src</rttsrc>-->
-   <rttsrc>/home/en/ForDcopier/Tools/RunTimeTester/src</rttsrc>
-   <rttLibDir>/home/en/ForDcopier/Tools/RunTimeTester/src</rttLibDir>
-   <logDir>/home/en/ForDcopier</logDir>
-   <reportDir>/home/en/ForDcopier/Tools/RunTimeTester/dataCopier/copier_cern/report</reportDir>
-   <rttcfg>/home/en/ForDcopier/Tools/RunTimeTester/dataCopier/copier_cern/cfg/cfg_nightly.xml</rttcfg>
-   <datasetDir>/castor/home/en/dcopier_dev/data</datasetDir>
-   <stageArea>/afs/cern.ch/atlas/project/RTT/data</stageArea>
-</dsmcfg>
-
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/CatalogMaker.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/CatalogMaker.py
deleted file mode 100755
index 626ec59619a0f40e0b4b4a9666ba50c9b30566fa..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/CatalogMaker.py
+++ /dev/null
@@ -1,81 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#Make a new RTT dataset Catalog
-import os.path, time, logging
-from DataSetCatalog     import DataSetCatalog
-from VersionedFileSaver import VersionedFileSaver
-from formatCollection import formatCollection
-
-logger = logging.getLogger('rtt')
-
-class CatalogMaker:
-    def __init__(self, paths, newCatalog, logger):
-        self.oldcat  = paths.dataSetCatalog
-        self.newcat  = newCatalog 
-        self.catalog = DataSetCatalog(logger, paths).catalog
-
-    def makeCatalog(self):
-        logger.debug('versioning old catalog')
-        vfs =VersionedFileSaver(8, os.path.dirname(self.oldcat))
-        vfs.saveFile(self.oldcat)
-
-        logger.debug('making new catalog')
-        fh = open(self.newcat, 'w')
-        fh.write('<?xml version="1.0"?>'+'\n')
-        fh.write('\n<!-- ********************************************************\n')
-        fh.write('\n            RTT Dataset Catalog                              \n')
-        fh.write('\n   Generated by  : DataSetManager                            \n')
-        fh.write('\n   Time          : '+time.ctime()+'                          \n')
-        fh.write('\n**********************************************************-->\n')
-        fh.write('\n<dsns>'+'\n')
-  
-        cdict = self.catalog
-        #cdict = self.cleanUp(cdict)
-        print 'No of datasets in new catalog: '+str(len(cdict.keys()))
-        for key in cdict.keys():
-            fh.write(' <dsn>'+'\n')
-            fh.write('    <logical>' + key +'</logical>'+'\n')
-            fh.write('    <physical>'+ cdict[key] +'</physical>'+'\n')
-            fh.write(' </dsn>'+'\n')
-        fh.write('</dsns>'+'\n')
-        fh.close()
-
-    def update(self, copiedDatasets, allDatasets):
-        'datasets is a list dataset objects'
-        datasets = copiedDatasets
-        #datasets.extend(self.datasetNotInCatalog(allDatasets))
-        catalog  = self.catalog
-        for ds in datasets:
-            print 'new dataset added to catalog '+str(ds.logicalName)
-            print str(ds.isToCopy)
-            if ds.dsSource != 'dq2':
-                catalog[ds.logicalName] = ds.physicalName
-            else:
-                for file in ds.physicalFiles:
-                    lname = os.path.join(ds.logicalName, os.path.basename(file))
-                    catalog[lname] = file
-                
-    def datasetNotInCatalog(self, requestedDatasets):
-        nic = [ds for ds in requestedDatasets
-               if ds.logicalName not in self.catalog.keys()]
-        dsnic = [ds for ds in nic if ds.isInDestDir]
-
-        return dsnic
-
-    def cleanUp(self, dict):
-        cleanDict ={}
-
-        for key in dict.keys():
-            if os.path.exists(dict[key]):
-                cleanDict[key]= dict[key]
-                #print 'OK : File exits on disk '+dict[key]
-            else:
-                print 'Error: Non existent file in catalog '+dict[key]
-
-        return cleanDict
-
-if __name__ == '__main__':
-    #pass
-    cm = CatalogMaker('/home/en/RTT/rttDataSetCatalog.xml', 'newCatalog.xml')
-    cm.makeCatalog()
-        
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/CopyAction.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/CopyAction.py
deleted file mode 100755
index 50bad5391e478e352db906934f40c2125475027c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/CopyAction.py
+++ /dev/null
@@ -1,292 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string,os,re, sys,time
-from popen2           import Popen4
-from formatCollection import formatCollection
-
-class BaseAction:
-    def __init__(self, filename):
-        self.name = filename
-
-
-#----------------------------------------------------------------------
-
-class CopyNonDQ2File(BaseAction):
-    def __init__(self , dsObject):
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-        BaseAction.__init__(self, self.name)
-
-
-    def copyFromCastor(self):
-        print 'castor file...'
-        dest = os.path.join(self.destDir,
-                                os.path.basename(self.name))
-        cmd = 'globus-url-copy gsiftp://castorgrid.cern.ch'
-        cmd +=  self.name +' '+'file://'+dest
-    
-        if not os.path.exists(dest):
-            self.dataset.copyStatus = 'success'
-            self.dataset.copyInfo   = 'File copied to \n'
-            self.dataset.copyInfo  += str(dest)
-            
-            print 'executing :'+cmd
-            print 'please wait...'
-            runCmd =Popen4(cmd)
-            answer = str(runCmd.fromchild.read())
-            print answer
-            pattern ='(E|e)+rror|not found'
-            if re.search(pattern, answer):
-                Popen4('rm -f '+dest)
-                self.dataset.copyStatus ='error'
-                self.dataset.copyInfo   = answer
-        else:
-            self.dataset.copyStatus = 'success'
-            self.dataset.copyInfo   = 'File already exists '
-
-
-    def copyFromAFS(self):
-         print 'afs file ...'
-         dest = os.path.join(self.destDir,
-                                os.path.basename(self.name))
-
-         cmd = 'cp '+filename +' '+dest
-     
-         if not os.path.exists(self.name):
-             self.dataset.copyStatus = 'error'
-             self.dataset.copyInfo   = 'file does not exist :'+self.ename
-         elif not os.path.exists(dest):
-             self.dataset.copyStatus = 'success'
-             self.dataset.copyInfo   = 'File copied to \n'
-             self.dataset.copyInfo  += str(dest)
-                
-             print 'executing :'+cmd
-             print 'please wait...'
-             runCmd = Popen4(cmd)
-             answer = str(runCmd.fromchild.read())
-             print answer
-             pattern ='(E|e)+rror|not found'
-             if re.search(pattern, answer):
-                 Popen4('rm -f '+dest)
-                 self.dataset.copyStatus ='error'
-                 self.dataset.copyInfo   = answer
-         else:
-             self.dataset.copyStatus = 'success'
-             self.dataset.copyInfo   = 'File already exists '
-
-
-    def doIt(self):
-        dict = {'afs'   : self.copyFromAFS,
-                'castor': self.copyFromCastor}
-
-        if self.dataset.dsSource in dict.keys():
-            print str(self.dataset.dsSource)
-            print 'Action to execute!!'
-            dict[self.dataset.dsSource]()
-        else:
-            print "No action taken because source of file is unknown"
-            print "File "+self.name
-        
-        
-#------------------------------------------------------------------------
-
-class CopyNonDQ2FileCERN(CopyNonDQ2File):
-    def __init__(self, dsObject):
-        print 'initialise CopyNonDQ2FileCERN'
-        CopyNonDQ2File.__init__(self, dsObject)
-
-    def copyFromCastor(self):
-        'copy file from castor'
-        self.dataset.physicalName = self.dataset.logicalName    
-        self.dataset.copyStatus   = 'success'
-        self.dataset.copyInfo     = 'File already at cern, no copy needed '
-
-    def copyFromAFS(self):
-        print 'copy file from afs'
-        self.dataset.physicalName = self.dataset.logicalName
-        self.dataset.copyStatus   = 'success'
-        self.dataset.copyInfo     = 'File already at cern, no copy needed '
-
-
-#------------------------------------------------------------------------
-        
-    
-
-class CopyCastorFile(BaseAction):
-    def __init__(self , dsObject):
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-        BaseAction.__init__(self, self.name)
-
-    def doIt(self):
-        filename = self.name
-        destDir  = self.destDir
-        src      = self.dataset.dsSource
-        print 'src : %s '%src
-        if src == 'afs':
-            pass
-                                                       
-        elif src == 'castor':
-            pass
-
-#-----------------------------------------------------------------------
-
-class ListCastorFile(BaseAction):
-    
-    def __init__(self , dsObject):
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-        BaseAction.__init__(self, self.name)
-
-    def doIt(self):
-        filename = string.strip(self.name)
-        destDir  = self.destDir
-        #cmd  = '/usr/local/lcg/edg/bin/'
-        cmd += 'edg-gridftp-ls gsiftp://castorgrid.cern.ch'
-        cmd +=  filename
-        print 'executing :'+cmd
-        print 'please wait...'
-        runCmd = Popen4(cmd)
-        answer = str(runCmd.fromchild.read())
-        answer = string.strip(answer)
-        print answer
-        if answer != filename :
-            self.dataset.copyStatus = 'error'
-            self.dataset.copyInfo   = answer
-
-
-"""
-DQ2 Classes
------------
-dsObject = dataset object created by dataset
-           maker
-name     = dataset name
-destDir  = dir where files are copied to.
-          The base destDir is set in a config file.
-          This is then 
-"""
-#---------------------------------------------------------------------
-
-class CopyDQ2Files:
-    def __init__(self , dsObject):
-
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-
-    def filesString(self):
-        s = ''
-        for file in self.dataset.filesToCopy:
-            s += ' '+os.path.basename(file)
-
-        return s
-
-
-    def doCopy(self):
-        self.cpstatus = {}
-        stage,castor  = self.setAreas()
-        print 'stage  ' +stage
-        print 'castor ' +castor
-        files = [os.path.basename(file) for file in self.dataset.filesToCopy]
-        for file in files:
-            dq2Getcmd  = 'dq2_get -s '+self.dataset.site
-            dq2Getcmd += ' -rc -d '+stage
-            dq2Getcmd += ' '+self.name
-            dq2Getcmd += ' '+file
-
-            temp = os.path.join(stage,  file)
-            dest = os.path.join(castor, file)
-            
-            try:
-                cmd    = Popen4('rfdir '+dest)
-                result = str(cmd.fromchild.read())
-                if re.search('No such file', result):
-                    print dq2Getcmd
-                    fromGrid   = Popen4(dq2Getcmd)
-                    answer = fromGrid.fromchild.readlines()
-                    self.handleStatus(self.name, answer, self.cpstatus)
-                    #fromGrid.wait()
-
-                    toCastor = Popen4('rfcp '+temp+ ' '+dest)
-                    toCastor.wait()
-                    
-                else:
-                    print 'File already available on castor'
-
-                print 'src file  '+temp
-                print 'dest file '+dest
-                time.sleep(10)
-                rmtemp = Popen4('rm -f '+temp)
-                rmtemp.wait()
-                print 'deleted '+temp
-            except Exception, e:
-                print 'problem getting data from grid %s'%str(e)
-
-        if 'error' in self.cpstatus.values():
-            self.dataset.copyStatus = 'error'
-            self.dataset.copyInfo   = str(formatCollection(self.cpstatus))
-        else:
-             self.dataset.copyStatus = 'success'
-
-    def handleStatus(self, name, result, dict):
-        pattern ='ERROR|not found'
-        for line in result:
-             if re.search(pattern, line):
-                 dict[name] = 'error'
-                 
-             elif re.search('Total:', line):
-                 total,failed = line.split('-')
-                 if string.strip(failed)[-1:]     == '0':
-                     if string.strip(total)[-1:]  != '0':
-                         dict[name] = 'success'
-
-    
-
-   
-    #------------------------------------------------------------------------
-    
-    def setAreas(self):
-        if self.dataset.destDir.startswith('/castor'):
-            self.stageArea = os.path.join(self.dataset.stageArea,
-                                          self.name)
-            print 'stage Area :'+self.stageArea
-            self.castorArea = os.path.join(self.dataset.destDir,
-                                           self.name)
-            print 'castor area '+self.castorArea
-            Popen4('mkdir '+self.stageArea)
-            Popen4('rfmkdir '+self.castorArea)
-            return (self.stageArea, self.castorArea)
-        
-#-----------------------------------------------------------------------------            
-
-class ListDQ2Files:
-    def __init__(self , dsObject):
-
-        self.dataset = dsObject
-        self.name    = self.dataset.logicalName
-        self.destDir = self.dataset.destDir
-
-    def filesString(self):
-        s = ''
-        for file in self.dataset.filesToCopy:
-            s += ' '+os.path.basename(file)
-
-        return s
-
-    def doCopy(self):
-
-        cmd  = 'dq2_ls '
-        cmd += self.name
-        cmd += self.filesString()
-        print 'executing :'+cmd
-        print 'please wait...'
-        runCmd = Popen4(cmd)
-        answer = str(runCmd.fromchild.read())
-        answer = string.strip(answer)
-        print answer
-
-  #------------------------------------------------------------------------
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMConfigReader.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMConfigReader.py
deleted file mode 100755
index b1c9416620622ec7d6de81891760b6b958bb3f63..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMConfigReader.py
+++ /dev/null
@@ -1,52 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-import string,logging,os
-from xml.dom.minidom import *
-from getText import getText
-
-logger = logging.getLogger('rtt')
-
-class DSMConfigReader:
-    def __init__(self, dsmcfg):
-        self.dsmcfg = dsmcfg
-        self.rttcfg = ''
-
-    def readDSMcfg(self):
-        logger.debug('Reading dsm configurations')
-        cfgDict = {}
-        dom     = parse(self.dsmcfg)
-        expectedTags = ['mode','datasetCatalog','rttsrc',
-                        'reportDir','rttcfg','datasetDir',
-                        'stageArea','rttLibDir', 'logDir']
-        for t in expectedTags:
-            tag = dom.getElementsByTagName(t)
-            cfgDict[t]=string.strip(getText(tag[0].childNodes))
-
-        cfgDict = self.addRTTcfg(cfgDict['rttcfg'], cfgDict)
-        return cfgDict
-
-    def addRTTcfg(self, rttcfg, cfgDict):
-        dict = cfgDict
-        dom  = parse(rttcfg)
-        tags = ['release', 'releaseType', 'site'] #tags of interest from rtt cfg
-        for t in tags:
-            tag     = dom.getElementsByTagName(t)
-            dict[t] = string.strip(getText(tag[0].childNodes))
-
-        if dict['release'] == 'nightly':
-            lcr  ='/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasProduction/latest_copied_release'
-            dict['release'] = os.readlink(lcr)
-            logger.debug ('nightly Tag is '+str(os.readlink(lcr)))
-
-        return dict
-
-if __name__ == '__main__':
-    cr = DSMConfigReader('/home/en/RTT/dsmProject/dsmCfg.xml')
-    cfg = cr.readDSMcfg()
-
-    print cfg
-
-
-       
-        
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMRunner.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMRunner.py
deleted file mode 100755
index 6b0976025ffcbbdf1fcb32d9f50731df613d0f51..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DSMRunner.py
+++ /dev/null
@@ -1,96 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os, os.path, logging
-
-from DSMConfigReader import DSMConfigReader
-from DataSetMaker    import DataSetMaker
-from Factory_Action  import ActionFactory
-from DataSet         import DataSetToBeCopied
-from ReportWriter    import ReportWriter
-
-
-#================================================================
-def usage():
-  usg  = '\nUsage: '+sys.argv[0]+' <cfg_file> \n\n'
-  usg += '<cfg_file> is an xml config file \n'
-  usg += 'specifying configuration parameters needed to \n'
-  usg += 'initialise the script\n'
-  usg += 'See sample config file at: ...address!'
-  print usg
-  sys.exit(0)
-#=============================================================
-
-
-class DSMRunner:
-    def __init__(self, dsmcfg):
-        self.dsmcfg     = dsmcfg
-        self.confReader = DSMConfigReader(self.dsmcfg)
-        self.cfg        = self.confReader.readDSMcfg()
-        
-        sys.path.append(self.cfg['rttsrc'])
-        
-
-    def getDataSets(self, files):
-        datasets= {}   # datasets[logicalName]=datasetObject 
-        for file in files:
-            #print'reading '+str(file)
-            dsmaker = DataSetMaker(file,self.cfg['datasetDir'])
-            ds      = dsmaker.makeDatasets()
-            for d in ds:
-                datasets[d.logicalName]= d 
-        return datasets
-        
-    def run(self):
-        from CatalogMaker     import CatalogMaker
-        from Factory_USRMaker import USRMakerFactory
-
-        from LoggerSetup      import setupLoggers        
-        setupLoggers(os.getcwd())
-        logger=logging.getLogger('rtt')
-        
-        logger.info( 'USR project maker imported')
-        usrMaker = USRMakerFactory().create(self.cfg['releaseType'])
-        usrMaker = usrMaker(self.cfg['rttcfg'])
-        usr      = usrMaker.makeUSR()
-        logger.info ('finding package configuration files ')
-        packages = usr.userStuff()
-        
-        logger.debug( 'usr found '+str(len(packages))+' packages')
-
-        self.packageConfigs = [p.pathToTestConfig for p in packages]
-        self.datasets       = self.getDataSets(self.packageConfigs)
-        self.tobeCopied     = [val for val in self.datasets.values()
-                               if not val.isInDestDir]
-        
-        logger.debug('no of datasets to be copied '+str(len(self.tobeCopied)))
-
-        af     = ActionFactory()
-        action = af.create(self.cfg['mode']) #action is to list/copy
-        for dataset in self.tobeCopied:      #a file on castor
-            dataset.__class__= DataSetToBeCopied
-            action(dataset).doIt()
-            
-        self.copyOK = [d for d in self.tobeCopied if d.copyStatus =='success']
-        
-        logger.info( 'write report ')
-        rw = ReportWriter(self.datasets, self.cfg, self.packageConfigs)
-        rw.write()
-
-        logger.info('Making new dataset catalog')
-        cm = CatalogMaker(self.cfg['datasetCatalog'],self.cfg['datasetCatalog'])
-        cm.update(self.copyOK)
-        cm.makeCatalog()
-
-
-
-if __name__ =='__main__':
-
-    if len(sys.argv) != 2:
-        usage()
-
-    dsmConfigFile = sys.argv[1]
-    
-    dr = DSMRunner(dsmConfigFile)
-    dr.run()
-
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSet.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSet.py
deleted file mode 100755
index 4d08a29d29aa55bedc84e9c7e6d0148188ad021a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSet.py
+++ /dev/null
@@ -1,153 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#Dataset is a collection of information about datasets
-import os.path
-
-
-class DataSet:
-    def __init__(self, logicalName, destDir, pkgCfg, catalog, stageArea=''):
-        self.logicalName  = logicalName
-        self.physicalName = ''
-        self.filesToCopy  = []
-        self.isInDestDir  = True
-        self.destDir      = destDir
-        self.packageCfg   = pkgCfg
-        self.catalog      = catalog 
-        self.dsSource     = 'unknown'
-        self.isToCopy     = False
-        self.stageArea    = stageArea
-        self.copyStatus   = 'unknown'
-        self.copyInfo     = 'not set'
-
-        self.setPhysicalName(self.logicalName, self.destDir)
-        self.setdsSource()
-        self.setIsToCopy()
-
-    def setPhysicalName(self, logicalName, destDir):
-        if self.logicalName in self.catalog.keys():
-            self.physicalName = self.catalog[self.logicalName]
-            return ''
-        name = os.path.basename(logicalName)
-        self.physicalName = os.path.join(destDir, name)
-
-    def setdsSource(self):
-        dict = {'/afs' : 'afs',
-                '/castor': 'castor'}
-        
-        for key in dict.keys():
-            if self.logicalName.startswith(key):
-                self.dsSource = dict[key]
-
-    def setIsToCopy(self):
-        if not self.logicalName in self.catalog.keys():
-            self.isToCopy    = True
-            self.isInDestDir = False
-
-
-class DataSetCERN(DataSet):
-    def __init__(self, logicalName, destDir, pkgCfg, catalog, stageArea=''):
-        DataSet.__init__(self, logicalName, destDir, pkgCfg, catalog, stageArea)
-        self.setPhysicalName(logicalName)
-        self.setIsToCopy()
-        
-    def setPhysicalName(self, logicalName):
-        self.physicalName = logicalName
-
-    def setIsToCopy(self):
-        if not os.path.isabs(self.physicalName):
-            print 'file to be copied '+str(self.physicalName)
-            self.isToCopy     = True
-            self.filesToCopy.append(self.logicalName)
-    
-class DataSetToBeCopied(DataSet):
-    def __init__(self, logicalName, destDir, pkgCfg, catalog, stageArea=''):
-        DataSet.__init__(self,
-                         logicalName,
-                         destDir,
-                         pkgCfg,
-                         catalog,
-                         stageArea)
-
-        #self.copyStatus =''
-        #self.copyInfo   ='Unknown'
-        self.copyStatusInit()
-        
-    def copyStatusInit(self):
-        self.copyStatus = 'unknown'
-        self.copyInfo   = 'No info yet'
-    
-
-class DQ2DataSet(DataSet):
-
-    def __init__(self, logicalName, destDir, pkgCfg, files, site, catalog, stageArea=''):
-        DataSet.__init__(self, logicalName, destDir, pkgCfg, catalog, stageArea)
-        self.physicalFiles = []
-        self.files         = files
-        self.filesToCopy   = []
-        self.site          = site
-        self.setLogicalFiles()
-        self.setdsSource()
-        #self.setPhysicalFiles()
-        self.setFilesToCopy()
-        self.setIsToCopy()
-        #print str(self.__dict__)
-        
-    def setdsSource(self):
-        self.dsSource = 'dq2'
-
-    def setPhysicalFiles(self):
-        self.physicalFiles=[os.path.join(self.destDir, file)
-                            for file in self.logicalFiles]
-
-    def setFilesToCopy(self):
-        copyCandidates   = [file for file in self.logicalFiles
-                            if not file in self.catalog.keys()]
-        copyCandidates   = [os.path.join(self.destDir, file)
-                            for file in copyCandidates]
-        self.filesToCopy = [file for file in copyCandidates
-                            if not os.path.exists(file)]
-        self.physicalFiles = self.filesToCopy                    
-        
-    def setIsToCopy(self):
-        if len(self.filesToCopy) > 0:
-            self.isToCopy    = True
-            self.isInDestDir = False
-            
-    def setLogicalFiles(self):
-        self.logicalFiles = [os.path.join(self.logicalName, file)
-                             for file in self.files]
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-"""
-    def setPhysicalFiles(self):
-        for file in self.files:
-            fn = os.path.join(self.destDir, self.logicalName)
-            fn = os.path.join(fn, file)
-            self.physicalFiles.append(fn)
-
-    def setFilesToCopy(self):
-        for file in self.physicalFiles:
-            if not os.path.exists(file):
-                self.filesToCopy.append(file)
-
-    def setIsToCopy(self):
-        if self.filesToCopy:
-            self.isToCopy    = True
-            self.isInDestDir = False
-    
-"""
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetMaker.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetMaker.py
deleted file mode 100755
index e90398a6bebbe42751786f473434c07c205c01cc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetMaker.py
+++ /dev/null
@@ -1,109 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#make dataset objects
-import string
-
-from xml.dom.minidom    import *
-from DataSet           import DataSet, DQ2DataSet
-#from myDataSet          import DataSet, DQ2DataSet
-from getText            import getText
-from formatCollection   import formatCollection
-
-
-class DataSetMaker:
-    def __init__(self, packageCfgFile, destDir, catalog='', stageArea=''):
-        self.packageCfg  = packageCfgFile
-        self.catalog     = catalog
-        self.destDir     = destDir
-        self.stageArea   = stageArea
-
-
-    def makeDatasets(self):
-        datasets  = []
-        print '====================================='
-        print 'Config to parse '+str(self.packageCfg)
-        print '====================================='
-        try:
-            dom       = parse(self.packageCfg)
-        except Exception, e:
-            print '====================================='
-            print 'Error in parsing '+str(self.packageCfg)
-            print '====================================='
-            print str(e)
-            return ([None],[None])
-
-        nonDQ2 = self.makeNonDQ2data(dom)
-        dq2    = self.makeDQ2data(dom)
-
-        return(nonDQ2, dq2)
-
-        #print 'DQ2 data'
-        #for d in dq2: print formatCollection((d.__dict__))
-        #print '-------------------------------------------------------------'
-        #print 'Non DQ2 data'
-        #for nd in nonDQ2: print formatCollection(nd.__dict__)
-        
-    def getDataset(self, tag):
-        datasetName = string.strip(getText(tag.childNodes))
-        return datasetName
-
-    def handleDQ2Tag(self, tag):
-        siteTag   = tag.getElementsByTagName('dq2site')
-        dq2site   = getText(siteTag[0].childNodes)
-        dsTag     = tag.getElementsByTagName('dq2dataset')
-        dq2dsName = getText(dsTag[0].childNodes)
-        
-        fileTags  = tag.getElementsByTagName('dq2file')
-        dq2files  = []
-        for t in fileTags:dq2files.append(getText(t.childNodes))
-        #print str((dq2dsName, dq2files))
-        return (dq2dsName, dq2files, dq2site)
-
-    def makeNonDQ2data(self, dom):
-        datasets  = []
-        tags      = dom.getElementsByTagName("dataset")
-        tags1     = dom.getElementsByTagName("datasetName")
-        tags.extend(tags1)
-        
-        dsNames   = [self.getDataset(tag) for tag in tags]
-        #print'non dq2 data'
-        #print str(len(dsNames))
-        for name in dsNames:
-            print name
-            if not name in datasets: #remove double counting
-                datasets.append(name)
-
-        dsObjs = [DataSet(name,
-                          self.destDir,
-                          self.packageCfg,
-                          self.catalog,
-                          self.stageArea)
-                     for name in datasets]
-
-        return dsObjs
-
-    def makeDQ2data(self, dom):
-
-        dq2Tags = dom.getElementsByTagName('dq2')
-        #dsNamesAndFiles is a list of tuples
-        dsNamesAndFiles = [self.handleDQ2Tag(tag) for tag in dq2Tags]
-
-        dsObjs =[DQ2DataSet(df[0],
-                            self.destDir,
-                            self.packageCfg,
-                            df[1],
-                            df[2],
-                            self.catalog,
-                            self.stageArea )
-                 for df in dsNamesAndFiles]
-        
-        return  dsObjs
-        
-
-if __name__ =='__main__':
-
-    packageCfgFile = '/home/en/RTT/TriggerRelease_TestConfiguration.xml'
-    destDir        = '/home/en/RTT' 
-
-    dm = DataSetMaker(packageCfgFile,destDir )
-    dm.makeDatasets()
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetManager.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetManager.py
deleted file mode 100755
index aec2239e9731950fd70b47a0ac0b4a751472f608..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DataSetManager.py
+++ /dev/null
@@ -1,144 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os, os.path, logging
-
-from DSMConfigReader import DSMConfigReader
-#from DataSetMaker    import DataSetMaker
-#from Factory_Action  import ActionFactory,DQ2ActionFactory
-#from DataSet         import DataSetToBeCopied
-#from ReportWriter    import ReportWriter
-
-#from validateXMLFile import validateXMLFile
-#================================================================
-def usage():
-  usg  = '\nUsage: '+sys.argv[0]+' <cfg_file> \n\n'
-  usg += '<cfg_file> is an xml config file \n'
-  usg += 'specifying configuration parameters needed to \n'
-  usg += 'initialise the script\n'
-  usg += 'See sample config file at: ...address!'
-  print usg
-  sys.exit(0)
-#=============================================================
-
-
-class DSMRunner:
-    def __init__(self, dsmcfg):
-        self.dsmcfg     = dsmcfg
-        self.confReader = DSMConfigReader(self.dsmcfg)
-        self.cfg        = self.confReader.readDSMcfg()
-        print str(self.cfg)
-        sys.path.append(self.cfg['rttsrc'])
-        print 'Sys.path ....'
-        print str(sys.path)
-
-    def getDataSets(self, files, catalog):
-        from DataSetMaker    import DataSetMaker
-      
-        nonDQ2data = []
-        dq2data    = []
-        validFiles = [file for file in files if os.path.exists(file)]
-        for file in validFiles:
-            #print'reading '+str(file)
-            dsmaker = DataSetMaker(file,
-                                   self.cfg['datasetDir'],
-                                   catalog,
-                                   self.cfg['stageArea'] )
-            ds      = dsmaker.makeDatasets()
-            ds      = [d for d in ds if d != None]
-            
-            nonDQ2      = [d for d in ds[0] if d != None]
-            dq2         = [d for d in ds[1] if d != None]
-        
-            nonDQ2data.extend(nonDQ2)
-            dq2data.extend(dq2)
-            #sys.exit(0)
-        return (nonDQ2data, dq2data)
-        
-    def run(self):
-        from DataSetMaker    import DataSetMaker
-        from Factory_Action  import ActionFactory,DQ2ActionFactory
-        from DataSet         import DataSetToBeCopied
-        from ReportWriter    import ReportWriter
-
-        from validateXMLFile import validateXMLFile
-
-      
-        from CatalogMaker     import CatalogMaker
-        from Factory_USRMaker import USRMakerFactory
-
-        from LoggerSetup      import setupLoggers        
-        setupLoggers(os.getcwd())
-        logger=logging.getLogger('rtt')
-        
-        logger.info('Making UserStuffRetriever')
-        usrMaker = USRMakerFactory().create(self.cfg['releaseType'])
-        usrMaker = usrMaker(self.cfg['rttcfg'], logger, self.cfg)
-        usr      = usrMaker.makeUSR()
-        
-        logger.info ('USR made succesfully, finding package configuration files ')  
-        packages            = usr.userStuff()
-        rttPackages         = [p for p in packages if p.wantsRTT]
-        logger.debug( 'usr found '+str(len(packages))+' packages')
-        self.packageConfigs = [p.pathToTestConfig for p in rttPackages]
-        logger.info(str(self.packageConfigs))
-        
-        #self.packageConfigs=['/home/en/testArea/toAddToCatalog.xml']
-
-        self.cm                = CatalogMaker(#self.cfg['datasetCatalog'],
-                                              usrMaker.paths,
-                                              self.cfg['datasetCatalog'],
-                                              logger)
-        self.catalog           = self.cm.catalog
-        self.datasets,self.dq2 = self.getDataSets(self.packageConfigs, self.catalog)  
-        self.tobeCopied        = [ds for ds in  self.datasets if ds.isToCopy]
-        self.dq2TobeCopied     = [ds for ds in self.dq2 if ds.isToCopy]
-       
-        logger.debug( 'No of dq2 datasets to be copied: '+str(len(self.dq2TobeCopied)))
-        
-        logger.debug('no of nonDQ2 datasets to be copied '+str(len(self.tobeCopied)))
-        logger.debug('no of DQ2 datasets to be copied '+str(len(self.dq2TobeCopied)))
-
-        af     = ActionFactory()
-        action = af.create(self.cfg['mode'], self.cfg['site']) #action is to list/copy
-        for dataset in self.tobeCopied:                        #a file on castor
-            dataset.__class__= DataSetToBeCopied
-            action(dataset).doIt()
-            
-        dqf     = DQ2ActionFactory()
-        action  = dqf.create(self.cfg['mode']) #action is to list/copy
-        for data in self.dq2TobeCopied:
-          print ' copying of files using dq2'
-          action(data).doCopy()
-            
-        self.copyOK = [d for d in self.tobeCopied if d.copyStatus =='success']
-        logger.info(str(len(self.copyOK)) +' dataset(s) copied successfully')
-
-        self.dq2CopyOk =[d for d in self.dq2TobeCopied
-                         if d.copyStatus == 'success' ]
-
-        self.datasets.extend(self.dq2)
-        self.copyOK.extend(self.dq2CopyOk)
-        
-        logger.info( 'writing report ')
-        rw = ReportWriter(self.datasets, self.cfg, self.packageConfigs)
-        rw.write(self.cfg['reportDir'])
-        logger.info('report written to '+str(rw.report))
-
-        logger.info('Making new dataset catalog')
-        self.cm.update(self.copyOK, self.datasets )
-        self.cm.makeCatalog()
-
-      
-
-if __name__ =='__main__':
-
-    if len(sys.argv) != 2:
-        usage()
-
-    dsmConfigFile = sys.argv[1]
-    #dsmConfigFile ='../cfg/dsm_cfg_nightly.xml'
-    
-    dr = DSMRunner(dsmConfigFile)
-    dr.run()
-
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/DiskRecordKeeper.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/DiskRecordKeeper.py
deleted file mode 100755
index 1605314f6b674866f44a8c5ec235a3aa64a6dbe3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/DiskRecordKeeper.py
+++ /dev/null
@@ -1,75 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path, time, os, anydbm
-from popen2 import Popen4
-
-class DiskRecordKeeper:
-    def __init__(self, dDir):
-        self.dir      = dDir
-        self.aTimes   = {}
-        self.fileInfo = {}
-
-        self.getDatasetNames()
-        self.fillLastAccessTimesDict()
-        self.storeAccessTimes()
-
-    def getDatasetNames(self):
-        #returns a list of datasets on disk
-        ls_Dir = Popen4('ls '+self.dir)
-        data   = ls_Dir.fromchild.readlines()
-        self.datasets=[]
-        for ds in data:
-            ds = os.path.join(self.dir, ds)
-            self.datasets.append(ds[0:-1])
-        #return  self.datasets
-
-    def fillLastAccessTimesDict(self):
-        for d in self.datasets:
-            #self.aTimes[d]=time.ctime(os.path.getatime(d))
-            #print os.stat(d)
-            self.fileInfo[d] = os.stat(d)
-            #self.aTimes[d]   = (os.stat(d)[6], time.ctime(os.stat(d)[7]))
-            self.aTimes[d]   = (os.stat(d)[7])
-
-    def storeAccessTimes(self):
-        dbH = anydbm.open('FILEACCESSRECORD', 'n')
-        for k in self.aTimes.keys():
-            dbH[k]= str(self.aTimes[k])
-
-    
-    def sortedDictValues(self, adict):
-        items = adict.items()
-        for item in items:
-            backitems=[ [v[1],v[0]] for v in items]
-        backitems.sort()
-        #print str(items)
-        return backitems
-        #return [value for key, value in items]
-
-
-
-if __name__ =='__main__':
-    #drk = DiskRecordKeeper('/unix/atlas/RTTdev/eric')
-    drk = DiskRecordKeeper('/unix/atlas1/RTTdata')
-
-    items = drk.sortedDictValues(drk.aTimes)
-    #print str(a)
-
-    print 'Last Access Time               size             File Name'
-    print '================               ====             =========\n\n'
-    for item in items:
-        size      = str(os.stat(item[1])[6]/1000000000.0)
-        accesstime = str( time.ctime(item[0]))
-        filename  = str(item[1])
-        
-        print  accesstime +'     '+size+' GB        '+filename
-        #print str( time.ctime(item[0])) +'    '+ size+' GB          '+str(item[1])
-
-    #print '\n\n\n\n'
-    #print 'Last Access Time             File Name'
-    #print '================             =========\n\n'
-    #for k in drk.aTimes.keys():
-    #    #print  k
-    #    print str(drk.aTimes[k])+'    '+ k
-    #    print '\n'
-    
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_Action.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_Action.py
deleted file mode 100755
index 6777b1a68c9a40b889a766dc22c44cec5224cb05..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_Action.py
+++ /dev/null
@@ -1,42 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from CopyAction  import CopyCastorFile, ListCastorFile
-from CopyAction  import CopyNonDQ2File, CopyNonDQ2FileCERN
-from CopyAction  import CopyDQ2Files,ListDQ2Files
-import sys
-
-class ActionFactory:
-    def __init__(self):
-        self.dict = {('active','UCLslc3')   : CopyNonDQ2File,
-                     ('active','Lancaster') : CopyNonDQ2File,
-                     ('active', 'CERN')     : CopyNonDQ2FileCERN,
-                     ('active', 'CERNslc4') : CopyNonDQ2FileCERN, 
-                     ('passive', None )     : ListCastorFile}
-
-    def create(self, mode, site):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if (mode,site) in self.dict.keys():
-            return self.dict[(mode,site)]
-        else:
-            print 'Unknown mode in DataSetManager cfg, exiting...mode: %s, site: %s' % (mode, site)
-            sys.exit()
-
-class DQ2ActionFactory:
-
-    def __init__(self):
-        self.dict = {'active' : CopyDQ2Files,
-                     'passive': ListDQ2Files}
-
-    def create(self, mode):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if mode in self.dict.keys():
-            return self.dict[mode]
-        else:
-            print 'Unknown mode in DataSetManager cfg, exiting...mode: %s' % mode
-            sys.exit()
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_USRMaker.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_USRMaker.py
deleted file mode 100755
index 00142a4955ad7ad43538bc9b0b1ded545c1b1e2b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/Factory_USRMaker.py
+++ /dev/null
@@ -1,20 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from UserStuffRetrieverMaker import USRProjectMaker
-
-
-class USRMakerFactory:
-    def __init__(self):
-        self.dict = {'project'   : USRProjectMaker}
-                  
-
-    def create(self, releaseType):
-        '''selects the action to be performed
-        active  - copy file from castor
-        passive - list file on castor
-        '''
-        if releaseType in self.dict.keys():
-            return self.dict[releaseType]
-        else:
-            print 'Unknown releaseType in RTT topLevelConfig cfg, exiting...'
-            sys.exit()
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/LoggerSetup.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/LoggerSetup.py
deleted file mode 100755
index ec95146ac32d06c434eff5a3a2fe9420648cc8b2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/LoggerSetup.py
+++ /dev/null
@@ -1,279 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-# ----------------------------------------------------------------
-# ----- IMPORTS --------------------------------------------------
-# ----------------------------------------------------------------
-
-import sys
-import os, os.path
-import logging
-
-from Tools import now, mkdirIfNew, rmdirMkdir, changePerm
-from ToVariableLoggingHandler import ToVariableLoggingHandler
-from exc2string import exc2string
-
-# ----------------------------------------------------------------
-
-# pseudo private, not copied out on imports of this module
-dirToWhichLogsOutput = ''
-
-# ----------------------------------------------------------------
-
-def getDirToWhichLogsOutput():
-    return dirToWhichLogsOutput
-
-# ----------------------------------------------------------------
-# ----- SETUP FUNCTIONS ------------------------------------------
-# ----------------------------------------------------------------
-
-def setupLoggers(parentDir):
-    'Create the loggers for the RTT.'
-    
-    # ------------------------------
-    # Create the output dirs
-    # ------------------------------    
-    global dirToWhichLogsOutput
-    dirToWhichLogsOutput = createLogDirectories(parentDir)
-    
-    # ------------------------------
-    # Setup the format
-    # ------------------------------
-    formatter = createFormatter()
-
-    # -----------------------------------
-    # Create the full name of the logs
-    # -----------------------------------
-    mainLogName    = createLogFile(dirToWhichLogsOutput,'RTT.log','RTT.log\n')
-    errorLogName   = createLogFile(dirToWhichLogsOutput,'RTTerror.log','RTTerror.log\n') 
-    warningLogName = createLogFile(dirToWhichLogsOutput,'RTTwarning.log','RTTwarning.log\n') 
-    infoLogName    = createLogFile(dirToWhichLogsOutput,'RTTinfo.log','RTTinfo.log\n')
-
-    dirSnapsName   = createLogFile(dirToWhichLogsOutput,'dirSnapShots.log','dirSnapShots.log\n')
-    # ----------------------------------
-    # create the file handler loggers
-    # ----------------------------------
-    logFileHandler     = createFileHandler(mainLogName,logging.DEBUG,formatter)
-    errorFileHandler   = createFileHandler(errorLogName,logging.ERROR,formatter)
-    warningFileHandler = createFileHandler(warningLogName,logging.WARNING,formatter)
-    infoFileHandler    = createFileHandler(infoLogName,logging.INFO,formatter)
-
-    dirSnapsHandler    = createFileHandler(dirSnapsName,logging.DEBUG,formatter)
-    # ----------------------------------
-    # create the stream handler logger
-    # ----------------------------------
-    logStreamHandler = createStreamHandler(logging.INFO,formatter)
-
-    # ----------------------------------
-    # create the variable logging handler
-    # ----------------------------------
-    variableLoggingHandler = createVariableLoggingHandler(logging.INFO, formatter)
-
-    # -------------------------------------------
-    # add the handlers and set the logger level
-    # -------------------------------------------    
-    logger = addHandlers('rtt',
-                         logFileHandler,
-                         infoFileHandler,
-                         warningFileHandler,
-                         errorFileHandler,
-                         logStreamHandler,
-                         variableLoggingHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    dirSnapsLogger = addHandlers('dirSnaps', dirSnapsHandler)
-    dirSnapsLogger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    # ---------------------------------------------------------
-    # Now do all the above for a separate special case logger
-    # ---------------------------------------------------------
-
-    smallLogName     = createLogFile(dirToWhichLogsOutput,'RTTstandalone.log','RTTstandalone.log\n')
-    smallFileHandler = createFileHandler(smallLogName,logging.DEBUG,formatter)
-
-    smallLogger = addHandlers('rtt.standalone',smallFileHandler)
-    smallLogger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    return dirToWhichLogsOutput
-# ----------------------------------------------------------------
-
-def setupLoggerForDebug(level=None):
-    formatter = createFormatter()
-
-    level = loggerLevelFactory(level)
-
-    logStreamHandler = createStreamHandler(level,formatter)
-    logger = addHandlers(None,logStreamHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-# ----------------------------------------------------------------
-
-def setupLoggerForTestSuite(logDir, level=None):
-    formatter = createFormatter()    
-
-    # convert to logging variable. Sets default if necessary
-    level = loggerLevelFactory(level) 
-
-    fullPathToLog = createLogFile(logDir,
-                                  'testSuite.log',
-                                  'testSuite.log\n')
-    fileHandler = createFileHandler(fullPathToLog,level,formatter)
-        
-    #    logger = addHandlers(None,fileHandler)
-    logger = addHandlers('rtt',fileHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    print logger.handlers
-    return fileHandler # so we can remove it
-
-# ----------------------------------------------------------------
-def resetLoggerForTestSuite(logger, handler, directory):
-    
-    level     = handler.level
-    formatter = handler.formatter
-    logger.removeHandler(handler)
-    fullPathToLog = createLogFile(directory,
-                                  'testSuite.log',
-                                  'testSuite.log\n')
-
-    fileHandler = createFileHandler(fullPathToLog, level, formatter)
-        
-    #    logger = addHandlers(None,fileHandler)
-    logger.addHandler(fileHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    print logger.handlers
-    return fileHandler # so we can reset it later
-# ----------------------------------------------------------------
-# ----- HELPER FUNCTIONS -----------------------------------------
-# ----------------------------------------------------------------
-
-def createLogDirectories(parentDir):
-    if not os.path.exists(parentDir) or not os.path.isabs(parentDir):
-        print 'Unable to setup the logger because the path: ' + str(parentDir)
-        print 'either does not exist, or is not absolute.'
-        sys.exit()
-            
-    try:
-        mkdirIfNew(os.path.join(parentDir,'logs'))
-    except Exception, e:        
-        print 'Unable to create the path: ' + str(os.path.join(parentDir,'logs'))
-        print str(e)
-        print exc2string(sys.exc_info())
-        sys.exit()
-
-    logsDirName = 'logs/logfiles_'+now()
-    fullLogsDirName = os.path.join(parentDir, logsDirName)
-
-    try:
-        if os.path.exists(fullLogsDirName):
-            rmdirMkdir(fullLogsDirName)
-        else:
-            mkdirIfNew(fullLogsDirName)
-    except Exception, e:
-        print 'Unable to create path: ' + str(fullLogsDirName)
-        print str(e)
-        print exc2string(sys.exc_info())
-        sys.exit()                         
-
-    return fullLogsDirName
-
-# ----------------------------------------------------------------
-
-def createFileHandler(name,level,formatter):
-    theHandler = logging.FileHandler(name)
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-
-# ----------------------------------------------------------------
-
-def createStreamHandler(level,formatter):
-    theHandler = logging.StreamHandler()
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-
-# ----------------------------------------------------------------
-
-def createVariableLoggingHandler(level,formatter):
-    theHandler = ToVariableLoggingHandler()
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-    
-# ----------------------------------------------------------------
-
-def createFormatter():    
-    msgFmt    = '%(asctime)s %(filename)s, %(lineno)s, %(levelname)s %(message)s'
-    tmeFmt    = '%d%b%y %H:%M:%S'
-    formatter = logging.Formatter(msgFmt, tmeFmt)
-    return formatter
-
-# ----------------------------------------------------------------
-
-def createLogFile(parentDir,fileName,startText='A log file',permission=0775):
-    if not os.path.exists(parentDir) or not os.path.isabs(parentDir):
-        print 'The parent directory: ' + str(parentDir)
-        print 'in which to create the logger file: ' + str(fileName)
-        print 'is either inexistant, or is not absolute. It must be both.'
-        logging.shutdown()
-        sys.exit()
-
-    try:
-        fullPath = os.path.join(parentDir,fileName)
-        open(fullPath,'w').write(startText)
-    except Exception, e:
-        print 'Unable to create and write to the log file: '
-        print str(fullPath)
-        print str(e)
-        print exc2string(sys.exc_info())
-        logging.shutdown()
-        sys.exit()
-
-    try:
-        changePerm(fullPath,permission)
-    except Exception, e:
-        print 'Unable to set permission: ' + str(permission)
-        print 'on the log file: ' + str(fullPath)
-        print str(e)
-        print exc2string(sys.exc_info())
-        logging.shutdown()
-        sys.exit()
-    
-    return fullPath
-    
-# ----------------------------------------------------------------
-
-def addHandlers(theLoggerName=None,*handlers):
-    'Creates a logger instance and adds handlers to it. Returns the instance.'
-
-    logger = logging.getLogger(theLoggerName)
-    [logger.addHandler(handler) for handler in handlers]
-
-    return logger
-
-# ----------------------------------------------------------------
-
-def loggerLevelFactory(level=None):    
-    levels = {
-        'DEBUG':     logging.DEBUG,
-        'INFO':      logging.INFO,
-        'WARNING':   logging.WARNING,
-        'ERROR':     logging.ERROR,
-        'CRITICAL':  logging.CRITICAL
-        }
-    if level in levels:
-        level=levels[level]
-    else:
-        print 'Unknown level %s\n' %level
-        print 'Possible levels: ', str(levels.keys())
-        print 'setting level to DEBUG'
-        
-        level = logging.DEBUG
-
-    return level
-
-# ----------------------------------------------------------------
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/ProxyHandler.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/ProxyHandler.py
deleted file mode 100644
index eb321bcababf42afe8cb571446e0467b6804ffd0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/ProxyHandler.py
+++ /dev/null
@@ -1,47 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from popen2 import Popen4
-
-class ProxyHandler:
-    def __init__(self):
-        self.setProxyDetails()
-
-    def getTimeLeft(self):
-        if hasattr(self, 'timeleft'):
-            msg  = 'Proxy time left in Hours : '
-            msg += self.timeleft
-            return msg
-        
-        elif hasattr(self, 'ERROR'):
-            return self.ERROR
-
-        else:
-            msg  = 'No information on grid proxy found. '
-            msg += 'Maybe Grid UI is not setup'
-            return msg           
-
-    def setProxyDetails(self):
-        dict = {}
-        lines = Popen4('grid-proxy-info ').fromchild.readlines()
-        lines = [line.strip() for line in lines]
-
-        for line in lines:
-            parts = line.split(':')
-            if len(parts) < 2:
-                pass
-            else:
-                dict[parts[0].strip()] = parts[1].strip()
-
-        self.__dict__.update(dict)
-
-        #print str(self.__dict__)
-     
-    
-
-if __name__ == "__main__":
-    ph = ProxyHandler()
-    print ph.getTimeLeft()
-        
-        
-        
-        
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/ReportWriter.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/ReportWriter.py
deleted file mode 100755
index 788f380dd5fa47772e7cc4b1ea47e9cc9421cae5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/ReportWriter.py
+++ /dev/null
@@ -1,137 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os,os.path, time, logging
-from ProxyHandler import ProxyHandler
-
-logger = logging.getLogger('rtt')
-
-class ReportWriter:
-    def __init__(self, datasets, configs, packageConfFiles):
-        self.datasets = datasets
-        self.configs  = configs
-        self.pcf      = packageConfFiles
-
-    def write(self, reportDir):
-        #print 'Datasets :'+str(self.datasets)
-        self.reportDir  = reportDir
-        ph              = ProxyHandler()
-        
-        self.notToCopy  = [d for d in self.datasets if  d.isInDestDir]
-        self.tobeCopied = [d for d in self.datasets if not d.isInDestDir]
-
-        self.sortedNTC  = self.sortByPackage(self.notToCopy)
-        
-        logger.debug( 'to be copied :'+str(self.tobeCopied))
-        
-        self.copyFailed = [d for d in self.tobeCopied if d.copyStatus !='success']
-        logger.debug( 'copy Failed '+str(self.copyFailed))
-        
-        self.copyOK     = [d for d in self.tobeCopied if d.copyStatus =='success']
-        logger.debug( 'copyOK '+str(self.copyOK))
-
-        report      = 'DatasetManager.report'
-        self.report = os.path.join(self.reportDir, report)
-        
-        try:
-            print 'Working dir : '+str(os.getcwd())
-            #os.chdir(os.getcwd())
-            fh = open(self.report, 'w')
-            #fh.write('\n************************************************************')
-            fh.write('\nDataSetManager Report\t\t' + time.ctime())
-            fh.write('\n************************************************************\n\n')
-
-            #fh.write('\nDate \t\t\t\t\t: '+ time.ctime())
-
-            fh.write('\n\nRelease      \t\t\t\t: '+str(self.configs.get('release')))
-            fh.write('\nRelease Type \t\t\t\t: '+str(self.configs.get('releaseType')))
-            fh.write('\nProxy Information\t\t\t: '+str(ph.getTimeLeft()))
-            #fh.write('\n\n************************************************************')
-            fh.write('\n\nTotal number of packages found by RTT \t: '+str(len(self.pcf)))
-            fh.write('\n\nTotal number of datasets requested    \t: '+str(len(self.datasets)))
-
-            fh.write('\nNo of datasets to be copied:          \t: '+str(len(self.tobeCopied)))
-
-            fh.write('\nNo of datasets where copyOK:          \t: '+str(len(self.copyOK)))
-            fh.write('\nNo of datasets where copyFailed:      \t: '+str(len(self.copyFailed)))
-
-            #fh.write('\n\nDETAILS')
-            #fh.write('\n****************************************************************\n')
-            fh.write('\n\n\nDetails on datasets where copy failed:')
-           # fh.write('\n****************************************************************')
-            for ds in self.copyFailed:
-                fh.write('\n====================================')
-                fh.write('\nName         :'+str(ds.logicalName))
-                if ds.dsSource =='dq2':
-                    for file in ds.files: fh.write('\n'+file)
-                fh.write('\nCopy Status  :'+str(ds.copyStatus))
-                fh.write('\nFailure Reason :\n')
-                fh.write(str(ds.copyInfo))
-                fh.write('\nPackageCfg : '+str(ds.packageCfg))
-                fh.write('\n====================================\n')    
-            #fh.write('\n****************************************************************\n\n')
-            fh.write('\n\nDetails on datasets where copy succeeded:')
-            #fh.write('\n****************************************************************')
-            for ds in self.copyOK:
-                fh.write('\n====================================')
-                fh.write('\nName         :'+str(ds.logicalName))
-                if ds.dsSource =='dq2':
-                    for file in ds.files: fh.write('\n'+file)
-                fh.write('\nCopy Status  :'+str(ds.copyStatus))
-                fh.write('\nPhysicalName :'+str(ds.physicalName))
-                fh.write('\n====================================\n\n')    
-            fh.write('\n****************************************************************\n\n')
-
-            fh.write('\nRequested datasets already available locally')
-            fh.write('\n****************************************************************')
-            for key in self.sortedNTC.keys():
-                fh.write('\n\n'+str('package: '+key))
-                for ds in self.sortedNTC[key]:
-                    fh.write('\n'+str(ds.logicalName))
-                    #if ds.files:
-                    if ds.dsSource =='dq2':
-                        for file in ds.files:
-                            fh.write('\n\t'+str(file))
-                fh.write('\n\t---------------------------------------------------------------')
-
-
-            #for ds in self.notToCopy:
-            #    fh.write('\n'+str(ds.logicalName))
-            #    #if ds.files:
-            #    if ds.dsSource =='dq2':
-            #        for file in ds.files:
-            #            fh.write('\n\t'+str(file))
-        
-            fh.close()
-
-        except Exception ,e :
-            logger.error('Could not create a report file '+str(self.report))
-            print str(e)
-
-    def sortByPackage(self, dsObjs):
-        dict = {}
-
-        for ds in dsObjs:
-            if ds.packageCfg in dict.keys():
-                dict[ds.packageCfg].append(ds)
-            else:
-                dict[ds.packageCfg]=[ds]
-
-        return dict
-    
-class WriteTest:
-    def __init__(self):
-        pass
-
-    def write(self):
-        fh=open('testReport.results', 'w')
-        fh.write('\nThe test file ')
-
-
-if __name__ =='__main__':
-
-    rt = WriteTest()
-    rt.write()
-                           
-    
-                            
-    
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/UserStuffRetrieverMaker.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/UserStuffRetrieverMaker.py
deleted file mode 100755
index a311141a70296d040fe645e51b74f58337303f00..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/UserStuffRetrieverMaker.py
+++ /dev/null
@@ -1,84 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTConfigParser            import * 
-from formatCollection           import formatCollection
-from Paths                      import NoPackagePaths
-from Factory_UserStuffRetriever import UserStuffRetrieverFactory
-from ParametersForCMT           import ParametersForCMT
-from makeRTTDOM                 import makeRTTDOM
-import sys, os
-    
-import logging
-logger = logging.getLogger('rtt')
-
-class USRBase:
-    def __init__(self, rttConfFile,
-                 subString = '_TestConfiguration', logger=''):
-        self.rttConfFile = rttConfFile
-        self.subString   = subString
-        self.logger      = logger
-
-
-class USRProjectMaker:
-    def __init__(self, rttConfFile,logger, copierConfig,
-                 subString = '_TestConfiguration'):
-        
-        self.copierConfig     = copierConfig
-        self.logger           = logger
-        self.rttConfFile      = rttConfFile
-        self.subString        = subString
-        self.summaryDocument  = makeRTTDOM('Summary')
-    
-    def makeUSR(self):
-        logger.info( 'Making UserStuffRetriever Project ')
-        cfg = RTTConfiguration(self.rttConfFile)
-        self.confDict = cfg.config
-
-        #extend original Config
-        self.confDict['originalBranch'] = cfg.originalBranch
-        self.confDict['findNightly']    = cfg.findNightly
-        self.confDict['valid']          = cfg.valid
-        self.confDict['isNightly']      = cfg.isNightly
-        
-        self.RTTSrcDir        = self.copierConfig.get('rttsrc')
-        self.RTTLibDir        = self.copierConfig.get('rttLibDir')
-        self.logFileDir       = self.copierConfig.get('logDir')
-        
-        if self.confDict['release'] == 'nightly':           
-            lcr  = '/afs/cern.ch/atlas/software/builds/nightlies/'
-            lcr += self.confDict['originalBranch']
-            lcr += '/AtlasProduction/latest_copied_release'
-            
-            self.confDict['release'] = os.readlink(lcr)
-            self.logger.info('nightly Tag is '+str(os.readlink(lcr)))
-        
-        self.logger.debug(formatCollection(self.confDict))
-        
-        self.parametersForCMT = ParametersForCMT(self.confDict, self.logger)
-
-        self.paths = NoPackagePaths(self.RTTSrcDir,
-                                    self.RTTLibDir,
-                                    self.confDict,
-                                    self.logFileDir,
-                                    self.summaryDocument.createElement,
-                                    self.summaryDocument.createTextNode,
-                                    self.parametersForCMT,
-                                    self.logger)
-       
-        usr    = UserStuffRetrieverFactory(logger).create(self.paths, self.subString)
-        return usr
-        
-if __name__ == '__main__':
-
-    from DSMConfigReader import DSMConfigReader
-
-    cr  = DSMConfigReader('/home/en/RTT/dsmProject/dsmCfg.xml')
-    cfg = cr.readDSMcfg()
-    print str(cfg)
-    sys.path.append(cfg['rttsrc'])
-    usrMaker = USRProjectMaker(cfg['rttcfg'])
-    
-    usr = usrMaker.makeUSR()
-
-    print str(usr)
-    
diff --git a/Tools/RunTimeTester/dataCopier/copier_cern/src/getText.py b/Tools/RunTimeTester/dataCopier/copier_cern/src/getText.py
deleted file mode 100755
index 97bac8d87e24d2cb16a1a2d3e88f1f5eb0d8fa0c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_cern/src/getText.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string
-
-def getText(nodeList):
-  text = ''
-  for node in nodeList:
-    if node.nodeType == node.TEXT_NODE:
-      text = text + node.data
-      text = string.strip(text)
-  return text
diff --git a/Tools/RunTimeTester/dataCopier/copier_ucl/dsmCron b/Tools/RunTimeTester/dataCopier/copier_ucl/dsmCron
deleted file mode 100755
index 809abf094743a3e9764733a2498475805bc51665..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_ucl/dsmCron
+++ /dev/null
@@ -1,5 +0,0 @@
-30  7   * * * /usr/bin/sg atlasrtt /home/en/RTT/dataCopier/runDSManager.sh > /home/en/RTT/dataCopier/dsm.log
-20  13  * * * /usr/local/bin/python2.3 /home/en/RTT/KitsInstall/dev_KitInstaller.py > /home/en/RTT/KitsInstall/dev_kit_install.log
-
-
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_ucl/dsm_cfg_nightly.xml b/Tools/RunTimeTester/dataCopier/copier_ucl/dsm_cfg_nightly.xml
deleted file mode 100755
index c1e79bdeb1d1021c4f327239660027b9a80878fd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_ucl/dsm_cfg_nightly.xml
+++ /dev/null
@@ -1,15 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE dsmcfg SYSTEM "/home/en/RTT/dsmProject/DatasetManager.dtd">
-
-<dsmcfg>
-   <mode>active</mode>
-   <datasetCatalog>/unix/atlasrtt/RTT/xml/rttDataSetCatalog.xml</datasetCatalog>
-   <rttsrc>/home/en/RTT/Grid/Tools/RunTimeTester/src</rttsrc>
-   <reportDir>/unix/www/html/atlas/AtlasTesting</reportDir>
-   <rttcfg>/home/en/RTT/dsmProject/cfg_nightly.xml</rttcfg>
-   <datasetDir>/unix/atlasrtt/RTTdata</datasetDir>
-  <!-- <datasetDir>/home/en/RTT</datasetDir> -->
-</dsmcfg>
-
-
diff --git a/Tools/RunTimeTester/dataCopier/copier_ucl/old.tgz b/Tools/RunTimeTester/dataCopier/copier_ucl/old.tgz
deleted file mode 100755
index 5ee79567fa6b757b2e95e07b57da823126805318..0000000000000000000000000000000000000000
Binary files a/Tools/RunTimeTester/dataCopier/copier_ucl/old.tgz and /dev/null differ
diff --git a/Tools/RunTimeTester/dataCopier/copier_ucl/runDSManager.sh b/Tools/RunTimeTester/dataCopier/copier_ucl/runDSManager.sh
deleted file mode 100755
index 07208f1df577369ddabf3b6f4bf5d14a02e2d2ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/copier_ucl/runDSManager.sh
+++ /dev/null
@@ -1,5 +0,0 @@
-#! /bin/bash
-set +e
-source /home/en/RTT/setupDQ2env.sh
-source /usr/local/lcg/etc/profile.d/grid_env.sh
-/usr/local/bin/python2.3 /home/en/RTT/dataCopier/src/DataSetManager.py  /home/en/RTT/dataCopier/dsm_cfg_nightly.xml
\ No newline at end of file
diff --git a/Tools/RunTimeTester/dataCopier/copier_ucl/src.tgz b/Tools/RunTimeTester/dataCopier/copier_ucl/src.tgz
deleted file mode 100755
index 89975176b6f5fa4a4932dbed8bd24eab7894152d..0000000000000000000000000000000000000000
Binary files a/Tools/RunTimeTester/dataCopier/copier_ucl/src.tgz and /dev/null differ
diff --git a/Tools/RunTimeTester/dataCopier/formatCollection.py b/Tools/RunTimeTester/dataCopier/formatCollection.py
deleted file mode 100755
index da19213c494cb88af961b3cf0d8ac10d37aaafaf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/formatCollection.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from string import replace
-def formatCollection(c):
-    return replace(str(c),',',',\n  ')
diff --git a/Tools/RunTimeTester/dataCopier/getText.py b/Tools/RunTimeTester/dataCopier/getText.py
deleted file mode 100755
index 97bac8d87e24d2cb16a1a2d3e88f1f5eb0d8fa0c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/getText.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string
-
-def getText(nodeList):
-  text = ''
-  for node in nodeList:
-    if node.nodeType == node.TEXT_NODE:
-      text = text + node.data
-      text = string.strip(text)
-  return text
diff --git a/Tools/RunTimeTester/dataCopier/validateXMLFile.py b/Tools/RunTimeTester/dataCopier/validateXMLFile.py
deleted file mode 100755
index 3cb0359a0249c6ef44995df66eb5396759096f8a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/dataCopier/validateXMLFile.py
+++ /dev/null
@@ -1,57 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"run validation against a DTD. The DTD location is written into the xml file."
-
-from xml.parsers.xmlproc      import xmlproc, xmlval
-import os.path
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-# -------------------------------------------------------------------------
-
-def validateXMLFile(fileName):
-    # logger.critical('Hacked version of validateXMLFile !!!! always returns True!!!!')
-    # logger.critical('UCL is down - no DTD!!!')
-    # return True
-
-    try:
-        validator = xmlval.XMLValidator()
-        validator.parse_resource(fileName)
-        logger.debug(fileName+' is validated against DTD')
-    except SystemExit, e:
-        logger.error(fileName+' is NOT validated against DTD ')
-        msg  = 'caught SytemExit: maybe xml file does not refer to a '
-        msg += 'DTD or is invalid against DTD'
-        logger.error(msg)
-        raise Exception, str(os.path.basename(fileName)) + " not valid against DTD"
-    except Exception, e:
-        logger.error(fileName+' is NOT validated against DTD ')
-        msg =  'Unknown exception: '+sys.exc_info()
-        logger.error(e)
-        logger.error(msg)
-        raise Exception, str(os.path.basename(fileName)) + ' not valid against DTD'
-
-if __name__ == '__main__':
-
-    import getopt, os, sys
-
-    try:
-        opts,args = getopt.getopt(sys.argv[1:],'')
-    except getopt.GetoptError:
-        print '1 usage validateXMLFile fn'
-        sys.exit(0)
-
-
-    if len(args) != 1:
-        print '2 usage validateXMLFile fn'
-        sys.exit(0)
-
-    file = args[0]
-    if not os.path.exists(file):
-        print 'unknown file', file
-        sys.exit(0)
-
-    validateXMLFile(file)
diff --git a/Tools/RunTimeTester/doc/Checkers.html b/Tools/RunTimeTester/doc/Checkers.html
deleted file mode 100755
index a07bee3bacba49f2d99bcf86e271d71348b80369..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Checkers.html
+++ /dev/null
@@ -1,178 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Checkers</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Checkers</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Checkers.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Checkers.py</a></font></td></tr></table>
-    <p><tt>Checkers&nbsp;that&nbsp;run&nbsp;on&nbsp;&nbsp;logfiles&nbsp;returned&nbsp;after&nbsp;processing.<br>
-&nbsp;<br>
-simpleLogChecker&nbsp;-&nbsp;A&nbsp;Log&nbsp;checker&nbsp;that&nbsp;examines&nbsp;log&nbsp;files&nbsp;for&nbsp;the&nbsp;success&nbsp;message<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;printed&nbsp;by&nbsp;the&nbsp;&nbsp;athena&nbsp;Applicationmanager.<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;Gets&nbsp;the&nbsp;logname&nbsp;from&nbsp;a&nbsp;JobInfo&nbsp;object.<br>
-&nbsp;<br>
-atlfastRegressionChecker&nbsp;-&nbsp;Checks&nbsp;the&nbsp;output&nbsp;of&nbsp;the&nbsp;Atlfast&nbsp;Comparator&nbsp;for&nbsp;differences<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;not&nbsp;=&nbsp;0.<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<br>
-PS</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="copy.html">copy</a><br>
-<a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-<a href="shutil.html">shutil</a><br>
-</td><td width="25%" valign=top><a href="string.html">string</a><br>
-<a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top><a href="time.html">time</a><br>
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Checkers.html#Checker">Checker</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="Checkers.html#AtlfastRegressionChecker">AtlfastRegressionChecker</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Checkers.html#HistogramDiffer">HistogramDiffer</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Checkers.html#SimpleErrorLogChecker">SimpleErrorLogChecker</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Checkers.html#SimpleFileDiffer">SimpleFileDiffer</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Checkers.html#SimpleLogChecker">SimpleLogChecker</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="AtlfastRegressionChecker">class <strong>AtlfastRegressionChecker</strong></a>(<a href="Checkers.html#Checker">Checker</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="AtlfastRegressionChecker-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="AtlfastRegressionChecker-check"><strong>check</strong></a>(self, jMinder)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Checkers.html#Checker">Checker</a>:<br>
-<dl><dt><a name="AtlfastRegressionChecker-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="AtlfastRegressionChecker-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Checker">class <strong>Checker</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Checker-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="Checker-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Checker-check"><strong>check</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Checker-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="HistogramDiffer">class <strong>HistogramDiffer</strong></a>(<a href="Checkers.html#Checker">Checker</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="HistogramDiffer-__init__"><strong>__init__</strong></a>(self, comparisonType)</dt></dl>
-
-<dl><dt><a name="HistogramDiffer-check"><strong>check</strong></a>(self, jMinder)</dt></dl>
-
-<dl><dt><a name="HistogramDiffer-isLegalComparisonType"><strong>isLegalComparisonType</strong></a>(self, comparisonType)</dt><dd><tt>#&nbsp;to&nbsp;move&nbsp;into&nbsp;Legals&nbsp;class</tt></dd></dl>
-
-<hr>
-Methods inherited from <a href="Checkers.html#Checker">Checker</a>:<br>
-<dl><dt><a name="HistogramDiffer-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="HistogramDiffer-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="SimpleErrorLogChecker">class <strong>SimpleErrorLogChecker</strong></a>(<a href="Checkers.html#Checker">Checker</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="SimpleErrorLogChecker-__init__"><strong>__init__</strong></a>(self, testString)</dt></dl>
-
-<dl><dt><a name="SimpleErrorLogChecker-check"><strong>check</strong></a>(self, jMinder)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Checkers.html#Checker">Checker</a>:<br>
-<dl><dt><a name="SimpleErrorLogChecker-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="SimpleErrorLogChecker-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="SimpleFileDiffer">class <strong>SimpleFileDiffer</strong></a>(<a href="Checkers.html#Checker">Checker</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="SimpleFileDiffer-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SimpleFileDiffer-check"><strong>check</strong></a>(self, jMinder)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Checkers.html#Checker">Checker</a>:<br>
-<dl><dt><a name="SimpleFileDiffer-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="SimpleFileDiffer-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="SimpleLogChecker">class <strong>SimpleLogChecker</strong></a>(<a href="Checkers.html#Checker">Checker</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="SimpleLogChecker-__init__"><strong>__init__</strong></a>(self, searchStringList, vetoStringList<font color="#909090">=[]</font>)</dt></dl>
-
-<dl><dt><a name="SimpleLogChecker-check"><strong>check</strong></a>(self, jMinder)</dt></dl>
-
-<dl><dt><a name="SimpleLogChecker-checkSearchString"><strong>checkSearchString</strong></a>(self, str, fileName)</dt><dd><tt>determine&nbsp;if&nbsp;all&nbsp;the&nbsp;substrings&nbsp;in&nbsp;list&nbsp;self.<strong>searchStrings</strong></tt></dd></dl>
-
-<dl><dt><a name="SimpleLogChecker-checkVetoString"><strong>checkVetoString</strong></a>(self, str, fileName)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Checkers.html#Checker">Checker</a>:<br>
-<dl><dt><a name="SimpleLogChecker-__call__"><strong>__call__</strong></a>(self, minder)</dt></dl>
-
-<dl><dt><a name="SimpleLogChecker-setLogger"><strong>setLogger</strong></a>(self, newLogger)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>failure</strong> = -1<br>
-<strong>globalReporter</strong> = &lt;Reporter.GlobalReporter instance&gt;<br>
-<strong>logger</strong> = &lt;logging.Logger instance&gt;<br>
-<strong>success</strong> = 0</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Coloriser.html b/Tools/RunTimeTester/doc/Coloriser.html
deleted file mode 100755
index 0690042d6129c6268ccde4793c59c1637853b448..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Coloriser.html
+++ /dev/null
@@ -1,73 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Coloriser</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Coloriser</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Coloriser.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Coloriser.py</a></font></td></tr></table>
-    <p><tt>Coloriser&nbsp;returns&nbsp;string&nbsp;of&nbsp;ccolored&nbsp;htm&nbsp;table&nbsp;elements<br>
-coloriserMN&nbsp;M-&nbsp;number&nbsp;of&nbsp;colors<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;N&nbsp;-&nbsp;number&nbsp;of&nbsp;html&nbsp;table&nbsp;entries<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<br>
-<a href="#Coloriser21">Coloriser21</a>&nbsp;2:two&nbsp;colors&nbsp;(green/red.&nbsp;User&nbsp;gives&nbsp;green&nbsp;string,&nbsp;all&nbsp;others&nbsp;red<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;1:always&nbsp;returns&nbsp;1&nbsp;table&nbsp;entry<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<br>
-<a href="#Coloriser32">Coloriser32</a>&nbsp;3:three&nbsp;colors&nbsp;(green/red/brown.&nbsp;User&nbsp;gives&nbsp;green&nbsp;string,&nbsp;and&nbsp;red&nbsp;strings<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;all&nbsp;others&nbsp;brown<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;2:&nbsp;always&nbsp;return&nbsp;with&nbsp;2&nbsp;entries&nbsp;(padd&nbsp;if&nbsp;neccessary)</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Coloriser.html#Coloriser21">Coloriser21</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Coloriser.html#Coloriser32">Coloriser32</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Coloriser21">class <strong>Coloriser21</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Coloriser21-__init__"><strong>__init__</strong></a>(self, greenString)</dt></dl>
-
-<dl><dt><a name="Coloriser21-colorise"><strong>colorise</strong></a>(self, string)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Coloriser32">class <strong>Coloriser32</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Coloriser32-__init__"><strong>__init__</strong></a>(self, greenString, redString)</dt></dl>
-
-<dl><dt><a name="Coloriser32-colorise"><strong>colorise</strong></a>(self, string)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Commander.html b/Tools/RunTimeTester/doc/Commander.html
deleted file mode 100755
index e37dc331488f5d49c818cb9f40fe364f00b6dfe3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Commander.html
+++ /dev/null
@@ -1,132 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Commander</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Commander</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Commander.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Commander.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Commander.html#Commander">Commander</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="Commander.html#DummyCommander">DummyCommander</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Commander.html#LSFBatchCommander">LSFBatchCommander</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Commander.html#LinuxBatchCommander">LinuxBatchCommander</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Commander.html#LinuxInteractiveCommander">LinuxInteractiveCommander</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Commander.html#PBSCommander">PBSCommander</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Commander">class <strong>Commander</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt><a href="#Commander">Commander</a>&nbsp;-&nbsp;a&nbsp;base&nbsp;class&nbsp;that&nbsp;issues&nbsp;commands&nbsp;to&nbsp;the&nbsp;operating&nbsp;system.<br>
-Derived&nbsp;classe&nbsp;overide&nbsp;the&nbsp;commands&nbsp;to&nbsp;allow&nbsp;for&nbsp;differences&nbsp;due&nbsp;to<br>
-different&nbsp;mode&nbsp;values.<br>
-&nbsp;<br>
-Commands:&nbsp;&nbsp;athena<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Commander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="DummyCommander">class <strong>DummyCommander</strong></a>(<a href="Commander.html#Commander">Commander</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="DummyCommander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LSFBatchCommander">class <strong>LSFBatchCommander</strong></a>(<a href="Commander.html#Commander">Commander</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Commands&nbsp;for&nbsp;runing&nbsp;on&nbsp;the&nbsp;LSF&nbsp;batch&nbsp;queues&nbsp;at&nbsp;CERN.<br>
-Switched&nbsp;off&nbsp;mailing&nbsp;Peter<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LSFBatchCommander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxBatchCommander">class <strong>LinuxBatchCommander</strong></a>(<a href="Commander.html#Commander">Commander</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Commands&nbsp;for&nbsp;running&nbsp;on&nbsp;Linux&nbsp;using&nbsp;the&nbsp;batch&nbsp;command.<br>
-Does&nbsp;not&nbsp;work&nbsp;because&nbsp;I&nbsp;could&nbsp;not&nbsp;make&nbsp;the&nbsp;linux&nbsp;batch&nbsp;command&nbsp;work.<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxBatchCommander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxInteractiveCommander">class <strong>LinuxInteractiveCommander</strong></a>(<a href="Commander.html#Commander">Commander</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Commands&nbsp;for&nbsp;running&nbsp;interactively&nbsp;(foreground)&nbsp;on&nbsp;Linux<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxInteractiveCommander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PBSCommander">class <strong>PBSCommander</strong></a>(<a href="Commander.html#Commander">Commander</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Commands&nbsp;for&nbsp;running&nbsp;jobs&nbsp;on&nbsp;the&nbsp;UCL&nbsp;HEP&nbsp;PBS&nbsp;queues<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PBSCommander-submitCommand"><strong>submitCommand</strong></a>(self, jDescriptor)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-commanderFactory"><strong>commanderFactory</strong></a>(mode)</dt><dd><tt>Returns&nbsp;a&nbsp;<a href="#Commander">Commander</a>&nbsp;instance&nbsp;according&nbsp;to&nbsp;mode</tt></dd></dl>
-</td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/ConfigReader.html b/Tools/RunTimeTester/doc/ConfigReader.html
deleted file mode 100755
index 12133ef75884827dace44357b456a373805efdb1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/ConfigReader.html
+++ /dev/null
@@ -1,72 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module ConfigReader</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>ConfigReader</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/ConfigReader.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/ConfigReader.py</a></font></td></tr></table>
-    <p><tt>Reads&nbsp;the&nbsp;jobs&nbsp;config&nbsp;file&nbsp;to&nbsp;determine&nbsp;which&nbsp;jobs&nbsp;to&nbsp;run.<br>
-&nbsp;<br>
-The&nbsp;presence&nbsp;of&nbsp;the&nbsp;5th&nbsp;parameter&nbsp;on&nbsp;a&nbsp;line&nbsp;in&nbsp;the&nbsp;config&nbsp;file&nbsp;is&nbsp;optional&nbsp;as<br>
-far&nbsp;as&nbsp;<a href="#ConfigReader">ConfigReader</a>&nbsp;is&nbsp;concerned,&nbsp;and&nbsp;gives&nbsp;the&nbsp;data&nbsp;set&nbsp;to&nbsp;use<br>
-(some&nbsp;jobs&nbsp;do&nbsp;not&nbsp;use&nbsp;data&nbsp;sets).<br>
-&nbsp;<br>
-The&nbsp;first&nbsp;parameter&nbsp;gives&nbsp;the&nbsp;job&nbsp;options&nbsp;file.&nbsp;If&nbsp;there&nbsp;is&nbsp;no&nbsp;extension&nbsp;to&nbsp;the<br>
-file&nbsp;name,&nbsp;'.py'&nbsp;is&nbsp;understood.&nbsp;Thus,&nbsp;to&nbsp;use&nbsp;the&nbsp;old&nbsp;'.txt'&nbsp;option&nbsp;files,&nbsp;supply<br>
-the&nbsp;options&nbsp;file&nbsp;with&nbsp;the&nbsp;.txt&nbsp;extension&nbsp;attatched&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;name.</tt></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="ConfigReader.html#ConfigReader">ConfigReader</a>
-</font></dt></dl>
- 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="ConfigReader">class <strong>ConfigReader</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="ConfigReader-__init__"><strong>__init__</strong></a>(self, confFile)</dt></dl>
-
-<dl><dt><a name="ConfigReader-fixJobOptions"><strong>fixJobOptions</strong></a>(self, jo)</dt></dl>
-
-<dl><dt><a name="ConfigReader-jobs"><strong>jobs</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'ConfigReader'</dl>
-
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './ConfigReader.py'<br>
-<strong>__name__</strong> = 'ConfigReader'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/GlobalDebugFlag.html b/Tools/RunTimeTester/doc/GlobalDebugFlag.html
deleted file mode 100755
index c73851d2d57a5ea2ff96d14d56d9a0fb27274798..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/GlobalDebugFlag.html
+++ /dev/null
@@ -1,25 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module GlobalDebugFlag</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>GlobalDebugFlag</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/GlobalDebugFlag.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/GlobalDebugFlag.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './GlobalDebugFlag.pyc'<br>
-<strong>__name__</strong> = 'GlobalDebugFlag'<br>
-<strong>dbg</strong> = 0</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/HTMLpage3.html b/Tools/RunTimeTester/doc/HTMLpage3.html
deleted file mode 100755
index e81e561dcd1588e79380d46740517d0b8d95d04f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/HTMLpage3.html
+++ /dev/null
@@ -1,45 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module HTMLpage3</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>HTMLpage3</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/HTMLpage3.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/HTMLpage3.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-writePage3"><strong>writePage3</strong></a>(paths, build, jobMinder, page3Path)</dt><dd><tt>Looks&nbsp;at&nbsp;the&nbsp;status&nbsp;file&nbsp;in&nbsp;the&nbsp;results&nbsp;subdirectory<br>
-for&nbsp;the&nbsp;release&nbsp;being&nbsp;processed&nbsp;(paths.release),&nbsp;together<br>
-with&nbsp;the&nbsp;template&nbsp;Results.html&nbsp;from&nbsp;the&nbsp;work&nbsp;directory<br>
-(self.<strong>workDir</strong>)&nbsp;to&nbsp;make&nbsp;an&nbsp;updated&nbsp;html&nbsp;file&nbsp;with&nbsp;the<br>
-current&nbsp;status.&nbsp;This&nbsp;page&nbsp;is&nbsp;written&nbsp;to&nbsp;the&nbsp;results<br>
-directory&nbsp;for&nbsp;this&nbsp;release.</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './HTMLpage3.py'<br>
-<strong>__name__</strong> = 'HTMLpage3'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/HTMLpages.html b/Tools/RunTimeTester/doc/HTMLpages.html
deleted file mode 100755
index f891d36048f186a920c20721998317fecd2ac433..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/HTMLpages.html
+++ /dev/null
@@ -1,54 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module HTMLpages</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>HTMLpages</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/HTMLpages.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/HTMLpages.py</a></font></td></tr></table>
-    <p><tt>Writes&nbsp;out&nbsp;html&nbsp;pages&nbsp;for&nbsp;the&nbsp;current&nbsp;RTT&nbsp;run.<br>
-&nbsp;<br>
-&nbsp;<br>
-Functions:<br>
-&nbsp;<br>
-resultsDir&nbsp;-&nbsp;returns&nbsp;the&nbsp;names&nbsp;of&nbsp;directories&nbsp;containg&nbsp;results.<br>
-writePage2&nbsp;-&nbsp;writes&nbsp;out&nbsp;an&nbsp;html&nbsp;page&nbsp;with&nbsp;jobname,&nbsp;dataset,&nbsp;statuses,&nbsp;and&nbsp;reference<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;release&nbsp;for&nbsp;all&nbsp;jobs.<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;Also&nbsp;calls&nbsp;writePage3&nbsp;to&nbsp;give&nbsp;detailed&nbsp;information&nbsp;for&nbsp;each&nbsp;page.</tt></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-resultsDirs"><strong>resultsDirs</strong></a>(build, paths)</dt></dl>
- <dl><dt><a name="-writePage2"><strong>writePage2</strong></a>(paths, build, jobMinders)</dt><dd><tt>Looks&nbsp;at&nbsp;the&nbsp;status&nbsp;file&nbsp;in&nbsp;the&nbsp;results&nbsp;subdirectory<br>
-for&nbsp;the&nbsp;release&nbsp;being&nbsp;processed&nbsp;(self.<strong>paths</strong>.release),&nbsp;together<br>
-with&nbsp;the&nbsp;template&nbsp;Results.html&nbsp;from&nbsp;the&nbsp;work&nbsp;directory<br>
-(self.<strong>workDir</strong>)&nbsp;to&nbsp;make&nbsp;an&nbsp;updated&nbsp;html&nbsp;file&nbsp;with&nbsp;the<br>
-current&nbsp;status.&nbsp;This&nbsp;page&nbsp;is&nbsp;written&nbsp;to&nbsp;the&nbsp;results<br>
-directory&nbsp;for&nbsp;this&nbsp;release.</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './HTMLpages.py'<br>
-<strong>__name__</strong> = 'HTMLpages'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/IChecker.py.html b/Tools/RunTimeTester/doc/IChecker.py.html
deleted file mode 100755
index d01a6eafb1696aa69f9685d04e7ea43056c2ec82..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/IChecker.py.html
+++ /dev/null
@@ -1,52 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module IChecker.py</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong><a href="IChecker.html"><font color="#ffffff">IChecker</font></a>.py</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/IChecker.py.hold">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/IChecker.py.hold</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="IChecker.py.html#IChecker">IChecker</a>
-</font></dt></dl>
- 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="IChecker">class <strong>IChecker</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="IChecker-check"><strong>check</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'IChecker.py'</dl>
-
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './IChecker.py.hold'<br>
-<strong>__name__</strong> = 'IChecker.py'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/JobDescriptor.html b/Tools/RunTimeTester/doc/JobDescriptor.html
deleted file mode 100755
index af29c72a469050cb42228ea96f0d5f42af2f0f9b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/JobDescriptor.html
+++ /dev/null
@@ -1,77 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module JobDescriptor</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>JobDescriptor</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/JobDescriptor.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/JobDescriptor.py</a></font></td></tr></table>
-    <p><tt>Class&nbsp;that&nbsp;stocks&nbsp;information&nbsp;about&nbsp;jobs&nbsp;to&nbsp;be&nbsp;run.&nbsp;One&nbsp;instance&nbsp;per&nbsp;job&nbsp;is&nbsp;created.<br>
-Responsible&nbsp;for&nbsp;setting&nbsp;up&nbsp;the&nbsp;run&nbsp;directory&nbsp;for&nbsp;a&nbsp;job.&nbsp;Eventually&nbsp;becomes&nbsp;a&nbsp;data&nbsp;member<br>
-of&nbsp;JobMinder.</tt></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="GlobalDebugFlag.html">GlobalDebugFlag</a><br>
-</td><td width="25%" valign=top><a href="anydbm.html">anydbm</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="shutil.html">shutil</a><br>
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="JobDescriptor.html#JobDescriptor">JobDescriptor</a>
-</font></dt></dl>
- 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JobDescriptor">class <strong>JobDescriptor</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="JobDescriptor-__init__"><strong>__init__</strong></a>(self, jobOptions, package, jobGroup, queue)</dt></dl>
-
-<dl><dt><a name="JobDescriptor-dump"><strong>dump</strong></a>(self)</dt><dd><tt>Dumps&nbsp;out&nbsp;jobname</tt></dd></dl>
-
-<dl><dt><a name="JobDescriptor-setupRunDir"><strong>setupRunDir</strong></a>(self, paths, build)</dt><dd><tt>Creates&nbsp;run&nbsp;dirctory&nbsp;(path&nbsp;calculated&nbsp;from&nbsp;release,&nbsp;build,<br>
-job&nbsp;options&nbsp;file&nbsp;name..)<br>
-Copies&nbsp;needed&nbsp;files&nbsp;to&nbsp;it<br>
-Sets&nbsp;u-p&nbsp;the&nbsp;ZEBRA&nbsp;link.<br>
-Uses&nbsp;ScriptWirter&nbsp;to&nbsp;write&nbsp;a&nbsp;run&nbsp;script&nbsp;to&nbsp;it.</tt></dd></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobDescriptor'</dl>
-
-<dl><dt><strong>jobIdentifier</strong> = 0</dl>
-
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './JobDescriptor.pyc'<br>
-<strong>__name__</strong> = 'JobDescriptor'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/JobGroupKits.html b/Tools/RunTimeTester/doc/JobGroupKits.html
deleted file mode 100755
index 80568d4aec32c90e348fb5209ffa498d42adbc98..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/JobGroupKits.html
+++ /dev/null
@@ -1,77 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module JobGroupKits</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>JobGroupKits</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/JobGroupKits.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/JobGroupKits.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="copy.html">copy</a><br>
-</td><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Borg.html#Borg">Borg.Borg</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="JobGroupKits.html#JobGroupKits">JobGroupKits</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JobGroupKits">class <strong>JobGroupKits</strong></a>(<a href="Borg.html#Borg">Borg.Borg</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>A&nbsp;class&nbsp;that&nbsp;contains&nbsp;information&nbsp;about&nbsp;all&nbsp;kits.&nbsp;Each&nbsp;kit<br>
-is&nbsp;represented&nbsp;by&nbsp;a&nbsp;dictionary.&nbsp;These&nbsp;dictionaries&nbsp;are&nbsp;stored&nbsp;in&nbsp;a<br>
-dictionary&nbsp;where&nbsp;they&nbsp;are&nbsp;keyed&nbsp;on&nbsp;the&nbsp;group&nbsp;name.<br>
-&nbsp;<br>
-The&nbsp;class&nbsp;is&nbsp;derived&nbsp;from&nbsp;the&nbsp;<a href="Borg.html#Borg">Borg</a>&nbsp;implementation&nbsp;of&nbsp;the&nbsp;Singleton&nbsp;pattern.<br>
-Instantiatiating&nbsp;the&nbsp;class&nbsp;with&nbsp;no&nbsp;argument&nbsp;gives&nbsp;a&nbsp;fiexed&nbsp;state.<br>
-Instantiating&nbsp;it&nbsp;with&nbsp;a&nbsp;list&nbsp;of&nbsp;kits&nbsp;changes&nbsp;the&nbsp;existing&nbsp;instances<br>
-everywhere.<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="JobGroupKits-__init__"><strong>__init__</strong></a>(self, rttKits<font color="#909090">=None</font>, userKits<font color="#909090">=None</font>)</dt><dd><tt>If&nbsp;no&nbsp;argument&nbsp;is&nbsp;given,&nbsp;the&nbsp;common&nbsp;object&nbsp;is&nbsp;returned<br>
-Passing&nbsp;a&nbsp;list&nbsp;of&nbsp;kits&nbsp;will&nbsp;change&nbsp;all&nbsp;the&nbsp;existing&nbsp;instances</tt></dd></dl>
-
-<dl><dt><a name="JobGroupKits-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobGroupKits-getAllKits"><strong>getAllKits</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobGroupKits-kit"><strong>kit</strong></a>(self, kitName)</dt></dl>
-
-<dl><dt><a name="JobGroupKits-nKits"><strong>nKits</strong></a>(self)</dt></dl>
-
-<hr>
-Data and other attributes inherited from <a href="Borg.html#Borg">Borg.Borg</a>:<br>
-<dl><dt><strong>shared_state</strong> = {}</dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/JobMinder.html b/Tools/RunTimeTester/doc/JobMinder.html
deleted file mode 100755
index 145ef1e4111e4d65f61a48cd29dbc3f5a5f2e658..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/JobMinder.html
+++ /dev/null
@@ -1,418 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module JobMinder</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>JobMinder</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/JobMinder.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/JobMinder.py</a></font></td></tr></table>
-    <p><tt>This&nbsp;file&nbsp;contains:<br>
-a&nbsp;factory&nbsp;function&nbsp;&nbsp;&nbsp;-&nbsp;&nbsp;&nbsp;to&nbsp;create&nbsp;instances&nbsp;of&nbsp;sub&nbsp;classes&nbsp;of&nbsp;<a href="#JobMinder">JobMinder</a><br>
-a&nbsp;base&nbsp;class&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;-&nbsp;&nbsp;&nbsp;<a href="#JobMinder">JobMinder</a><br>
-&nbsp;<br>
-<a href="#JobMinder">JobMinder</a>&nbsp;acts&nbsp;like&nbsp;a&nbsp;state&nbsp;machine.&nbsp;It&nbsp;has&nbsp;states:&nbsp;queued,&nbsp;running,&nbsp;success,&nbsp;error,<br>
-it&nbsp;nows&nbsp;how&nbsp;to&nbsp;proceed&nbsp;from&nbsp;one&nbsp;state&nbsp;to&nbsp;the&nbsp;next,&nbsp;and&nbsp;species&nbsp;the&nbsp;actions&nbsp;to&nbsp;be<br>
-performed&nbsp;when&nbsp;the&nbsp;state&nbsp;changes.&nbsp;The&nbsp;way&nbsp;jobs&nbsp;are&nbsp;submited,&nbsp;and&nbsp;the&nbsp;names&nbsp;of<br>
-log&nbsp;files&nbsp;are&nbsp;examples&nbsp;of&nbsp;items&nbsp;that&nbsp;change&nbsp;with&nbsp;mode.<br>
-&nbsp;<br>
-Subclassing&nbsp;is&nbsp;used&nbsp;to&nbsp;specialse&nbsp;according&nbsp;to&nbsp;mode.<br>
-&nbsp;<br>
-The&nbsp;base&nbsp;class&nbsp;also&nbsp;specifies&nbsp;common&nbsp;tasks&nbsp;that&nbsp;must&nbsp;be&nbsp;performed&nbsp;across&nbsp;job&nbsp;mode<br>
-boundaries:<br>
-&nbsp;<br>
--&nbsp;database&nbsp;manipulations&nbsp;for&nbsp;jobs&nbsp;which&nbsp;store&nbsp;reference&nbsp;files<br>
--&nbsp;various&nbsp;protection&nbsp;manipulations&nbsp;to&nbsp;allow&nbsp;other&nbsp;mebers&nbsp;of&nbsp;the&nbsp;same&nbsp;UNIX&nbsp;group<br>
-&nbsp;&nbsp;to&nbsp;manipulate&nbsp;files&nbsp;on&nbsp;future&nbsp;runs.<br>
--&nbsp;copying&nbsp;files&nbsp;to&nbsp;be&nbsp;stored&nbsp;at&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;job<br>
--&nbsp;actions&nbsp;to&nbsp;be&nbsp;taken&nbsp;of&nbsp;there&nbsp;is&nbsp;a&nbsp;forced&nbsp;shutdown<br>
-&nbsp;&nbsp;(triggered&nbsp;by&nbsp;receiving&nbsp;a&nbsp;signal&nbsp;15)<br>
-PS</tt></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="anydbm.html">anydbm</a><br>
-<a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="shutil.html">shutil</a><br>
-<a href="string.html">string</a><br>
-</td><td width="25%" valign=top><a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="JobMinder.html#JobMinder">JobMinder</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="JobMinder.html#LXBatchJobMinder">LXBatchJobMinder</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#LinuxBatchJobMinder">LinuxBatchJobMinder</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#LinuxInteractiveJobMinder">LinuxInteractiveJobMinder</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#PBSJobMinder">PBSJobMinder</a>
-</font></dt></dl>
-</dd>
-<dt><font face="helvetica, arial"><a href="JobMinder.html#LXBatchJobMinderFactory">LXBatchJobMinderFactory</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#LinuxBatchJobMinderFactory">LinuxBatchJobMinderFactory</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#LinuxInteractiveJobMinderFactory">LinuxInteractiveJobMinderFactory</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobMinder.html#PBSJobMinderFactory">PBSJobMinderFactory</a>
-</font></dt></dl>
- 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JobMinder">class <strong>JobMinder</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="JobMinder-__init__"><strong>__init__</strong></a>(self, jDescriptor, popen)</dt><dd><tt>resDir&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;-&nbsp;results&nbsp;directory:&nbsp;recieves&nbsp;the&nbsp;logfile<br>
-status</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-copyKeepFiles"><strong>copyKeepFiles</strong></a>(self)</dt><dd><tt>#&nbsp;copy&nbsp;files&nbsp;to&nbsp;be&nbsp;kept&nbsp;(log&nbsp;histos,&nbsp;ntuples...)&nbsp;to&nbsp;results&nbsp;dir</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-forceCloseDown"><strong>forceCloseDown</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-fromError"><strong>fromError</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-fromQueued"><strong>fromQueued</strong></a>(self)</dt><dd><tt>#&nbsp;subclasses&nbsp;to&nbsp;override&nbsp;this&nbsp;method</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-fromRunning"><strong>fromRunning</strong></a>(self)</dt><dd><tt>#&nbsp;subclasses&nbsp;to&nbsp;override&nbsp;this&nbsp;method</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-fromSuccess"><strong>fromSuccess</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-performChecks"><strong>performChecks</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-poll"><strong>poll</strong></a>(self)</dt><dd><tt>makes&nbsp;status&nbsp;transitions</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-runPostScripts"><strong>runPostScripts</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-setProtection"><strong>setProtection</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobMinder-setProtections"><strong>setProtections</strong></a>(self, path, perm)</dt></dl>
-
-<dl><dt><a name="JobMinder-setStatus"><strong>setStatus</strong></a>(self, status)</dt><dd><tt>#&nbsp;Do&nbsp;not&nbsp;set&nbsp;the&nbsp;status&nbsp;member&nbsp;data&nbsp;directly!</tt></dd></dl>
-
-<dl><dt><a name="JobMinder-setStatusList"><strong>setStatusList</strong></a>(self, statusL)</dt></dl>
-
-<dl><dt><a name="JobMinder-storeRefFile"><strong>storeRefFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LXBatchJobMinder">class <strong>LXBatchJobMinder</strong></a>(<a href="JobMinder.html#JobMinder">JobMinder</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>A&nbsp;<a href="#JobMinder">JobMinder</a>&nbsp;class&nbsp;for&nbsp;running&nbsp;on&nbsp;the&nbsp;LX&nbsp;batch&nbsp;queues<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LXBatchJobMinder-__init__"><strong>__init__</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-fromQueued"><strong>fromQueued</strong></a>(self)</dt><dd><tt>Finds&nbsp;out&nbsp;from&nbsp;popen&nbsp;if&nbsp;queued&nbsp;on&nbsp;Interactive<br>
-stores&nbsp;the&nbsp;submission&nbsp;log</tt></dd></dl>
-
-<dl><dt><a name="LXBatchJobMinder-fromRunning"><strong>fromRunning</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-getLogFile"><strong>getLogFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = '<font color="#c040c0">\n</font>    A JobMinder class for running on the LX batch queues<font color="#c040c0">\n</font>    '</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-<hr>
-Methods inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><a name="LXBatchJobMinder-copyKeepFiles"><strong>copyKeepFiles</strong></a>(self)</dt><dd><tt>#&nbsp;copy&nbsp;files&nbsp;to&nbsp;be&nbsp;kept&nbsp;(log&nbsp;histos,&nbsp;ntuples...)&nbsp;to&nbsp;results&nbsp;dir</tt></dd></dl>
-
-<dl><dt><a name="LXBatchJobMinder-forceCloseDown"><strong>forceCloseDown</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-fromError"><strong>fromError</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-fromSuccess"><strong>fromSuccess</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-performChecks"><strong>performChecks</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-poll"><strong>poll</strong></a>(self)</dt><dd><tt>makes&nbsp;status&nbsp;transitions</tt></dd></dl>
-
-<dl><dt><a name="LXBatchJobMinder-runPostScripts"><strong>runPostScripts</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-setProtection"><strong>setProtection</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-setProtections"><strong>setProtections</strong></a>(self, path, perm)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-setStatus"><strong>setStatus</strong></a>(self, status)</dt><dd><tt>#&nbsp;Do&nbsp;not&nbsp;set&nbsp;the&nbsp;status&nbsp;member&nbsp;data&nbsp;directly!</tt></dd></dl>
-
-<dl><dt><a name="LXBatchJobMinder-setStatusList"><strong>setStatusList</strong></a>(self, statusL)</dt></dl>
-
-<dl><dt><a name="LXBatchJobMinder-storeRefFile"><strong>storeRefFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LXBatchJobMinderFactory">class <strong>LXBatchJobMinderFactory</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LXBatchJobMinderFactory-create"><strong>create</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxBatchJobMinder">class <strong>LinuxBatchJobMinder</strong></a>(<a href="JobMinder.html#JobMinder">JobMinder</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxBatchJobMinder-__init__"><strong>__init__</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-fromQueued"><strong>fromQueued</strong></a>(self)</dt><dd><tt>Finds&nbsp;out&nbsp;from&nbsp;popen&nbsp;if&nbsp;queued&nbsp;on&nbsp;Interactive<br>
-stores&nbsp;the&nbsp;submission&nbsp;log</tt></dd></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-fromRunning"><strong>fromRunning</strong></a>(self)</dt><dd><tt>Looks&nbsp;to&nbsp;see&nbsp;if&nbsp;the&nbsp;log&nbsp;has&nbsp;appeared<br>
-Checks&nbsp;it&nbsp;for&nbsp;errors&nbsp;when&nbsp;it&nbsp;appears</tt></dd></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-<hr>
-Methods inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><a name="LinuxBatchJobMinder-copyKeepFiles"><strong>copyKeepFiles</strong></a>(self)</dt><dd><tt>#&nbsp;copy&nbsp;files&nbsp;to&nbsp;be&nbsp;kept&nbsp;(log&nbsp;histos,&nbsp;ntuples...)&nbsp;to&nbsp;results&nbsp;dir</tt></dd></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-forceCloseDown"><strong>forceCloseDown</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-fromError"><strong>fromError</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-fromSuccess"><strong>fromSuccess</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-performChecks"><strong>performChecks</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-poll"><strong>poll</strong></a>(self)</dt><dd><tt>makes&nbsp;status&nbsp;transitions</tt></dd></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-runPostScripts"><strong>runPostScripts</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-setProtection"><strong>setProtection</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-setProtections"><strong>setProtections</strong></a>(self, path, perm)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-setStatus"><strong>setStatus</strong></a>(self, status)</dt><dd><tt>#&nbsp;Do&nbsp;not&nbsp;set&nbsp;the&nbsp;status&nbsp;member&nbsp;data&nbsp;directly!</tt></dd></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-setStatusList"><strong>setStatusList</strong></a>(self, statusL)</dt></dl>
-
-<dl><dt><a name="LinuxBatchJobMinder-storeRefFile"><strong>storeRefFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxBatchJobMinderFactory">class <strong>LinuxBatchJobMinderFactory</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxBatchJobMinderFactory-create"><strong>create</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxInteractiveJobMinder">class <strong>LinuxInteractiveJobMinder</strong></a>(<a href="JobMinder.html#JobMinder">JobMinder</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxInteractiveJobMinder-__init__"><strong>__init__</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-fromQueued"><strong>fromQueued</strong></a>(self)</dt><dd><tt>Finds&nbsp;out&nbsp;from&nbsp;popen&nbsp;if&nbsp;queued&nbsp;on&nbsp;Interactive<br>
-stores&nbsp;the&nbsp;submission&nbsp;log</tt></dd></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-fromRunning"><strong>fromRunning</strong></a>(self)</dt><dd><tt>Looks&nbsp;to&nbsp;see&nbsp;if&nbsp;the&nbsp;log&nbsp;has&nbsp;appeared<br>
-Checks&nbsp;it&nbsp;for&nbsp;errors&nbsp;when&nbsp;it&nbsp;appears</tt></dd></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-<hr>
-Methods inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><a name="LinuxInteractiveJobMinder-copyKeepFiles"><strong>copyKeepFiles</strong></a>(self)</dt><dd><tt>#&nbsp;copy&nbsp;files&nbsp;to&nbsp;be&nbsp;kept&nbsp;(log&nbsp;histos,&nbsp;ntuples...)&nbsp;to&nbsp;results&nbsp;dir</tt></dd></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-forceCloseDown"><strong>forceCloseDown</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-fromError"><strong>fromError</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-fromSuccess"><strong>fromSuccess</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-performChecks"><strong>performChecks</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-poll"><strong>poll</strong></a>(self)</dt><dd><tt>makes&nbsp;status&nbsp;transitions</tt></dd></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-runPostScripts"><strong>runPostScripts</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-setProtection"><strong>setProtection</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-setProtections"><strong>setProtections</strong></a>(self, path, perm)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-setStatus"><strong>setStatus</strong></a>(self, status)</dt><dd><tt>#&nbsp;Do&nbsp;not&nbsp;set&nbsp;the&nbsp;status&nbsp;member&nbsp;data&nbsp;directly!</tt></dd></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-setStatusList"><strong>setStatusList</strong></a>(self, statusL)</dt></dl>
-
-<dl><dt><a name="LinuxInteractiveJobMinder-storeRefFile"><strong>storeRefFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LinuxInteractiveJobMinderFactory">class <strong>LinuxInteractiveJobMinderFactory</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LinuxInteractiveJobMinderFactory-create"><strong>create</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PBSJobMinder">class <strong>PBSJobMinder</strong></a>(<a href="JobMinder.html#JobMinder">JobMinder</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PBSJobMinder-__init__"><strong>__init__</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-findLog"><strong>findLog</strong></a>(self)</dt><dd><tt>Looks&nbsp;to&nbsp;see&nbsp;if&nbsp;the&nbsp;log&nbsp;has&nbsp;appeared<br>
-Checks&nbsp;it&nbsp;for&nbsp;errors&nbsp;when&nbsp;it&nbsp;appears</tt></dd></dl>
-
-<dl><dt><a name="PBSJobMinder-fromQueued"><strong>fromQueued</strong></a>(self)</dt><dd><tt>Finds&nbsp;out&nbsp;from&nbsp;popen&nbsp;if&nbsp;queued&nbsp;on&nbsp;PBS<br>
-stores&nbsp;the&nbsp;submission&nbsp;log</tt></dd></dl>
-
-<dl><dt><a name="PBSJobMinder-fromRunning"><strong>fromRunning</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-<hr>
-Methods inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><a name="PBSJobMinder-copyKeepFiles"><strong>copyKeepFiles</strong></a>(self)</dt><dd><tt>#&nbsp;copy&nbsp;files&nbsp;to&nbsp;be&nbsp;kept&nbsp;(log&nbsp;histos,&nbsp;ntuples...)&nbsp;to&nbsp;results&nbsp;dir</tt></dd></dl>
-
-<dl><dt><a name="PBSJobMinder-forceCloseDown"><strong>forceCloseDown</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-fromError"><strong>fromError</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-fromSuccess"><strong>fromSuccess</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-performChecks"><strong>performChecks</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-poll"><strong>poll</strong></a>(self)</dt><dd><tt>makes&nbsp;status&nbsp;transitions</tt></dd></dl>
-
-<dl><dt><a name="PBSJobMinder-runPostScripts"><strong>runPostScripts</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-setProtection"><strong>setProtection</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-setProtections"><strong>setProtections</strong></a>(self, path, perm)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-setStatus"><strong>setStatus</strong></a>(self, status)</dt><dd><tt>#&nbsp;Do&nbsp;not&nbsp;set&nbsp;the&nbsp;status&nbsp;member&nbsp;data&nbsp;directly!</tt></dd></dl>
-
-<dl><dt><a name="PBSJobMinder-setStatusList"><strong>setStatusList</strong></a>(self, statusL)</dt></dl>
-
-<dl><dt><a name="PBSJobMinder-storeRefFile"><strong>storeRefFile</strong></a>(self)</dt></dl>
-
-<hr>
-Data and non-method functions inherited from <a href="JobMinder.html#JobMinder">JobMinder</a>:<br>
-<dl><dt><strong>lDbg</strong> = 1</dl>
-
-</td></tr></table> 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PBSJobMinderFactory">class <strong>PBSJobMinderFactory</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PBSJobMinderFactory-create"><strong>create</strong></a>(self, jDescriptor, popen)</dt></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'JobMinder'</dl>
-
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './JobMinder.py'<br>
-<strong>__name__</strong> = 'JobMinder'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/JobOptionsGenerator.html b/Tools/RunTimeTester/doc/JobOptionsGenerator.html
deleted file mode 100755
index 87c267d27bab5d245fed9e8e9c3605eb9173c6ec..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/JobOptionsGenerator.html
+++ /dev/null
@@ -1,75 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module JobOptionsGenerator</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>JobOptionsGenerator</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/JobOptionsGenerator.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/JobOptionsGenerator.py</a></font></td></tr></table>
-    <p><tt>Standalone&nbsp;script&nbsp;to&nbsp;generate&nbsp;job&nbsp;option&nbsp;files&nbsp;for&nbsp;Atlfast&nbsp;using&nbsp;a&nbsp;template<br>
-script,&nbsp;and&nbsp;passing&nbsp;in&nbsp;parameters&nbsp;to&nbsp;sepcigy&nbsp;AF&nbsp;options.<br>
-Author:&nbsp;P&nbsp;Sherwood,&nbsp;using&nbsp;work&nbsp;of&nbsp;J.&nbsp;Couchman</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-<a href="shutil.html">shutil</a><br>
-</td><td width="25%" valign=top><a href="string.html">string</a><br>
-<a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top><a href="time.html">time</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="JobOptionsGenerator.html#JobOptionsGenerator">JobOptionsGenerator</a>
-</font></dt><dt><font face="helvetica, arial"><a href="JobOptionsGenerator.html#JobOptions_Generator">JobOptions_Generator</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JobOptionsGenerator">class <strong>JobOptionsGenerator</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="JobOptionsGenerator-__init__"><strong>__init__</strong></a>(self, shareDir)</dt></dl>
-
-<dl><dt><a name="JobOptionsGenerator-makeAll"><strong>makeAll</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobOptionsGenerator-makeAtlfast"><strong>makeAtlfast</strong></a>(self)</dt><dd><tt>create&nbsp;jobOptions&nbsp;files,&nbsp;and&nbsp;write&nbsp;out&nbsp;a&nbsp;config&nbsp;table&nbsp;for&nbsp;Atlfast</tt></dd></dl>
-
-<dl><dt><a name="JobOptionsGenerator-makeGenerators"><strong>makeGenerators</strong></a>(self)</dt><dd><tt>create&nbsp;jobOptions&nbsp;files</tt></dd></dl>
-
-<dl><dt><a name="JobOptionsGenerator-makeJobOptions"><strong>makeJobOptions</strong></a>(self, subs, jobOptions)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JobOptions_Generator">class <strong>JobOptions_Generator</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="JobOptions_Generator-BSdatajobOptions"><strong>BSdatajobOptions</strong></a>(self, jobOptions, jDescriptor)</dt></dl>
-
-<dl><dt><a name="JobOptions_Generator-BSdatasetCommand"><strong>BSdatasetCommand</strong></a>(self, jd)</dt></dl>
-
-<dl><dt><a name="JobOptions_Generator-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JobOptions_Generator-datasetCommand"><strong>datasetCommand</strong></a>(self, jd)</dt></dl>
-
-<dl><dt><a name="JobOptions_Generator-dc2jobOptions"><strong>dc2jobOptions</strong></a>(self, jobOptions, jDescriptor)</dt></dl>
-
-</td></tr></table></td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Legals.html b/Tools/RunTimeTester/doc/Legals.html
deleted file mode 100755
index f3aec9aca8635c79ff4dae43e0671a2ee29ba00d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Legals.html
+++ /dev/null
@@ -1,166 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Legals</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Legals</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Legals.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Legals.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="copy.html">copy</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Legals.html#LegalsBase">LegalsBase</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="Legals.html#LegalsMonolithicN0X">LegalsMonolithicN0X</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Legals.html#LegalsMonolithicNX0">LegalsMonolithicNX0</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Legals.html#LegalsProject">LegalsProject</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LegalsBase">class <strong>LegalsBase</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LegalsBase-__init__"><strong>__init__</strong></a>(self, hardCoded, logger)</dt></dl>
-
-<dl><dt><a name="LegalsBase-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsBase-convertLogicalNightly"><strong>convertLogicalNightly</strong></a>(self, release)</dt><dd><tt>nightly&nbsp;release&nbsp;names&nbsp;in&nbsp;config&nbsp;files&nbsp;are&nbsp;atlrel_x<br>
-convert&nbsp;to&nbsp;the&nbsp;name&nbsp;appropriate&nbsp;to&nbsp;the&nbsp;branch&nbsp;and&nbsp;releaseType<br>
-if&nbsp;the&nbsp;release&nbsp;is&nbsp;of&nbsp;the&nbsp;form&nbsp;atlrel_x.&nbsp;Return&nbsp;unchanged&nbsp;otherwise<br>
-&nbsp;eg&nbsp;for&nbsp;numbered&nbsp;releases</tt></dd></dl>
-
-<dl><dt><a name="LegalsBase-getNumberedReleases"><strong>getNumberedReleases</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsBase-isLegal"><strong>isLegal</strong></a>(self, key, val)</dt></dl>
-
-<dl><dt><a name="LegalsBase-isNightlyInRelease"><strong>isNightlyInRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsBase-isNightlyRelease"><strong>isNightlyRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsBase-validKeys"><strong>validKeys</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsBase-vals"><strong>vals</strong></a>(self, key)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LegalsMonolithicN0X">class <strong>LegalsMonolithicN0X</strong></a>(<a href="Legals.html#LegalsBase">LegalsBase</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LegalsMonolithicN0X-__init__"><strong>__init__</strong></a>(self, hardCoded, logger)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Legals.html#LegalsBase">LegalsBase</a>:<br>
-<dl><dt><a name="LegalsMonolithicN0X-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-convertLogicalNightly"><strong>convertLogicalNightly</strong></a>(self, release)</dt><dd><tt>nightly&nbsp;release&nbsp;names&nbsp;in&nbsp;config&nbsp;files&nbsp;are&nbsp;atlrel_x<br>
-convert&nbsp;to&nbsp;the&nbsp;name&nbsp;appropriate&nbsp;to&nbsp;the&nbsp;branch&nbsp;and&nbsp;releaseType<br>
-if&nbsp;the&nbsp;release&nbsp;is&nbsp;of&nbsp;the&nbsp;form&nbsp;atlrel_x.&nbsp;Return&nbsp;unchanged&nbsp;otherwise<br>
-&nbsp;eg&nbsp;for&nbsp;numbered&nbsp;releases</tt></dd></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-getNumberedReleases"><strong>getNumberedReleases</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-isLegal"><strong>isLegal</strong></a>(self, key, val)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-isNightlyInRelease"><strong>isNightlyInRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-isNightlyRelease"><strong>isNightlyRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-validKeys"><strong>validKeys</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicN0X-vals"><strong>vals</strong></a>(self, key)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LegalsMonolithicNX0">class <strong>LegalsMonolithicNX0</strong></a>(<a href="Legals.html#LegalsBase">LegalsBase</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>#&nbsp;----------------------------------------------------------------------&nbsp;&nbsp;&nbsp;<br>
-#&nbsp;//////////////////////////////////////////////////////////////////////<br>
-#&nbsp;----------------------------------------------------------------------<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LegalsMonolithicNX0-__init__"><strong>__init__</strong></a>(self, hardCoded, logger)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Legals.html#LegalsBase">LegalsBase</a>:<br>
-<dl><dt><a name="LegalsMonolithicNX0-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-convertLogicalNightly"><strong>convertLogicalNightly</strong></a>(self, release)</dt><dd><tt>nightly&nbsp;release&nbsp;names&nbsp;in&nbsp;config&nbsp;files&nbsp;are&nbsp;atlrel_x<br>
-convert&nbsp;to&nbsp;the&nbsp;name&nbsp;appropriate&nbsp;to&nbsp;the&nbsp;branch&nbsp;and&nbsp;releaseType<br>
-if&nbsp;the&nbsp;release&nbsp;is&nbsp;of&nbsp;the&nbsp;form&nbsp;atlrel_x.&nbsp;Return&nbsp;unchanged&nbsp;otherwise<br>
-&nbsp;eg&nbsp;for&nbsp;numbered&nbsp;releases</tt></dd></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-getNumberedReleases"><strong>getNumberedReleases</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-isLegal"><strong>isLegal</strong></a>(self, key, val)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-isNightlyInRelease"><strong>isNightlyInRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-isNightlyRelease"><strong>isNightlyRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-validKeys"><strong>validKeys</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsMonolithicNX0-vals"><strong>vals</strong></a>(self, key)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LegalsProject">class <strong>LegalsProject</strong></a>(<a href="Legals.html#LegalsBase">LegalsBase</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LegalsProject-__init__"><strong>__init__</strong></a>(self, hardCoded, logger)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Legals.html#LegalsBase">LegalsBase</a>:<br>
-<dl><dt><a name="LegalsProject-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsProject-convertLogicalNightly"><strong>convertLogicalNightly</strong></a>(self, release)</dt><dd><tt>nightly&nbsp;release&nbsp;names&nbsp;in&nbsp;config&nbsp;files&nbsp;are&nbsp;atlrel_x<br>
-convert&nbsp;to&nbsp;the&nbsp;name&nbsp;appropriate&nbsp;to&nbsp;the&nbsp;branch&nbsp;and&nbsp;releaseType<br>
-if&nbsp;the&nbsp;release&nbsp;is&nbsp;of&nbsp;the&nbsp;form&nbsp;atlrel_x.&nbsp;Return&nbsp;unchanged&nbsp;otherwise<br>
-&nbsp;eg&nbsp;for&nbsp;numbered&nbsp;releases</tt></dd></dl>
-
-<dl><dt><a name="LegalsProject-getNumberedReleases"><strong>getNumberedReleases</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsProject-isLegal"><strong>isLegal</strong></a>(self, key, val)</dt></dl>
-
-<dl><dt><a name="LegalsProject-isNightlyInRelease"><strong>isNightlyInRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsProject-isNightlyRelease"><strong>isNightlyRelease</strong></a>(self, rel)</dt></dl>
-
-<dl><dt><a name="LegalsProject-validKeys"><strong>validKeys</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LegalsProject-vals"><strong>vals</strong></a>(self, key)</dt></dl>
-
-</td></tr></table></td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/PathCompleter.html b/Tools/RunTimeTester/doc/PathCompleter.html
deleted file mode 100755
index 86ebac030f9c59d7c84fb8821717b228d3080e68..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/PathCompleter.html
+++ /dev/null
@@ -1,63 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module PathCompleter</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>PathCompleter</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/PathCompleter.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/PathCompleter.py</a></font></td></tr></table>
-    <p><tt><a href="#PathCompleter">PathCompleter</a>&nbsp;&nbsp;builds&nbsp;paths&nbsp;to&nbsp;a&nbsp;file&nbsp;from&nbsp;a&nbsp;release&nbsp;and&nbsp;&nbsp;package.</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="re.html">re</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="PathCompleter.html#PathCompleter">PathCompleter</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PathCompleter">class <strong>PathCompleter</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PathCompleter-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PathCompleter-fileFromReleasedFile"><strong>fileFromReleasedFile</strong></a>(self, package, file, paths)</dt><dd><tt>returns&nbsp;path&nbsp;to&nbsp;file&nbsp;in&nbsp;a&nbsp;release</tt></dd></dl>
-
-<dl><dt><a name="PathCompleter-releaseAreaCMTfile"><strong>releaseAreaCMTfile</strong></a>(self, package, paths)</dt><dd><tt>returns&nbsp;path&nbsp;to&nbsp;the&nbsp;cmt&nbsp;directory&nbsp;given&nbsp;release&nbsp;and&nbsp;<br>
-package&nbsp;name</tt></dd></dl>
-
-<dl><dt><a name="PathCompleter-releaseAreaCMTfile2"><strong>releaseAreaCMTfile2</strong></a>(self, package, paths)</dt><dd><tt>returns&nbsp;path&nbsp;to&nbsp;the&nbsp;cmt&nbsp;directory&nbsp;given&nbsp;release&nbsp;and&nbsp;<br>
-package&nbsp;name<br>
-&nbsp;<br>
-Timing&nbsp;tests&nbsp;show&nbsp;that&nbsp;releaseAreaCMTfile2&nbsp;is&nbsp;a&nbsp;little&nbsp;faster&nbsp;than&nbsp;releaseAreaCMTfile</tt></dd></dl>
-
-<dl><dt><a name="PathCompleter-versionDir"><strong>versionDir</strong></a>(self, package, paths)</dt><dd><tt>returns&nbsp;path&nbsp;to&nbsp;CMT&nbsp;added&nbsp;directory&nbsp;(ie&nbsp;package-xx-yy-zz)</tt></dd></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Paths.html b/Tools/RunTimeTester/doc/Paths.html
deleted file mode 100755
index cfc0b624448f839672124c60f142fb38b6da8257..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Paths.html
+++ /dev/null
@@ -1,143 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Paths</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Paths</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Paths.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Paths.py</a></font></td></tr></table>
-    <p><tt>#&nbsp;from&nbsp;siteSpecificsFactory&nbsp;import&nbsp;siteSpecificsFactory</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="Reporter.html">Reporter</a><br>
-<a href="copy.html">copy</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-<a href="string.html">string</a><br>
-</td><td width="25%" valign=top><a href="sys.html">sys</a><br>
-<a href="time.html">time</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Paths.html#NoPackagePaths">NoPackagePaths</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="Paths.html#Paths">Paths</a>
-</font></dt></dl>
-</dd>
-<dt><font face="helvetica, arial"><a href="Paths.html#PathsForDescriptors">PathsForDescriptors</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Paths.html#PathsForTestRun">PathsForTestRun</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="NoPackagePaths">class <strong>NoPackagePaths</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt><a href="#Paths">Paths</a>&nbsp;is&nbsp;a&nbsp;bag&nbsp;object&nbsp;of&nbsp;random&nbsp;bits&nbsp;and&nbsp;pieces.<br>
-It&nbsp;is&nbsp;instantiated&nbsp;using&nbsp;information&nbsp;from&nbsp;various&nbsp;sources,&nbsp;but&nbsp;mainly<br>
-from&nbsp;the&nbsp;top&nbsp;level&nbsp;RTT&nbsp;configuration&nbsp;file,&nbsp;This&nbsp;raw&nbsp;information&nbsp;is&nbsp;read&nbsp;in<br>
-via&nbsp;a&nbsp;parser&nbsp;which&nbsp;puts&nbsp;into&nbsp;the&nbsp;rttConfig&nbsp;object&nbsp;&nbsp;passed&nbsp;to&nbsp;this&nbsp;class.<br>
-The&nbsp;constructor&nbsp;manipulates&nbsp;the&nbsp;information&nbsp;so&nbsp;that&nbsp;it&nbsp;is&nbsp;useful&nbsp;to&nbsp;the<br>
-rest&nbsp;of&nbsp;the&nbsp;program.<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="NoPackagePaths-__init__"><strong>__init__</strong></a>(self, rttTagBeingRun, rttConfig, timer, logDir, logger)</dt><dd><tt>rttConfig&nbsp;is&nbsp;an&nbsp;object&nbsp;containing&nbsp;of&nbsp;parameters&nbsp;obtained&nbsp;by&nbsp;reading&nbsp;the<br>
-top&nbsp;level&nbsp;configuration&nbsp;file.<br>
-&nbsp;<br>
-This&nbsp;constructor&nbsp;constructs&nbsp;a&nbsp;number&nbsp;of&nbsp;paths.</tt></dd></dl>
-
-<dl><dt><a name="NoPackagePaths-__str__"><strong>__str__</strong></a>(self)</dt><dd><tt>#------------------------------------------------------------------------</tt></dd></dl>
-
-<dl><dt><a name="NoPackagePaths-createResultsDirDictionary"><strong>createResultsDirDictionary</strong></a>(self, resBase)</dt><dd><tt>Sets&nbsp;up&nbsp;a&nbsp;dictionary&nbsp;containing&nbsp;the&nbsp;various&nbsp;paths&nbsp;concerned&nbsp;with<br>
-results.<br>
-e.g.&nbsp;&lt;resBase&gt;/opt/N.0.X/10.0.1/kit<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&lt;resBase&gt;/opt/N.0.X/10.0.1/release</tt></dd></dl>
-
-<dl><dt><a name="NoPackagePaths-createWorkDirDictionary"><strong>createWorkDirDictionary</strong></a>(self, workBase)</dt></dl>
-
-<dl><dt><a name="NoPackagePaths-dump"><strong>dump</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="NoPackagePaths-dumpHeader"><strong>dumpHeader</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="NoPackagePaths-updateWithNightlyRelease"><strong>updateWithNightlyRelease</strong></a>(self, release)</dt><dd><tt>Update&nbsp;paths.release,&nbsp;rttConfig&nbsp;and&nbsp;the&nbsp;work/res&nbsp;dirs&nbsp;with&nbsp;the&nbsp;found<br>
-(by&nbsp;Tester.findNightlyRelease:&nbsp;which&nbsp;calls&nbsp;this&nbsp;method)&nbsp;latest&nbsp;nightly&nbsp;release.<br>
-Until&nbsp;this&nbsp;point&nbsp;self.<strong>release</strong>&nbsp;=&nbsp;'nightly'.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Paths">class <strong>Paths</strong></a>(<a href="Paths.html#NoPackagePaths">NoPackagePaths</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Paths-__init__"><strong>__init__</strong></a>(self, rttConfig, rttTagBeingRun, timer, logDir, cvsPackage, rttPackage, logger)</dt></dl>
-
-<dl><dt><a name="Paths-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Paths-checkPath"><strong>checkPath</strong></a>(self, path)</dt><dd><tt>#------------------------------------------------------------------------</tt></dd></dl>
-
-<dl><dt><a name="Paths-dumpHeader"><strong>dumpHeader</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="Paths.html#NoPackagePaths">NoPackagePaths</a>:<br>
-<dl><dt><a name="Paths-createResultsDirDictionary"><strong>createResultsDirDictionary</strong></a>(self, resBase)</dt><dd><tt>Sets&nbsp;up&nbsp;a&nbsp;dictionary&nbsp;containing&nbsp;the&nbsp;various&nbsp;paths&nbsp;concerned&nbsp;with<br>
-results.<br>
-e.g.&nbsp;&lt;resBase&gt;/opt/N.0.X/10.0.1/kit<br>
-&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&lt;resBase&gt;/opt/N.0.X/10.0.1/release</tt></dd></dl>
-
-<dl><dt><a name="Paths-createWorkDirDictionary"><strong>createWorkDirDictionary</strong></a>(self, workBase)</dt></dl>
-
-<dl><dt><a name="Paths-dump"><strong>dump</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Paths-updateWithNightlyRelease"><strong>updateWithNightlyRelease</strong></a>(self, release)</dt><dd><tt>Update&nbsp;paths.release,&nbsp;rttConfig&nbsp;and&nbsp;the&nbsp;work/res&nbsp;dirs&nbsp;with&nbsp;the&nbsp;found<br>
-(by&nbsp;Tester.findNightlyRelease:&nbsp;which&nbsp;calls&nbsp;this&nbsp;method)&nbsp;latest&nbsp;nightly&nbsp;release.<br>
-Until&nbsp;this&nbsp;point&nbsp;self.<strong>release</strong>&nbsp;=&nbsp;'nightly'.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PathsForDescriptors">class <strong>PathsForDescriptors</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PathsForDescriptors-__init__"><strong>__init__</strong></a>(self, paths)</dt></dl>
-
-<dl><dt><a name="PathsForDescriptors-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PathsForTestRun">class <strong>PathsForTestRun</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PathsForTestRun-__init__"><strong>__init__</strong></a>(self, paths)</dt></dl>
-
-<dl><dt><a name="PathsForTestRun-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>globalReporter</strong> = &lt;Reporter.GlobalReporter instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/PostScripts.html b/Tools/RunTimeTester/doc/PostScripts.html
deleted file mode 100755
index 37ffc62b7bb11e1a3d4c9a62a7aca4ddb5786fbe..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/PostScripts.html
+++ /dev/null
@@ -1,90 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module PostScripts</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>PostScripts</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/PostScripts.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/PostScripts.py</a></font></td></tr></table>
-    <p><tt>Functions&nbsp;to&nbsp;run&nbsp;after&nbsp;a&nbsp;jobs&nbsp;have&nbsp;completed.&nbsp;Organised&nbsp;by&nbsp;job&nbsp;group</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="re.html">re</a><br>
-</td><td width="25%" valign=top><a href="time.html">time</a><br>
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="PostScripts.html#RecExCommonPostScript">RecExCommonPostScript</a>
-</font></dt><dt><font face="helvetica, arial"><a href="PostScripts.html#SitePostScripts">SitePostScripts</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="PostScripts.html#EgammaWatcherPostScript">EgammaWatcherPostScript</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="EgammaWatcherPostScript">class <strong>EgammaWatcherPostScript</strong></a>(<a href="PostScripts.html#SitePostScripts">SitePostScripts</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Class methods defined here:<br>
-<dl><dt><a name="EgammaWatcherPostScript-run"><strong>run</strong></a>(self, site)<font color="#909090"><font face="helvetica, arial"> from __builtin__.classobj</font></font></dt><dd><tt>run&nbsp;some&nbsp;root&nbsp;macros</tt></dd></dl>
-
-<hr>
-Data and other attributes defined here:<br>
-<dl><dt><strong>psLabel</strong> = 'site'</dl>
-
-<hr>
-Class methods inherited from <a href="PostScripts.html#SitePostScripts">SitePostScripts</a>:<br>
-<dl><dt><a name="EgammaWatcherPostScript-setSiteSpecifics"><strong>setSiteSpecifics</strong></a>(self, site)<font color="#909090"><font face="helvetica, arial"> from __builtin__.classobj</font></font></dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RecExCommonPostScript">class <strong>RecExCommonPostScript</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Class methods defined here:<br>
-<dl><dt><a name="RecExCommonPostScript-run"><strong>run</strong></a>(self, logFile)<font color="#909090"><font face="helvetica, arial"> from __builtin__.classobj</font></font></dt><dd><tt>Extract&nbsp;lines&nbsp;from&nbsp;log&nbsp;and&nbsp;store&nbsp;in&nbsp;logExtract</tt></dd></dl>
-
-<hr>
-Data and other attributes defined here:<br>
-<dl><dt><strong>psLabel</strong> = 'log'</dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="SitePostScripts">class <strong>SitePostScripts</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Class methods defined here:<br>
-<dl><dt><a name="SitePostScripts-setSiteSpecifics"><strong>setSiteSpecifics</strong></a>(self, site)<font color="#909090"><font face="helvetica, arial"> from __builtin__.classobj</font></font></dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/RTTSConfigReader.html b/Tools/RunTimeTester/doc/RTTSConfigReader.html
deleted file mode 100755
index 180a5e2ba28493ea244352708965a8a2657b9d7f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/RTTSConfigReader.html
+++ /dev/null
@@ -1,72 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module RTTSConfigReader</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>RTTSConfigReader</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/RTTSConfigReader.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/RTTSConfigReader.py</a></font></td></tr></table>
-    <p><tt>A&nbsp;configuration&nbsp;file&nbsp;reader:</tt></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="RTTSConfigReader.html#RTTSConfigReader">RTTSConfigReader</a>
-</font></dt></dl>
- 
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RTTSConfigReader">class <strong>RTTSConfigReader</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2>&nbsp;</td></tr>
-<tr><td>&nbsp;&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="RTTSConfigReader-__init__"><strong>__init__</strong></a>(self, confFile, basePath)</dt><dd><tt>Provides&nbsp;default&nbsp;values.<br>
-Read&nbsp;config&nbsp;file.<br>
-If&nbsp;paths&nbsp;are&nbsp;given&nbsp;as&nbsp;relative&nbsp;paths,&nbsp;make&nbsp;them&nbsp;absolute&nbsp;wrt&nbsp;to&nbsp;&nbsp;basePath<br>
-this&nbsp;makes&nbsp;the&nbsp;code&nbsp;for&nbsp;robust&nbsp;if&nbsp;against&nbsp;the&nbsp;effect&nbsp;of&nbsp;changing&nbsp;directories.<br>
-Perform&nbsp;simple&nbsp;validation&nbsp;checks.<br>
-Throw&nbsp;exception&nbsp;if&nbsp;config&nbsp;file&nbsp;parameter&nbsp;is&nbsp;invalid.</tt></dd></dl>
-
-<dl><dt><a name="RTTSConfigReader-fixPath"><strong>fixPath</strong></a>(self, path)</dt><dd><tt>&nbsp;convert&nbsp;relative&nbsp;path&nbsp;to&nbsp;an&nbsp;absolute&nbsp;path</tt></dd></dl>
-
-<dl><dt><a name="RTTSConfigReader-params"><strong>params</strong></a>(self)</dt><dd><tt>Return&nbsp;a&nbsp;list&nbsp;of&nbsp;the&nbsp;values&nbsp;read&nbsp;from&nbsp;the&nbsp;config&nbsp;file.</tt></dd></dl>
-
-<hr>
-Data and non-method functions defined here:<br>
-<dl><dt><strong>__doc__</strong> = None</dl>
-
-<dl><dt><strong>__module__</strong> = 'RTTSConfigReader'</dl>
-
-<dl><dt><strong>lDbg</strong> = 0</dl>
-
-</td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './RTTSConfigReader.py'<br>
-<strong>__name__</strong> = 'RTTSConfigReader'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/RTTSException.html b/Tools/RunTimeTester/doc/RTTSException.html
deleted file mode 100755
index 3870b04c31d0474760cbaabe0e03cd35a13cf325..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/RTTSException.html
+++ /dev/null
@@ -1,149 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module RTTSException</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>RTTSException</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/RTTSException.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/RTTSException.py</a></font></td></tr></table>
-    <p><tt><a href="exceptions.html#Exception">Exception</a>&nbsp;class&nbsp;that&nbsp;carries&nbsp;an&nbsp;optional&nbsp;datum,&nbsp;eg&nbsp;error&nbsp;string</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="exceptions.html#Exception">exceptions.Exception</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="RTTSException.html#RTTSException">RTTSException</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="RTTSException.html#BadlyConfiguredPackage">BadlyConfiguredPackage</a>
-</font></dt><dt><font face="helvetica, arial"><a href="RTTSException.html#ConfigFileError">ConfigFileError</a>
-</font></dt><dt><font face="helvetica, arial"><a href="RTTSException.html#TimeoutException">TimeoutException</a>
-</font></dt><dt><font face="helvetica, arial"><a href="RTTSException.html#UnableToProcessRTTconffile">UnableToProcessRTTconffile</a>
-</font></dt></dl>
-</dd>
-</dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="BadlyConfiguredPackage">class <strong>BadlyConfiguredPackage</strong></a>(<a href="RTTSException.html#RTTSException">RTTSException</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="RTTSException.html#BadlyConfiguredPackage">BadlyConfiguredPackage</a></dd>
-<dd><a href="RTTSException.html#RTTSException">RTTSException</a></dd>
-<dd><a href="exceptions.html#Exception">exceptions.Exception</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="BadlyConfiguredPackage-__init__"><strong>__init__</strong></a>(self, args<font color="#909090">=None</font>)</dt></dl>
-
-<hr>
-Methods inherited from <a href="RTTSException.html#RTTSException">RTTSException</a>:<br>
-<dl><dt><a name="BadlyConfiguredPackage-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="exceptions.html#Exception">exceptions.Exception</a>:<br>
-<dl><dt><a name="BadlyConfiguredPackage-__getitem__"><strong>__getitem__</strong></a>(...)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="ConfigFileError">class <strong>ConfigFileError</strong></a>(<a href="RTTSException.html#RTTSException">RTTSException</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="RTTSException.html#ConfigFileError">ConfigFileError</a></dd>
-<dd><a href="RTTSException.html#RTTSException">RTTSException</a></dd>
-<dd><a href="exceptions.html#Exception">exceptions.Exception</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="ConfigFileError-__init__"><strong>__init__</strong></a>(self, args<font color="#909090">=None</font>)</dt></dl>
-
-<hr>
-Methods inherited from <a href="RTTSException.html#RTTSException">RTTSException</a>:<br>
-<dl><dt><a name="ConfigFileError-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="exceptions.html#Exception">exceptions.Exception</a>:<br>
-<dl><dt><a name="ConfigFileError-__getitem__"><strong>__getitem__</strong></a>(...)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RTTSException">class <strong>RTTSException</strong></a>(<a href="exceptions.html#Exception">exceptions.Exception</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="RTTSException-__init__"><strong>__init__</strong></a>(self, args<font color="#909090">=None</font>)</dt></dl>
-
-<dl><dt><a name="RTTSException-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="exceptions.html#Exception">exceptions.Exception</a>:<br>
-<dl><dt><a name="RTTSException-__getitem__"><strong>__getitem__</strong></a>(...)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="TimeoutException">class <strong>TimeoutException</strong></a>(<a href="RTTSException.html#RTTSException">RTTSException</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="RTTSException.html#TimeoutException">TimeoutException</a></dd>
-<dd><a href="RTTSException.html#RTTSException">RTTSException</a></dd>
-<dd><a href="exceptions.html#Exception">exceptions.Exception</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="TimeoutException-__init__"><strong>__init__</strong></a>(self, args<font color="#909090">=None</font>)</dt></dl>
-
-<hr>
-Methods inherited from <a href="RTTSException.html#RTTSException">RTTSException</a>:<br>
-<dl><dt><a name="TimeoutException-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="exceptions.html#Exception">exceptions.Exception</a>:<br>
-<dl><dt><a name="TimeoutException-__getitem__"><strong>__getitem__</strong></a>(...)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="UnableToProcessRTTconffile">class <strong>UnableToProcessRTTconffile</strong></a>(<a href="RTTSException.html#RTTSException">RTTSException</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="RTTSException.html#UnableToProcessRTTconffile">UnableToProcessRTTconffile</a></dd>
-<dd><a href="RTTSException.html#RTTSException">RTTSException</a></dd>
-<dd><a href="exceptions.html#Exception">exceptions.Exception</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="UnableToProcessRTTconffile-__init__"><strong>__init__</strong></a>(self, args<font color="#909090">=None</font>)</dt></dl>
-
-<hr>
-Methods inherited from <a href="RTTSException.html#RTTSException">RTTSException</a>:<br>
-<dl><dt><a name="UnableToProcessRTTconffile-__str__"><strong>__str__</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="exceptions.html#Exception">exceptions.Exception</a>:<br>
-<dl><dt><a name="UnableToProcessRTTconffile-__getitem__"><strong>__getitem__</strong></a>(...)</dt></dl>
-
-</td></tr></table></td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Runner.html b/Tools/RunTimeTester/doc/Runner.html
deleted file mode 100755
index 0fbfa179a8268a561176d1474845d9a5ec2da2be..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Runner.html
+++ /dev/null
@@ -1,17 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Runner</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Runner</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Runner.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Runner.py</a></font></td></tr></table>
-    <p><tt>Short&nbsp;script&nbsp;that&nbsp;is&nbsp;used&nbsp;to&nbsp;start&nbsp;the&nbsp;RTT.<br>
-Could&nbsp;be&nbsp;run&nbsp;from&nbsp;a&nbsp;cron&nbsp;job.<br>
-User&nbsp;specifies&nbsp;the&nbsp;location&nbsp;of&nbsp;the&nbsp;RTT&nbsp;scripts,&nbsp;and&nbsp;the&nbsp;location&nbsp;of&nbsp;the<br>
-two&nbsp;configuration&nbsp;files.</tt></p>
-
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/ScriptWriter.html b/Tools/RunTimeTester/doc/ScriptWriter.html
deleted file mode 100755
index e9e9004423b6d4295f497e476ff5999a67e6d93f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/ScriptWriter.html
+++ /dev/null
@@ -1,482 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module ScriptWriter</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>ScriptWriter</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/ScriptWriter.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/ScriptWriter.py</a></font></td></tr></table>
-    <p><tt><a href="#ScriptWriter">ScriptWriter</a>&nbsp;write&nbsp;is&nbsp;a&nbsp;base&nbsp;class&nbsp;for&nbsp;writing&nbsp;shell&nbsp;scripts&nbsp;to&nbsp;run&nbsp;the&nbsp;jobs.<br>
-The&nbsp;shell&nbsp;scripts&nbsp;are&nbsp;typically&nbsp;written&nbsp;to&nbsp;a&nbsp;run&nbsp;directory,&nbsp;and&nbsp;are&nbsp;submitted&nbsp;to&nbsp;a<br>
-queue&nbsp;from&nbsp;there.<br>
-&nbsp;<br>
-The&nbsp;base&nbsp;class&nbsp;provides&nbsp;a&nbsp;default&nbsp;implementation.</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#PBSTestScriptWriter">PBSTestScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#RecExCommonScriptWriter">RecExCommonScriptWriter</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#JiveXMLScriptWriter">JiveXMLScriptWriter</a>
-</font></dt></dl>
-</dd>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#TrigReleaseIDScriptWriter">TrigReleaseIDScriptWriter</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#TrigReleaseCaloScriptWriter">TrigReleaseCaloScriptWriter</a>
-</font></dt></dl>
-</dd>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#TriggerReleaseScriptWriter">TriggerReleaseScriptWriter</a>
-</font></dt></dl>
-</dd>
-<dt><font face="helvetica, arial"><a href="ScriptWriter.html#KitValidationScriptWriter">KitValidationScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#MonitoringScriptWriter">MonitoringScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#NullScriptWriter">NullScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#QAmetricsScriptWriter">QAmetricsScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#RTTSelfTestScriptWriter">RTTSelfTestScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#RTTTrivialJobScriptWriter">RTTTrivialJobScriptWriter</a>
-</font></dt><dt><font face="helvetica, arial"><a href="ScriptWriter.html#RuleCheckerScriptWriter">RuleCheckerScriptWriter</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="AthenaScriptWriter">class <strong>AthenaScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="AthenaScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt><dd><tt>Provides&nbsp;a&nbsp;run&nbsp;script&nbsp;for&nbsp;Athena&nbsp;jobs</tt></dd></dl>
-
-<dl><dt><a name="AthenaScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>Command&nbsp;for&nbsp;running&nbsp;athena.&nbsp;from&nbsp;about&nbsp;8.2.0,&nbsp;packages&nbsp;started&nbsp;switching&nbsp;to<br>
-using&nbsp;python&nbsp;scripts&nbsp;to&nbsp;run&nbsp;athena.</tt></dd></dl>
-
-<dl><dt><a name="AthenaScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="AthenaScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="AthenaScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt><dd><tt>Default&nbsp;shell&nbsp;script&nbsp;for&nbsp;running&nbsp;a&nbsp;job.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="JiveXMLScriptWriter">class <strong>JiveXMLScriptWriter</strong></a>(<a href="ScriptWriter.html#RecExCommonScriptWriter">RecExCommonScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="ScriptWriter.html#JiveXMLScriptWriter">JiveXMLScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#RecExCommonScriptWriter">RecExCommonScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="JiveXMLScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#RecExCommonScriptWriter">RecExCommonScriptWriter</a>:<br>
-<dl><dt><a name="JiveXMLScriptWriter-G3flag"><strong>G3flag</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>add&nbsp;in&nbsp;flags&nbsp;to&nbsp;run&nbsp;on&nbsp;DC1</tt></dd></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-dataSet"><strong>dataSet</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>:<br>
-<dl><dt><a name="JiveXMLScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="JiveXMLScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="KitValidationScriptWriter">class <strong>KitValidationScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="KitValidationScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt><dd><tt>Obtain&nbsp;information&nbsp;pertaining&nbsp;to&nbsp;the&nbsp;job</tt></dd></dl>
-
-<dl><dt><a name="KitValidationScriptWriter-kitValidationCommand"><strong>kitValidationCommand</strong></a>(self)</dt><dd><tt>Command&nbsp;for&nbsp;running&nbsp;KitValidation.</tt></dd></dl>
-
-<dl><dt><a name="KitValidationScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt><dd><tt>Default&nbsp;shell&nbsp;script&nbsp;for&nbsp;running&nbsp;a&nbsp;job.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="MonitoringScriptWriter">class <strong>MonitoringScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="MonitoringScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="MonitoringScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="NullScriptWriter">class <strong>NullScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="NullScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="NullScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PBSTestScriptWriter">class <strong>PBSTestScriptWriter</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PBSTestScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="PBSTestScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="QAmetricsScriptWriter">class <strong>QAmetricsScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="QAmetricsScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="QAmetricsScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RTTSelfTestScriptWriter">class <strong>RTTSelfTestScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="RTTSelfTestScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="RTTSelfTestScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt><dd><tt>Checkout&nbsp;head&nbsp;version&nbsp;of&nbsp;RTT.&nbsp;This&nbsp;will&nbsp;be&nbsp;shipped&nbsp;back&nbsp;to&nbsp;the<br>
-machine&nbsp;running&nbsp;RTT&nbsp;&nbsp;Scripts,&nbsp;where&nbsp;Test&nbsp;Suite&nbsp;can&nbsp;be&nbsp;run&nbsp;as&nbsp;an<br>
-RTT&nbsp;test.&nbsp;Dont&nbsp;run&nbsp;TestSuite&nbsp;on&nbsp;the&nbsp;batch&nbsp;machine&nbsp;to&nbsp;avoid<br>
-requiring&nbsp;Python&nbsp;on&nbsp;that&nbsp;machine.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RTTTrivialJobScriptWriter">class <strong>RTTTrivialJobScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="RTTTrivialJobScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="RTTTrivialJobScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt><dd><tt>Checkout&nbsp;head&nbsp;version&nbsp;of&nbsp;RTT.&nbsp;This&nbsp;will&nbsp;be&nbsp;shipped&nbsp;back&nbsp;to&nbsp;the<br>
-machine&nbsp;running&nbsp;RTT&nbsp;&nbsp;Scripts,&nbsp;where&nbsp;Test&nbsp;Suite&nbsp;can&nbsp;be&nbsp;run&nbsp;as&nbsp;an<br>
-RTT&nbsp;test.&nbsp;Dont&nbsp;run&nbsp;TestSuite&nbsp;on&nbsp;the&nbsp;batch&nbsp;machine&nbsp;to&nbsp;avoid<br>
-requiring&nbsp;Python&nbsp;on&nbsp;that&nbsp;machine.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RecExCommonScriptWriter">class <strong>RecExCommonScriptWriter</strong></a>(<a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Script&nbsp;to&nbsp;run&nbsp;RecExCommon.<br>
-This&nbsp;needs&nbsp;to&nbsp;perform&nbsp;various&nbsp;RecExCommon&nbsp;tasks.<br>
-Also,&nbsp;for&nbsp;reading&nbsp;DC1&nbsp;data,&nbsp;the&nbsp;ZEBRA&nbsp;.P&nbsp;link,&nbsp;which&nbsp;was&nbsp;set&nbsp;when&nbsp;the<br>
-run&nbsp;directories&nbsp;were&nbsp;setup,&nbsp;needs&nbsp;to&nbsp;be&nbsp;protected&nbsp;because&nbsp;the&nbsp;REC&nbsp;tasks<br>
-will&nbsp;overwrite&nbsp;it.<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="ScriptWriter.html#RecExCommonScriptWriter">RecExCommonScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="RecExCommonScriptWriter-G3flag"><strong>G3flag</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>add&nbsp;in&nbsp;flags&nbsp;to&nbsp;run&nbsp;on&nbsp;DC1</tt></dd></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-dataSet"><strong>dataSet</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>:<br>
-<dl><dt><a name="RecExCommonScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="RecExCommonScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="RuleCheckerScriptWriter">class <strong>RuleCheckerScriptWriter</strong></a>(<a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="RuleCheckerScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="RuleCheckerScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt><dd><tt>CMT&nbsp;stuff&nbsp;needed&nbsp;to&nbsp;run&nbsp;RuleChecker&nbsp;once&nbsp;a&nbsp;directory&nbsp;is&nbsp;checked&nbsp;out.<br>
-Empirically,&nbsp;I&nbsp;find&nbsp;gmake&nbsp;has&nbsp;to&nbsp;be&nbsp;run&nbsp;before&nbsp;gmake&nbsp;rulechecker,<br>
-I&nbsp;think&nbsp;because&nbsp;the&nbsp;include&nbsp;files&nbsp;need&nbsp;to&nbsp;be&nbsp;installed.<br>
-&nbsp;<br>
-Temporarily&nbsp;(14/7/04)&nbsp;I&nbsp;also&nbsp;need&nbsp;to&nbsp;check&nbsp;out&nbsp;CodeCheck.<br>
-Hopefully&nbsp;this&nbsp;will&nbsp;be&nbsp;unecessary&nbsp;soon.</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="ScriptWriter">class <strong>ScriptWriter</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="ScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt><dd><tt>Common&nbsp;stuff&nbsp;for&nbsp;all&nbsp;Scriptwriters&nbsp;-&nbsp;use&nbsp;Descriptor&nbsp;base&nbsp;path</tt></dd></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="TrigReleaseCaloScriptWriter">class <strong>TrigReleaseCaloScriptWriter</strong></a>(<a href="ScriptWriter.html#TrigReleaseIDScriptWriter">TrigReleaseIDScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="ScriptWriter.html#TrigReleaseCaloScriptWriter">TrigReleaseCaloScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#TrigReleaseIDScriptWriter">TrigReleaseIDScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="TrigReleaseCaloScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#TrigReleaseIDScriptWriter">TrigReleaseIDScriptWriter</a>:<br>
-<dl><dt><a name="TrigReleaseCaloScriptWriter-dataSet"><strong>dataSet</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>:<br>
-<dl><dt><a name="TrigReleaseCaloScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>Command&nbsp;for&nbsp;running&nbsp;athena.&nbsp;from&nbsp;about&nbsp;8.2.0,&nbsp;packages&nbsp;started&nbsp;switching&nbsp;to<br>
-using&nbsp;python&nbsp;scripts&nbsp;to&nbsp;run&nbsp;athena.</tt></dd></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="TrigReleaseCaloScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="TrigReleaseIDScriptWriter">class <strong>TrigReleaseIDScriptWriter</strong></a>(<a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="ScriptWriter.html#TrigReleaseIDScriptWriter">TrigReleaseIDScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="TrigReleaseIDScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-dataSet"><strong>dataSet</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>:<br>
-<dl><dt><a name="TrigReleaseIDScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>Command&nbsp;for&nbsp;running&nbsp;athena.&nbsp;from&nbsp;about&nbsp;8.2.0,&nbsp;packages&nbsp;started&nbsp;switching&nbsp;to<br>
-using&nbsp;python&nbsp;scripts&nbsp;to&nbsp;run&nbsp;athena.</tt></dd></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="TrigReleaseIDScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="TriggerReleaseScriptWriter">class <strong>TriggerReleaseScriptWriter</strong></a>(<a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt>Method resolution order:</dt>
-<dd><a href="ScriptWriter.html#TriggerReleaseScriptWriter">TriggerReleaseScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a></dd>
-<dd><a href="ScriptWriter.html#ScriptWriter">ScriptWriter</a></dd>
-</dl>
-<hr>
-Methods defined here:<br>
-<dl><dt><a name="TriggerReleaseScriptWriter-__init__"><strong>__init__</strong></a>(self, jDescriptor)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-dataSet"><strong>dataSet</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-runScript"><strong>runScript</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="ScriptWriter.html#AthenaScriptWriter">AthenaScriptWriter</a>:<br>
-<dl><dt><a name="TriggerReleaseScriptWriter-athenaCommand"><strong>athenaCommand</strong></a>(self, release<font color="#909090">=830</font>)</dt><dd><tt>Command&nbsp;for&nbsp;running&nbsp;athena.&nbsp;from&nbsp;about&nbsp;8.2.0,&nbsp;packages&nbsp;started&nbsp;switching&nbsp;to<br>
-using&nbsp;python&nbsp;scripts&nbsp;to&nbsp;run&nbsp;athena.</tt></dd></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-checkExtension"><strong>checkExtension</strong></a>(self, option)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-commandLineFlags"><strong>commandLineFlags</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-commandLineOptionLabel"><strong>commandLineOptionLabel</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-jobOptionsFile"><strong>jobOptionsFile</strong></a>(self)</dt><dd><tt>Returns&nbsp;the&nbsp;path&nbsp;to&nbsp;the&nbsp;job&nbsp;options&nbsp;file&nbsp;that&nbsp;is&nbsp;to&nbsp;be<br>
-tacked&nbsp;on&nbsp;the&nbsp;end&nbsp;of&nbsp;the&nbsp;athena&nbsp;command.<br>
----&nbsp;If&nbsp;DC1&nbsp;or&nbsp;dataless&nbsp;job:&nbsp;just&nbsp;return&nbsp;the&nbsp;job&nbsp;options&nbsp;passed&nbsp;in&nbsp;&lt;options&gt;&nbsp;tag<br>
----&nbsp;If&nbsp;DC2&nbsp;or&nbsp;BS&nbsp;data:&nbsp;return&nbsp;an&nbsp;RTT&nbsp;created&nbsp;job&nbsp;options&nbsp;that&nbsp;'includes'&nbsp;the<br>
-user&nbsp;one&nbsp;in&nbsp;the&nbsp;&lt;options&gt;&nbsp;tag&nbsp;but&nbsp;then&nbsp;adds&nbsp;the&nbsp;data&nbsp;in&nbsp;the&nbsp;&lt;dataset&gt;&nbsp;tag(s)<br>
-to&nbsp;the&nbsp;appropriate&nbsp;Python&nbsp;class&nbsp;(DC2&nbsp;=&nbsp;EventSelector.InputCollections,<br>
-BS&nbsp;=&nbsp;ByteStreamEventStorageInputSvc.FullFileName).</tt></dd></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-jobOptionsListHandler"><strong>jobOptionsListHandler</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-lsatlF"><strong>lsatlF</strong></a>(self, when)</dt></dl>
-
-<dl><dt><a name="TriggerReleaseScriptWriter-poolInsertString"><strong>poolInsertString</strong></a>(self)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-libLineCalculator"><strong>libLineCalculator</strong></a>(siteSpecifics, mode)</dt><dd><tt>This&nbsp;function&nbsp;calculates&nbsp;the&nbsp;LD_LIBRARY_PATH&nbsp;line.<br>
-This&nbsp;line&nbsp;will&nbsp;be&nbsp;placed&nbsp;in&nbsp;a&nbsp;shell&nbsp;script&nbsp;that&nbsp;will&nbsp;be&nbsp;run&nbsp;using<br>
-Popen4.<br>
-&nbsp;<br>
-The&nbsp;explicit&nbsp;value&nbsp;of&nbsp;LD_LIBRARY_PATH&nbsp;is&nbsp;written&nbsp;out&nbsp;so&nbsp;the&nbsp;file<br>
-can&nbsp;be&nbsp;run&nbsp;by&nbsp;hand&nbsp;under&nbsp;the&nbsp;same&nbsp;conditions&nbsp;that&nbsp;the&nbsp;p4open&nbsp;object<br>
-sees.</tt></dd></dl>
- <dl><dt><a name="-writeToDir"><strong>writeToDir</strong></a>(script)</dt></dl>
-</td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/SiteSpecifics.html b/Tools/RunTimeTester/doc/SiteSpecifics.html
deleted file mode 100755
index 09846d4bce4a6e28c9600094baddb2f04a3d3d19..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/SiteSpecifics.html
+++ /dev/null
@@ -1,165 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module SiteSpecifics</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>SiteSpecifics</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/SiteSpecifics.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/SiteSpecifics.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="SiteSpecifics.html#CERNSiteSpecifics">CERNSiteSpecifics</a>
-</font></dt><dt><font face="helvetica, arial"><a href="SiteSpecifics.html#PetersMacSiteSpecifics">PetersMacSiteSpecifics</a>
-</font></dt><dt><font face="helvetica, arial"><a href="SiteSpecifics.html#UCLslc3SiteSpecifics">UCLslc3SiteSpecifics</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="CERNSiteSpecifics">class <strong>CERNSiteSpecifics</strong></a>(<a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>For&nbsp;running&nbsp;at&nbsp;CERN<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="CERNSiteSpecifics-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-rootCommand"><strong>rootCommand</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-rootSys"><strong>rootSys</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>:<br>
-<dl><dt><a name="CERNSiteSpecifics-addToLibs"><strong>addToLibs</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-postTogether"><strong>postTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-preAthena"><strong>preAthena</strong></a>(self)</dt><dd><tt>Actions&nbsp;required&nbsp;before&nbsp;invoking&nbsp;athena&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-preTogether"><strong>preTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-site"><strong>site</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="CERNSiteSpecifics-togetherCommand"><strong>togetherCommand</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PetersMacSiteSpecifics">class <strong>PetersMacSiteSpecifics</strong></a>(<a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>For&nbsp;running&nbsp;on&nbsp;Peter&nbsp;Sherwood's&nbsp;laptop<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PetersMacSiteSpecifics-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-rootCommand"><strong>rootCommand</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>:<br>
-<dl><dt><a name="PetersMacSiteSpecifics-addToLibs"><strong>addToLibs</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-postTogether"><strong>postTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-preAthena"><strong>preAthena</strong></a>(self)</dt><dd><tt>Actions&nbsp;required&nbsp;before&nbsp;invoking&nbsp;athena&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-preTogether"><strong>preTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-rootSys"><strong>rootSys</strong></a>(self)</dt><dd><tt>How&nbsp;a&nbsp;root&nbsp;env&nbsp;var&nbsp;is&nbsp;called&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-site"><strong>site</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PetersMacSiteSpecifics-togetherCommand"><strong>togetherCommand</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="SiteSpecifics">class <strong>SiteSpecifics</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Base&nbsp;class<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="SiteSpecifics-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SiteSpecifics-addToLibs"><strong>addToLibs</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SiteSpecifics-postTogether"><strong>postTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SiteSpecifics-preAthena"><strong>preAthena</strong></a>(self)</dt><dd><tt>Actions&nbsp;required&nbsp;before&nbsp;invoking&nbsp;athena&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="SiteSpecifics-preTogether"><strong>preTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SiteSpecifics-rootCommand"><strong>rootCommand</strong></a>(self)</dt><dd><tt>How&nbsp;root&nbsp;is&nbsp;called&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="SiteSpecifics-rootSys"><strong>rootSys</strong></a>(self)</dt><dd><tt>How&nbsp;a&nbsp;root&nbsp;env&nbsp;var&nbsp;is&nbsp;called&nbsp;at&nbsp;a&nbsp;specific&nbsp;site.</tt></dd></dl>
-
-<dl><dt><a name="SiteSpecifics-site"><strong>site</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="SiteSpecifics-togetherCommand"><strong>togetherCommand</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="UCLslc3SiteSpecifics">class <strong>UCLslc3SiteSpecifics</strong></a>(<a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>)</font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>Running&nbsp;on&nbsp;SLC3&nbsp;machines&nbsp;at&nbsp;UCL<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="UCLslc3SiteSpecifics-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-addToLibs"><strong>addToLibs</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-preAthena"><strong>preAthena</strong></a>(self)</dt><dd><tt>Set&nbsp;up&nbsp;env&nbsp;variables&nbsp;required&nbsp;by&nbsp;Gaudi,&nbsp;but&nbsp;not&nbsp;usually&nbsp;set&nbsp;at&nbsp;UCL</tt></dd></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-rootCommand"><strong>rootCommand</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-rootSys"><strong>rootSys</strong></a>(self)</dt></dl>
-
-<hr>
-Methods inherited from <a href="SiteSpecifics.html#SiteSpecifics">SiteSpecifics</a>:<br>
-<dl><dt><a name="UCLslc3SiteSpecifics-postTogether"><strong>postTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-preTogether"><strong>preTogether</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-site"><strong>site</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="UCLslc3SiteSpecifics-togetherCommand"><strong>togetherCommand</strong></a>(self)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Status.html b/Tools/RunTimeTester/doc/Status.html
deleted file mode 100755
index 2279e4c541a965c39c29b440bc39902fbfaf49cd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Status.html
+++ /dev/null
@@ -1,100 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Status</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Status</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Status.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Status.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="Status.html#PackageStatus">PackageStatus</a>
-</font></dt><dt><font face="helvetica, arial"><a href="Status.html#Status">Status</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="PackageStatus">class <strong>PackageStatus</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="PackageStatus-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-done"><strong>done</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-error"><strong>error</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-failure"><strong>failure</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-get"><strong>get</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-init"><strong>init</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-notShutdownState"><strong>notShutdownState</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-ongoing"><strong>ongoing</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-packageTimeout"><strong>packageTimeout</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-rttTimeout"><strong>rttTimeout</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="PackageStatus-shutdown"><strong>shutdown</strong></a>(self)</dt></dl>
-
-</td></tr></table> <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Status">class <strong>Status</strong></a></font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Status-__init__"><strong>__init__</strong></a>(self, status)</dt></dl>
-
-<dl><dt><a name="Status-isFailedPrimaryStatus"><strong>isFailedPrimaryStatus</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Status-isFailedSecondaryStatus"><strong>isFailedSecondaryStatus</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Status-primary"><strong>primary</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Status-secondary"><strong>secondary</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Status-set"><strong>set</strong></a>(self, status)</dt></dl>
-
-<dl><dt><a name="Status-setList"><strong>setList</strong></a>(self, statusList)</dt></dl>
-
-<dl><dt><a name="Status-setPrimary"><strong>setPrimary</strong></a>(self, status)</dt></dl>
-
-<dl><dt><a name="Status-setSecondary"><strong>setSecondary</strong></a>(self, status)</dt></dl>
-
-<dl><dt><a name="Status-status"><strong>status</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="Status-time"><strong>time</strong></a>(self)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Tester.html b/Tools/RunTimeTester/doc/Tester.html
deleted file mode 100755
index b31676c61439548cafb917f89a3a55ed923403ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Tester.html
+++ /dev/null
@@ -1,177 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Tester</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Tester</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Tester.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Tester.py</a></font></td></tr></table>
-    <p><tt>A&nbsp;Giant&nbsp;class&nbsp;which&nbsp;does&nbsp;everything&nbsp;in&nbsp;its&nbsp;constructor&nbsp;!??!<br>
-Does&nbsp;roughly&nbsp;the&nbsp;following:<br>
-&nbsp;<br>
-1.&nbsp;Reads&nbsp;the&nbsp;RTT&nbsp;configuration&nbsp;file,&nbsp;which&nbsp;is&nbsp;used&nbsp;to&nbsp;set&nbsp;up&nbsp;a&nbsp;large&nbsp;number<br>
-&nbsp;&nbsp;&nbsp;of&nbsp;paths&nbsp;and&nbsp;other&nbsp;variables<br>
-&nbsp;<br>
-2.&nbsp;Uses&nbsp;factory&nbsp;method&nbsp;to&nbsp;set&nbsp;up&nbsp;a&nbsp;Commander&nbsp;(mode&nbsp;specific&nbsp;commands)<br>
-&nbsp;<br>
-3.&nbsp;Deletes&nbsp;existing&nbsp;&nbsp;run&nbsp;directories<br>
-&nbsp;<br>
-4.&nbsp;Reads&nbsp;the&nbsp;Jobs&nbsp;configuration&nbsp;file&nbsp;-&nbsp;this&nbsp;sets&nbsp;up&nbsp;a&nbsp;list&nbsp;of&nbsp;Job&nbsp;Descriptors<br>
-&nbsp;&nbsp;&nbsp;(one&nbsp;JobDescriptor&nbsp;per&nbsp;job)<br>
-&nbsp;&nbsp;&nbsp;JobDescriptors&nbsp;contain&nbsp;info&nbsp;about&nbsp;a&nbsp;job.&nbsp;They&nbsp;also&nbsp;create&nbsp;a&nbsp;run&nbsp;directory&nbsp;for<br>
-&nbsp;&nbsp;&nbsp;thir&nbsp;job.<br>
-&nbsp;<br>
-5.&nbsp;Submit&nbsp;the&nbsp;jobs.&nbsp;One&nbsp;JobMinder&nbsp;per&nbsp;job&nbsp;is&nbsp;created.&nbsp;JobMinders&nbsp;incoporate&nbsp;the<br>
-&nbsp;&nbsp;&nbsp;JobDescriptor,&nbsp;and&nbsp;then&nbsp;monitor&nbsp;the&nbsp;progress&nbsp;of&nbsp;the&nbsp;job.<br>
-&nbsp;<br>
-6.&nbsp;A&nbsp;poll&nbsp;command&nbsp;is&nbsp;periodically&nbsp;sent&nbsp;to&nbsp;the&nbsp;JobMinders.&nbsp;These&nbsp;check&nbsp;their&nbsp;job&nbsp;status.<br>
-&nbsp;&nbsp;&nbsp;This&nbsp;status&nbsp;is&nbsp;interogated,&nbsp;and&nbsp;html&nbsp;pages&nbsp;are&nbsp;updated&nbsp;with&nbsp;new&nbsp;status&nbsp;info.<br>
-&nbsp;<br>
-7.&nbsp;After&nbsp;all&nbsp;jobs&nbsp;have&nbsp;completed&nbsp;perform&nbsp;cleanp&nbsp;activities</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="HTMLWriter.html">HTMLWriter</a><br>
-<a href="Reporter.html">Reporter</a><br>
-<a href="anydbm.html">anydbm</a><br>
-</td><td width="25%" valign=top><a href="copy.html">copy</a><br>
-<a href="logging.html">logging</a><br>
-<a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="signal.html">signal</a><br>
-<a href="sys.html">sys</a><br>
-<a href="time.html">time</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="__builtin__.html#object">__builtin__.object</a>
-</font></dt><dd>
-<dl>
-<dt><font face="helvetica, arial"><a href="Tester.html#Tester">Tester</a>
-</font></dt></dl>
-</dd>
-</dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="Tester">class <strong>Tester</strong></a>(<a href="__builtin__.html#object">__builtin__.object</a>)</font></td></tr>
-    
-<tr><td bgcolor="#ffc8d8"><tt>&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="Tester-__init__"><strong>__init__</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-badConfFilePath"><strong>badConfFilePath</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-badRTTtag"><strong>badRTTtag</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-basicCleanUp"><strong>basicCleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-blackList"><strong>blackList</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-cleanUp"><strong>cleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-configureMonitorMinders"><strong>configureMonitorMinders</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createDirectoryMaker"><strong>createDirectoryMaker</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createFailureTestRun"><strong>createFailureTestRun</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createGlobalRunTimer"><strong>createGlobalRunTimer</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createInitialWebPages"><strong>createInitialWebPages</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createKits"><strong>createKits</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createLegals"><strong>createLegals</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createNICOS"><strong>createNICOS</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createNoPackagePaths"><strong>createNoPackagePaths</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createRunSummariser"><strong>createRunSummariser</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createStatusLogger"><strong>createStatusLogger</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createTestRun"><strong>createTestRun</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-createTestRuns"><strong>createTestRuns</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-doMonitoringOutput"><strong>doMonitoringOutput</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-doRunSummariserOutput"><strong>doRunSummariserOutput</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-doTimingOutput"><strong>doTimingOutput</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-execute"><strong>execute</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-executeTestRuns"><strong>executeTestRuns</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-finalise"><strong>finalise</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-findNightlyRelease"><strong>findNightlyRelease</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-fullCleanUp"><strong>fullCleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-getTestRunMonitorData"><strong>getTestRunMonitorData</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-getTestRunTimes"><strong>getTestRunTimes</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-inform"><strong>inform</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-initVariables"><strong>initVariables</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-initialise"><strong>initialise</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-makeLog"><strong>makeLog</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-monitorDataToFile"><strong>monitorDataToFile</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-onSignal"><strong>onSignal</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-permissionsCleanUp"><strong>permissionsCleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-readTopLevelConfigFile"><strong>readTopLevelConfigFile</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-rttRunHasTimedOut"><strong>rttRunHasTimedOut</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-statusLogCleanUp"><strong>statusLogCleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-statusLogCleanUpOnly"><strong>statusLogCleanUpOnly</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-thisNightlyRunWasStartedYesterday"><strong>thisNightlyRunWasStartedYesterday</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-updateWebPage2"><strong>updateWebPage2</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-validateInputParams"><strong>validateInputParams</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-webPagesCleanUp"><strong>webPagesCleanUp</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<dl><dt><a name="Tester-xmlFromMonDict"><strong>xmlFromMonDict</strong></a> = traced_f(*a, **k)</dt></dl>
-
-<hr>
-Data and other attributes defined here:<br>
-<dl><dt><strong>__dict__</strong> = &lt;dictproxy object&gt;<dd><tt>dictionary&nbsp;for&nbsp;instance&nbsp;variables&nbsp;(if&nbsp;defined)</tt></dl>
-
-<dl><dt><strong>__weakref__</strong> = &lt;attribute '__weakref__' of 'Tester' objects&gt;<dd><tt>list&nbsp;of&nbsp;weak&nbsp;references&nbsp;to&nbsp;the&nbsp;<a href="__builtin__.html#object">object</a>&nbsp;(if&nbsp;defined)</tt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>globalReporter</strong> = &lt;Reporter.GlobalReporter instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/Timer.html b/Tools/RunTimeTester/doc/Timer.html
deleted file mode 100755
index 4f1f9232fb51ac8a3add668f0275128b5a0ae5b8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/Timer.html
+++ /dev/null
@@ -1,30 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module Timer</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>Timer</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Timer.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/Timer.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="time.html">time</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-timer"><strong>timer</strong></a>()</dt></dl>
-</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/cmtLines.html b/Tools/RunTimeTester/doc/cmtLines.html
deleted file mode 100755
index f3273b100dc53626264596395b6ce6d7a1780f01..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/cmtLines.html
+++ /dev/null
@@ -1,62 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module cmtLines</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>cmtLines</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/cmtLines.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/cmtLines.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-checkConditionsForProjectBuildRunning"><strong>checkConditionsForProjectBuildRunning</strong></a>(desc)</dt><dd><tt>check&nbsp;we&nbsp;have&nbsp;made&nbsp;the&nbsp;project&nbsp;build&nbsp;cmt&nbsp;requirements&nbsp;files<br>
-for&nbsp;now&nbsp;I&nbsp;am&nbsp;going&nbsp;to&nbsp;hardwire&nbsp;a&nbsp;lot&nbsp;of&nbsp;the&nbsp;code.&nbsp;This&nbsp;is&nbsp;less&nbsp;than<br>
-ideal&nbsp;as&nbsp;we&nbsp;are&nbsp;assuming&nbsp;what&nbsp;is&nbsp;done&nbsp;in&nbsp;other&nbsp;parts&nbsp;of&nbsp;the&nbsp;code<br>
-(UserStuffRetrieverProject)<br>
-&nbsp;Need&nbsp;to&nbsp;make&nbsp;an&nbsp;USR&nbsp;in&nbsp;order&nbsp;to&nbsp;refind&nbsp;the&nbsp;cmt&nbsp;requirements<br>
-&nbsp;files...</tt></dd></dl>
- <dl><dt><a name="-cmtLines"><strong>cmtLines</strong></a>(jDescriptor)</dt><dd><tt>Local&nbsp;packages&nbsp;can&nbsp;only&nbsp;be&nbsp;run&nbsp;if<br>
-&nbsp;<br>
-1.&nbsp;mode&nbsp;=&nbsp;LinuxInteractive&nbsp;mode<br>
-2.&nbsp;CMTPATH&nbsp;is&nbsp;set<br>
-&nbsp;<br>
-because&nbsp;the&nbsp;user&nbsp;has&nbsp;to&nbsp;set&nbsp;up&nbsp;use&nbsp;statements,&nbsp;typically&nbsp;in&nbsp;the<br>
-TestRelease&nbsp;requirements&nbsp;file.<br>
-&nbsp;<br>
-Returns&nbsp;lines&nbsp;of&nbsp;script&nbsp;needed&nbsp;to&nbsp;set&nbsp;up&nbsp;cmt&nbsp;if&nbsp;this&nbsp;is&nbsp;being<br>
-done&nbsp;from&nbsp;scratch(eg&nbsp;for&nbsp;nightly&nbsp;running)<br>
-Returns&nbsp;an&nbsp;empty&nbsp;string&nbsp;otherwise<br>
-(for&nbsp;running&nbsp;from&nbsp;checked-out&nbsp;code,&nbsp;where&nbsp;it&nbsp;is<br>
-assumed&nbsp;the&nbsp;user&nbsp;has&nbsp;performed&nbsp;the&nbsp;cmt&nbsp;operations.</tt></dd></dl>
- <dl><dt><a name="-cmtLinesProject"><strong>cmtLinesProject</strong></a>(jDescriptor)</dt><dd><tt>Returns&nbsp;lines&nbsp;of&nbsp;script&nbsp;needed&nbsp;to&nbsp;set&nbsp;up&nbsp;cmt&nbsp;if&nbsp;this&nbsp;is&nbsp;being<br>
-done&nbsp;from&nbsp;scratch(eg&nbsp;for&nbsp;nightly&nbsp;running)<br>
-Returns&nbsp;an&nbsp;empty&nbsp;string&nbsp;otherwise<br>
-(for&nbsp;running&nbsp;from&nbsp;checked-out&nbsp;code,&nbsp;where&nbsp;it&nbsp;is<br>
-assumed&nbsp;the&nbsp;user&nbsp;has&nbsp;performed&nbsp;the&nbsp;cmt&nbsp;operations.</tt></dd></dl>
- <dl><dt><a name="-conditionsForLocalRunning"><strong>conditionsForLocalRunning</strong></a>()</dt></dl>
-</td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/dbgF.html b/Tools/RunTimeTester/doc/dbgF.html
deleted file mode 100755
index 239ecaa0e50ba42c456bc2c7dd4d3df470774727..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/dbgF.html
+++ /dev/null
@@ -1,22 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module dbgF</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>dbgF</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/dbgF.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/dbgF.py</a></font></td></tr></table>
-    <p><tt>Formats&nbsp;and&nbsp;time&nbsp;stamps&nbsp;&nbsp;debug&nbsp;statements</tt></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-dbgF"><strong>dbgF</strong></a>(obj, text)</dt></dl>
-</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/findNightlyRelease.html b/Tools/RunTimeTester/doc/findNightlyRelease.html
deleted file mode 100755
index 011902708e3dd64ac5d6dded50a219932e1a5b79..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/findNightlyRelease.html
+++ /dev/null
@@ -1,47 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module findNightlyRelease</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>findNightlyRelease</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/findNightlyRelease.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/findNightlyRelease.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="time.html">time</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-findNightlyRelease"><strong>findNightlyRelease</strong></a>(paths, buildType)</dt><dd><tt>Finds&nbsp;which&nbsp;tag&nbsp;todays&nbsp;nightly&nbsp;has&nbsp;been&nbsp;given<br>
-(atlrel_0,&nbsp;atlrel_1&nbsp;etc....)&nbsp;and&nbsp;whether&nbsp;it&nbsp;has&nbsp;finished&nbsp;building<br>
-&nbsp;<br>
-Input:&nbsp;string&nbsp;buildType&nbsp;-&nbsp;opt&nbsp;or&nbsp;dbg<br>
-&nbsp;<br>
-Returns:&nbsp;integer&nbsp;finished&nbsp;-&nbsp;1&nbsp;if&nbsp;nightly&nbsp;build&nbsp;is&nbsp;complete&nbsp;else&nbsp;0</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './findNightlyRelease.pyc'<br>
-<strong>__name__</strong> = 'findNightlyRelease'<br>
-<strong>gdbg</strong> = 0</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/index.html b/Tools/RunTimeTester/doc/index.html
deleted file mode 100755
index 0956d2c862ddc6142238475dd953cb551bb9de51..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/index.html
+++ /dev/null
@@ -1,150 +0,0 @@
-
-<html>
-  <head>
-    <title>Index for pydoc generated doc files</title>
-  </head>
-
-  <body>
-    <h3>RunTimeTester-00-00-93</h3>
-    <ul>
-      <li><a href= "./AthenaJobDescriptor.html">AthenaJobDescriptor.html</a></li>
-      <li><a href= "./BaseDirectoryMaker.html">BaseDirectoryMaker.html</a></li>
-      <li><a href= "./BlackLister.html">BlackLister.html</a></li>
-      <li><a href= "./Borg.html">Borg.html</a></li>
-      <li><a href= "./cfg2xml.html">cfg2xml.html</a></li>
-      <li><a href= "./Checkers.html">Checkers.html</a></li>
-      <li><a href= "./childrenToDescendants.html">childrenToDescendants.html</a></li>
-      <li><a href= "./cmtLines.html">cmtLines.html</a></li>
-      <li><a href= "./CMTtools.html">CMTtools.html</a></li>
-      <li><a href= "./Coloriser.html">Coloriser.html</a></li>
-      <li><a href= "./Commander.html">Commander.html</a></li>
-      <li><a href= "./createRunDir.html">createRunDir.html</a></li>
-      <li><a href= "./CVSPackage.html">CVSPackage.html</a></li>
-      <li><a href= "./dataFiles2String.html">dataFiles2String.html</a></li>
-      <li><a href= "./DataSetCatalog.html">DataSetCatalog.html</a></li>
-      <li><a href= "./DBFileRetriever.html">DBFileRetriever.html</a></li>
-      <li><a href= "./dbgFf.html">dbgFf.html</a></li>
-      <li><a href= "./dbgF.html">dbgF.html</a></li>
-      <li><a href= "./DBIntegrityChecker.html">DBIntegrityChecker.html</a></li>
-      <li><a href= "./Defs.html">Defs.html</a></li>
-      <li><a href= "./Descriptor.html">Descriptor.html</a></li>
-      <li><a href= "./DirectoryMaker.html">DirectoryMaker.html</a></li>
-      <li><a href= "./Evaluate.html">Evaluate.html</a></li>
-      <li><a href= "./exc2string.html">exc2string.html</a></li>
-      <li><a href= "./Factory_BaseClass.html">Factory_BaseClass.html</a></li>
-      <li><a href= "./Factory_CmtLines.html">Factory_CmtLines.html</a></li>
-      <li><a href= "./Factory_Commander.html">Factory_Commander.html</a></li>
-      <li><a href= "./Factory_DirectoryMaker.html">Factory_DirectoryMaker.html</a></li>
-      <li><a href= "./Factory_HardCoded.html">Factory_HardCoded.html</a></li>
-      <li><a href= "./Factory_JobDescriptor.html">Factory_JobDescriptor.html</a></li>
-      <li><a href= "./Factory_JobMaker.html">Factory_JobMaker.html</a></li>
-      <li><a href= "./Factory_Legals.html">Factory_Legals.html</a></li>
-      <li><a href= "./Factory_Minder.html">Factory_Minder.html</a></li>
-      <li><a href= "./Factory_PathNameBuilder.html">Factory_PathNameBuilder.html</a></li>
-      <li><a href= "./Factory_RunTimer.html">Factory_RunTimer.html</a></li>
-      <li><a href= "./Factory_ScriptWriter.html">Factory_ScriptWriter.html</a></li>
-      <li><a href= "./Factory_SiteSpecifics.html">Factory_SiteSpecifics.html</a></li>
-      <li><a href= "./Factory_TestRun.html">Factory_TestRun.html</a></li>
-      <li><a href= "./Factory_UserStuffRetriever.html">Factory_UserStuffRetriever.html</a></li>
-      <li><a href= "./FindInstalledDirsMonolithic.html">FindInstalledDirsMonolithic.html</a></li>
-      <li><a href= "./FindInstalledDirsProject.html">FindInstalledDirsProject.html</a></li>
-      <li><a href= "./findInstalledDirs.html">findInstalledDirs.html</a></li>
-      <li><a href= "./findOptionsPaths.html">findOptionsPaths.html</a></li>
-      <li><a href= "./findPackages.html">findPackages.html</a></li>
-      <li><a href= "./findRTThome.html">findRTThome.html</a></li>
-      <li><a href= "./formatCollection.html">formatCollection.html</a></li>
-      <li><a href= "./Geneology.html">Geneology.html</a></li>
-      <li><a href= "./getInheritanceHeirarchy.html">getInheritanceHeirarchy.html</a></li>
-      <li><a href= "./GetUserConfigData.html">GetUserConfigData.html</a></li>
-      <li><a href= "./globalReporterTest.html">globalReporterTest.html</a></li>
-      <li><a href= "./GroupKits.html">GroupKits.html</a></li>
-      <li><a href= "./GroupsStatus.html">GroupsStatus.html</a></li>
-      <li><a href= "./hardCoded.html">hardCoded.html</a></li>
-      <li><a href= "./HardCoded.html">HardCoded.html</a></li>
-      <li><a href= "./HTMLWriter.html">HTMLWriter.html</a></li>
-      <li><a href= "./JobGroupDirectoryMaker.html">JobGroupDirectoryMaker.html</a></li>
-      <li><a href= "./JobGroupKit.html">JobGroupKit.html</a></li>
-      <li><a href= "./JobGroupKits.html">JobGroupKits.html</a></li>
-      <li><a href= "./JobGroupsParser.html">JobGroupsParser.html</a></li>
-      <li><a href= "./JobMaker.html">JobMaker.html</a></li>
-      <li><a href= "./JobOptionsGenerator.html">JobOptionsGenerator.html</a></li>
-      <li><a href= "./JobsXMLReader.html">JobsXMLReader.html</a></li>
-      <li><a href= "./kitFromConfFile.html">kitFromConfFile.html</a></li>
-      <li><a href= "./Launcher.html">Launcher.html</a></li>
-      <li><a href= "./Legals.html">Legals.html</a></li>
-      <li><a href= "./logAnalyser.html">logAnalyser.html</a></li>
-      <li><a href= "./LoggerSetup.html">LoggerSetup.html</a></li>
-      <li><a href= "./MacroRunner.html">MacroRunner.html</a></li>
-      <li><a href= "./MethodTimer.html">MethodTimer.html</a></li>
-      <li><a href= "./MethodTracer.html">MethodTracer.html</a></li>
-      <li><a href= "./MinderDebugger.html">MinderDebugger.html</a></li>
-      <li><a href= "./Minder.html">Minder.html</a></li>
-      <li><a href= "./moduleImportFinder.html">moduleImportFinder.html</a></li>
-      <li><a href= "./ModuleLoader.html">ModuleLoader.html</a></li>
-      <li><a href= "./MonitorJobMinder.html">MonitorJobMinder.html</a></li>
-      <li><a href= "./NameAndStatus.html">NameAndStatus.html</a></li>
-      <li><a href= "./NICOS.html">NICOS.html</a></li>
-      <li><a href= "./Null.html">Null.html</a></li>
-      <li><a href= "./PathCompleter.html">PathCompleter.html</a></li>
-      <li><a href= "./PathNameBuilder.html">PathNameBuilder.html</a></li>
-      <li><a href= "./Paths.html">Paths.html</a></li>
-      <li><a href= "./PostScripts.html">PostScripts.html</a></li>
-      <li><a href= "./qaMetrics2OnePage.html">qaMetrics2OnePage.html</a></li>
-      <li><a href= "./ReleaseJobDescriptor.html">ReleaseJobDescriptor.html</a></li>
-      <li><a href= "./releaseToNum.html">releaseToNum.html</a></li>
-      <li><a href= "./Reporter.html">Reporter.html</a></li>
-      <li><a href= "./requirementsWriter.html">requirementsWriter.html</a></li>
-      <li><a href= "./rmTree.html">rmTree.html</a></li>
-      <li><a href= "./RootMacroRunner.html">RootMacroRunner.html</a></li>
-      <li><a href= "./RTTConfigParser.html">RTTConfigParser.html</a></li>
-      <li><a href= "./RTTDateTime.html">RTTDateTime.html</a></li>
-      <li><a href= "./RttLibraryTools.html">RttLibraryTools.html</a></li>
-      <li><a href= "./RTTpath.html">RTTpath.html</a></li>
-      <li><a href= "./RTTRunMonitor.html">RTTRunMonitor.html</a></li>
-      <li><a href= "./RTTRunSummariser.html">RTTRunSummariser.html</a></li>
-      <li><a href= "./RTTSException.html">RTTSException.html</a></li>
-      <li><a href= "./RTTSummForRoot.html">RTTSummForRoot.html</a></li>
-      <li><a href= "./Runner.html">Runner.html</a></li>
-      <li><a href= "./RunTimer.html">RunTimer.html</a></li>
-      <li><a href= "./ScriptWriter.html">ScriptWriter.html</a></li>
-      <li><a href= "./setupLoggerForDebug.html">setupLoggerForDebug.html</a></li>
-      <li><a href= "./setupLoggerForTestSuite.html">setupLoggerForTestSuite.html</a></li>
-      <li><a href= "./shellCommand.html">shellCommand.html</a></li>
-      <li><a href= "./SiteSpecifics.html">SiteSpecifics.html</a></li>
-      <li><a href= "./srcFileCounter.html">srcFileCounter.html</a></li>
-      <li><a href= "./StampObject.html">StampObject.html</a></li>
-      <li><a href= "./StatusLogger.html">StatusLogger.html</a></li>
-      <li><a href= "./Status.html">Status.html</a></li>
-      <li><a href= "./SuperParser.html">SuperParser.html</a></li>
-      <li><a href= "./Tester.html">Tester.html</a></li>
-      <li><a href= "./TestRun.html">TestRun.html</a></li>
-      <li><a href= "./TestSuiteChecker.html">TestSuiteChecker.html</a></li>
-      <li><a href= "./TestSuiteGlobals.html">TestSuiteGlobals.html</a></li>
-      <li><a href= "./TestSuiteLogging.html">TestSuiteLogging.html</a></li>
-      <li><a href= "./TestSuitePeter.html">TestSuitePeter.html</a></li>
-      <li><a href= "./TestSuiteRefDicts.html">TestSuiteRefDicts.html</a></li>
-      <li><a href= "./TestSuiteRun.html">TestSuiteRun.html</a></li>
-      <li><a href= "./TestSuiteTests.html">TestSuiteTests.html</a></li>
-      <li><a href= "./TextExtractor.html">TextExtractor.html</a></li>
-      <li><a href= "./TimedDataHist.html">TimedDataHist.html</a></li>
-      <li><a href= "./Timer.html">Timer.html</a></li>
-      <li><a href= "./Tools2.html">Tools2.html</a></li>
-      <li><a href= "./Tools.html">Tools.html</a></li>
-      <li><a href= "./ToVariableLoggingHandler.html">ToVariableLoggingHandler.html</a></li>
-      <li><a href= "./UserStuffRetriever2.html">UserStuffRetriever2.html</a></li>
-      <li><a href= "./UserStuffRetrieverProject.html">UserStuffRetrieverProject.html</a></li>
-      <li><a href= "./UserTestDescriptor.html">UserTestDescriptor.html</a></li>
-      <li><a href= "./validateXMLFile.html">validateXMLFile.html</a></li>
-      <li><a href= "./VersionedFileSaver.html">VersionedFileSaver.html</a></li>
-      <li><a href= "./viol2xml.html">viol2xml.html</a></li>
-      <li><a href= "./WatcherJobDescriptor.html">WatcherJobDescriptor.html</a></li>
-      <li><a href= "./WatcherJobMinder.html">WatcherJobMinder.html</a></li>
-      <li><a href= "./WorkerJobDescriptor.html">WorkerJobDescriptor.html</a></li>
-      <li><a href= "./WorkerJobMinder.html">WorkerJobMinder.html</a></li>
-      <li><a href= "./XMLFileSplicer.html">XMLFileSplicer.html</a></li>
-      <li><a href= "./XMLTools.html">XMLTools.html</a></li>
-      <li><a href= "./index.html">index.html</a></li>
-
-    </ul>
-  </body>
-</html>
diff --git a/Tools/RunTimeTester/doc/jobGroupKitFactory.html b/Tools/RunTimeTester/doc/jobGroupKitFactory.html
deleted file mode 100755
index d69ac84f6f84239053b1f9747fc4d6a64b9f4ea4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/jobGroupKitFactory.html
+++ /dev/null
@@ -1,42 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module jobGroupKitFactory</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>jobGroupKitFactory</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/jobGroupKitFactory.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/jobGroupKitFactory.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-jobGroupKitFactory"><strong>jobGroupKitFactory</strong></a>(group)</dt><dd><tt>&nbsp;provide&nbsp;an&nbsp;instance&nbsp;of&nbsp;a&nbsp;JobGroupKit:<br>
-You&nbsp;know&nbsp;the&nbsp;name&nbsp;of&nbsp;a&nbsp;group&nbsp;-&nbsp;and&nbsp;you&nbsp;need&nbsp;the&nbsp;kit&nbsp;to&nbsp;give&nbsp;you<br>
-pieces&nbsp;relevant&nbsp;to&nbsp;that&nbsp;group</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './jobGroupKitFactory.pyc'<br>
-<strong>__name__</strong> = 'jobGroupKitFactory'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/logAnalyser.html b/Tools/RunTimeTester/doc/logAnalyser.html
deleted file mode 100755
index 59f156c79cb8fd9cfebc5efc7b713ab513d3bcac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/logAnalyser.html
+++ /dev/null
@@ -1,58 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module logAnalyser</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>logAnalyser</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/logAnalyser.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/logAnalyser.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ee77aa">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Classes</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#ee77aa"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl>
-<dt><font face="helvetica, arial"><a href="logAnalyser.html#LogAnalyser">LogAnalyser</a>
-</font></dt></dl>
- <p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#ffc8d8">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#000000" face="helvetica, arial"><a name="LogAnalyser">class <strong>LogAnalyser</strong></a></font></td></tr>
-    
-<tr bgcolor="#ffc8d8"><td rowspan=2><tt>&nbsp;&nbsp;&nbsp;</tt></td>
-<td colspan=2><tt>reads&nbsp;in&nbsp;a&nbsp;list&nbsp;of&nbsp;strings,&nbsp;and&nbsp;files&nbsp;for&nbsp;the&nbsp;presence<br>
-of&nbsp;these&nbsp;strings.&nbsp;Write&nbsp;counts&nbsp;to&nbsp;an&nbsp;html&nbsp;file<br>&nbsp;</tt></td></tr>
-<tr><td>&nbsp;</td>
-<td width="100%">Methods defined here:<br>
-<dl><dt><a name="LogAnalyser-__init__"><strong>__init__</strong></a>(self)</dt></dl>
-
-<dl><dt><a name="LogAnalyser-findStrings"><strong>findStrings</strong></a>(self, dir<font color="#909090">='.'</font>)</dt></dl>
-
-<dl><dt><a name="LogAnalyser-printPage"><strong>printPage</strong></a>(self)</dt></dl>
-
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/makeDocIndex.html b/Tools/RunTimeTester/doc/makeDocIndex.html
deleted file mode 100755
index 1b25fcea16d9730a09aa4d1a8ecb66637c939c80..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/makeDocIndex.html
+++ /dev/null
@@ -1,37 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module makeDocIndex</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>makeDocIndex</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/CMT/8.3.0/Tools/RunTimeTester/RunTimeTester-00-00-07/src/makeDocIndex.py">/local_disk/sherwood/CMT/8.3.0/Tools/RunTimeTester/RunTimeTester-00-00-07/src/makeDocIndex.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="string.html">string</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './makeDocIndex.pyc'<br>
-<strong>__name__</strong> = 'makeDocIndex'<br>
-<strong>file</strong> = 'makeDocIndex.pyc'<br>
-<strong>fn</strong> = 'index.html'<br>
-<strong>html</strong> = '<font color="#c040c0">\n</font>    &lt;/ul&gt;<font color="#c040c0">\n</font>  &lt;/body&gt;<font color="#c040c0">\n</font>&lt;/html&gt;<font color="#c040c0">\n</font>'<br>
-<strong>ofile</strong> = &lt;open file 'index.html', mode 'w'&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/makeDocIndex.py b/Tools/RunTimeTester/doc/makeDocIndex.py
deleted file mode 100755
index 56cbd428ed201c6b4988ed98ec7b1be693328d98..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/makeDocIndex.py
+++ /dev/null
@@ -1,43 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-
-import os, os.path, string
-
-fn='index.html'
-if os.path.exists(fn):
-    os.remove(fn)
-
-ofile=open(fn,'w')
-
-
-html="""
-<html>
-  <head>
-    <title>Index for pydoc generated doc files</title>
-  </head>
-
-  <body>
-    <h3>%s</h3>
-    <ul>
-""" % (os.getcwd().split('/'))[-2]
-
-
-ofile.write(html)
-for file in os.listdir("."):
-    print 
-    if os.path.splitext(file)[1] == '.html':
-        ofile.write('      <li><a href= "./'+file+'">'+file+'</a></li>\n')
-
-html="""
-    </ul>
-  </body>
-</html>
-"""
-
-ofile.write(html)
-
-
-
-
-
diff --git a/Tools/RunTimeTester/doc/releaseToNum.html b/Tools/RunTimeTester/doc/releaseToNum.html
deleted file mode 100755
index bf16779891f1dd8f90a17edb19870f1aad7406a7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/releaseToNum.html
+++ /dev/null
@@ -1,23 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module releaseToNum</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>releaseToNum</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/releaseToNum.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/releaseToNum.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-releaseToNum"><strong>releaseToNum</strong></a>(str)</dt><dd><tt>Converts&nbsp;the&nbsp;release&nbsp;string&nbsp;x.y.z&nbsp;to&nbsp;a&nbsp;number<br>
-(eg&nbsp;"8.3.0"-&gt;&nbsp;830)</tt></dd></dl>
-</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/requirementsWriter.html b/Tools/RunTimeTester/doc/requirementsWriter.html
deleted file mode 100755
index 641c9c9b02823ed5c8bccbb16827cda901e6f6e5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/requirementsWriter.html
+++ /dev/null
@@ -1,42 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module requirementsWriter</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>requirementsWriter</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/requirementsWriter.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/requirementsWriter.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="Reporter.html">Reporter</a><br>
-</td><td width="25%" valign=top><a href="logging.html">logging</a><br>
-</td><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="string.html">string</a><br>
-</td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-requirementsWriter"><strong>requirementsWriter</strong></a>(distArea, release, runPath, runType, packages<font color="#909090">=[]</font>)</dt><dd><tt>returns&nbsp;the&nbsp;'home&nbsp;directory'&nbsp;requirements&nbsp;file</tt></dd></dl>
- <dl><dt><a name="-requirementsWriter_Project"><strong>requirementsWriter_Project</strong></a>(runPath, packages<font color="#909090">=[]</font>)</dt><dd><tt>returns&nbsp;the&nbsp;'home&nbsp;directory'&nbsp;requirements&nbsp;file</tt></dd></dl>
-</td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><strong>globalReporter</strong> = &lt;Reporter.GlobalReporter instance&gt;<br>
-<strong>logger</strong> = &lt;logging.Logger instance&gt;</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/rmTree.html b/Tools/RunTimeTester/doc/rmTree.html
deleted file mode 100755
index 1fcbc8d22a7f25218eb6a2396931835c4348c603..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/rmTree.html
+++ /dev/null
@@ -1,31 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module rmTree</title>
-</head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>rmTree</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/rmTree.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-93/src/rmTree.py</a></font></td></tr></table>
-    <p></p>
-<p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top><a href="sys.html">sys</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table><p>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;</td>
-<td width="100%"><dl><dt><a name="-rmTree"><strong>rmTree</strong></a>(top)</dt></dl>
-</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/scriptWriterFactory.html b/Tools/RunTimeTester/doc/scriptWriterFactory.html
deleted file mode 100755
index 9debaf045436151d936a4124e5ec0db606f0cecf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/scriptWriterFactory.html
+++ /dev/null
@@ -1,34 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module scriptWriterFactory</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>scriptWriterFactory</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/scriptWriterFactory.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/scriptWriterFactory.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-scriptWriterFactory"><strong>scriptWriterFactory</strong></a>(jDescriptor)</dt><dd><tt>&nbsp;provide&nbsp;an&nbsp;instance&nbsp;of&nbsp;a&nbsp;ScriptWriter:<br>
-You&nbsp;give&nbsp;the&nbsp;job&nbsp;group&nbsp;-&nbsp;and&nbsp;you&nbsp;need&nbsp;the&nbsp;kit&nbsp;to&nbsp;give&nbsp;you<br>
-pieces&nbsp;relevant&nbsp;to&nbsp;that&nbsp;group</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './scriptWriterFactory.pyc'<br>
-<strong>__name__</strong> = 'scriptWriterFactory'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/doc/siteSpecificsFactory.html b/Tools/RunTimeTester/doc/siteSpecificsFactory.html
deleted file mode 100755
index a69fb7e04838f2f8c26ee28867dec7a2086e0a5d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/doc/siteSpecificsFactory.html
+++ /dev/null
@@ -1,42 +0,0 @@
-
-<!doctype html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
-<html><head><title>Python: module siteSpecificsFactory</title>
-<style type="text/css"><!--
-TT { font-family: lucidatypewriter, lucida console, courier }
---></style></head><body bgcolor="#f0f0f8">
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="heading">
-<tr bgcolor="#7799ee">
-<td valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial">&nbsp;<br><big><big><strong>siteSpecificsFactory</strong></big></big></font></td
-><td align=right valign=bottom
-><font color="#ffffff" face="helvetica, arial"><a href=".">index</a><br><a href="file:/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/siteSpecificsFactory.py">/local_disk/sherwood/RTT/Tools/RunTimeTester/RunTimeTester-00-00-10/src/siteSpecificsFactory.py</a></font></td></tr></table>
-    <p></p>
-
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#aa55cc">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#fffff" face="helvetica, arial"><big><strong>Modules</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#aa55cc"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><table width="100%" summary="list"><tr><td width="25%" valign=top><a href="os.html">os</a><br>
-</td><td width="25%" valign=top></td><td width="25%" valign=top></td><td width="25%" valign=top></td></tr></table></td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#eeaa77">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Functions</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#eeaa77"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><dl><dt><a name="-siteSpecificsFactory"><strong>siteSpecificsFactory</strong></a>(site)</dt><dd><tt>&nbsp;provide&nbsp;an&nbsp;instance&nbsp;of&nbsp;a&nbsp;SiteSpecifics:<br>
-You&nbsp;give&nbsp;the&nbsp;site&nbsp;of&nbsp;a&nbsp;group&nbsp;-&nbsp;and&nbsp;you&nbsp;get&nbsp;an&nbsp;object&nbsp;with<br>
-details&nbsp;for&nbsp;that&nbsp;site.</tt></dd></dl>
-</td></tr></table>
-<table width="100%" cellspacing=0 cellpadding=2 border=0 summary="section">
-<tr bgcolor="#55aa55">
-<td colspan=3 valign=bottom>&nbsp;<br>
-<font color="#ffffff" face="helvetica, arial"><big><strong>Data</strong></big></font></td></tr>
-    
-<tr><td bgcolor="#55aa55"><tt>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</tt></td><td>&nbsp;&nbsp;</td>
-<td width="100%"><strong>__file__</strong> = './siteSpecificsFactory.pyc'<br>
-<strong>__name__</strong> = 'siteSpecificsFactory'</td></tr></table>
-</body></html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/get_release_2017.py b/Tools/RunTimeTester/get_release_2017.py
deleted file mode 100644
index 57a4743fb0a8ef33d5539f1248b31fbd0dd8568e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/get_release_2017.py
+++ /dev/null
@@ -1,195 +0,0 @@
-"""get_release: returns the path to a build provided that such a build
-exists, and that it is recent."""
-
-import os
-import re
-from time import sleep
-from datetime import (datetime,
-                      time,
-                      timedelta)
-import PathConstants
-
-fn_re = re.compile(r'^(?P<year>20\d\d)-(?P<month>\d\d)-(?P<day>\d\d)T\d\d\d\d$')
-build_day_start = 20
-
-def fn2dt(fn): return datetime.strptime(fn, "%Y-%m-%dT%H%M")
-
-def fn2asetup_date(fn):
-    m = fn_re.match(fn)
-    return 'r%s-%s-%s' % (m.group('year'), m.group('month'), m.group('day'))
-             
-def _get_build_in_period(branch, period_start, period_end):
-    """
-    period_end - end of build persiod under conssideration:
-                 either build_day_start for today or for tomorrow
-                 depending on the start time of the RTT.
-
-    full dir eg 
-    /cvmfs/atlas-nightlies.cern.ch/repo/sw/master/2017-03-09T2245
-    """
-
-    # find if any branchs are recent
-    basedir = PathConstants.branchdir_2017
-    found = [fn for fn in os.listdir(os.path.join(basedir, branch))
-             if fn_re.match(fn)]
-
-    def is_recent(fn):
-        fn_dt = fn2dt(fn)
-        return  fn_dt >= period_start and fn_dt < period_end 
-
-
-    # print found
-    recent = [f for f in found if is_recent(f)]
-    if not recent:
-        print 'no file available for period %s - %s in %s' % (
-            str(period_start),
-            str(period_end),
-            os.path.join(PathConstants.branchdir_2017, branch)
-        )
-        return ''
-    # examine build for matching cmt config paramters
-    # start search form the most recent, to back in time.
-
-    print 'recent', recent
-    recent = sorted(
-        recent,
-        cmp=lambda x,y: cmp(fn2dt(x), fn2dt(y)),
-        reverse=True)
-    return recent[0]
-
-
-def _get_release(release_datetime):
-    """return the release (rel_X) for a datetime within the current
-    build persiod. The current build period starts at the build_start_time
-    of the same day if the rtt starts up at a time after the build_start_time,
-    otherwise the build_start_time of the previous day"""
-
-    
-    # python calls Sunday 6, Monday 0, ATLAS: Sunday 0, Monday 1
-    rel = release_datetime.weekday() + 1
-    if rel == 7: rel = 0
-
-    # but if the build starts after build_day_start, consider
-    # it to be the next day
-
-    # so, a build starting after build_day_start on sunday
-    # will be assigned 6 by python, converted to 0 above.
-    # further it will be reassigned to 1 as it starts after
-    # build_day_start
-    if release_datetime.hour >= build_day_start: rel += 1
-    if rel == 7: rel = 0
-
-    return 'rel_%d' % rel
-
-def get_period_start(dt):
-    
-    result = dt.replace(
-        hour=build_day_start,
-        minute=0,
-        second=0,
-        microsecond=0)
-
-    if dt.hour < build_day_start:
-        result = result - timedelta(days=1)
-
-    return result
-
-
-def get_release(branch, cmtconfig, check_datetime=None):
-    """Return a release (string matching rel_[0,6]) if a build occurs
-    during a period of interest, '' otherwise.
-    If the RTT check time is after build_day_start, the period of
-    interest starts at the same day at build_day_start, otherwise
-    at build_day_start of the previous day. It terminates 24 hours later.
-
-    If no such build exists, the function waits for it to arrive, or
-    the current time exits the persiod of interest.
-    """
-    if check_datetime is None: check_datetime = datetime.now()
-    
-    period_start = get_period_start(check_datetime)
-    print period_start
-    period_end = period_start + timedelta(days=1)
-
-    while True:
-        fn  = _get_build_in_period(branch,
-                                   period_start,
-                                   period_end)
-        if not fn:
-            print 'get_release_2017 no file available, sleeping'
-            sleep(60*5) # no file available, sleep 5 mins
-        else:
-            return (fn, _get_release(fn2dt(fn)), fn2asetup_date(fn))
-    
-if __name__ == '__main__':
-    import sys
-    branch = sys.argv[1]
-    cmtconfig = sys.argv[2]
-
-    # eg python get_release_2017.py master x86_64-slc6-gcc49-opt
-
-
-
-    print 'calling _get_build_in_period', branch, cmtconfig
-    period_start = get_period_start(datetime.now())
-    period_end = period_start + timedelta(days=1)
-    
-    fn = _get_build_in_period(branch, period_start, period_end)
-    print 'release file: -->%s<--' % fn
-
-    # the call to get release waits for a suitable build to appear,
-    # so may hang.
-    # fn, rel = get_release(datetime.now(), branch, cmtconfig)
-    # print 'ran get_release', branch, cmtconfig
-    # print fn, rel
-
-    print 'calling _get_release'
-
-    early = build_day_start - 1
-    late = build_day_start + 1
-
-    assert early in range(24)
-    assert late in range(24)
-    
-    sunday_early = datetime(year=2017,
-                              month=3,
-                              day=12,
-                              hour=early)
-
-    sunday_late = sunday_early.replace(hour=late)
-
-    assert _get_release(sunday_early) == 'rel_0'
-    assert _get_release(sunday_late) == 'rel_1'
-
-
-    print '\ntemp debug\n'
-    now = datetime.now()
-    now_1 = now - timedelta(days=1)
-    now_2 = now - timedelta(days=2)
-    now_3 = now - timedelta(days=3)
-
-    conf = 'x86_64-slc6-gcc62-opt'
-    branch = 'master'
-
-    print branch, conf
-    print get_release(branch, conf)
-    print get_release(branch, conf, now)
-    print get_release(branch, conf, now_1)
-    print get_release(branch, conf, now_2)
-    print get_release(branch, conf, now_3)
-
-
-    branch = '21.0'
-    print 
-    print branch, conf
-    print
-    print get_release(branch, conf)
-    print get_release(branch, conf, now)
-    print get_release(branch, conf, now_1)
-    print get_release(branch, conf, now_2)
-    print get_release(branch, conf, now_3)
-
-    print 'ok'
-
-    
-    
diff --git a/Tools/RunTimeTester/releaseTools/AthenaLogStrings.dat b/Tools/RunTimeTester/releaseTools/AthenaLogStrings.dat
deleted file mode 100755
index e528924d7b6c7f2d3f7ddb67ff8ff9d97d49a9a7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/AthenaLogStrings.dat
+++ /dev/null
@@ -1,6 +0,0 @@
-/afs/cern.ch/atlas/project/RTT/Work/rel_3/dev/build/i686-slc3-gcc323-opt
-*_log
-RDBAccessSvc::connect()                   ERROR Exception caught:
-Could not load module GaudiAud
-dbConnection is not correctly initialized.
-exception
diff --git a/Tools/RunTimeTester/releaseTools/BadJobsFromRTTSummary.py b/Tools/RunTimeTester/releaseTools/BadJobsFromRTTSummary.py
deleted file mode 100755
index 94bd61f6788cdbe73d54592ce0675117cb013ac4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/BadJobsFromRTTSummary.py
+++ /dev/null
@@ -1,123 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path, string, time, xml.dom.minidom, sys, getopt
-from getRTTSummaryAsDOM import getRTTSummaryAsDOM
-sys.path.append('../src')
-from Evaluate import Evaluate
-
-def badJobsFromRTTSummary(dom):
-    
-    de = dom.documentElement
-
-    packages = de.getElementsByTagName('package')
-    npackages = 0
-    nbadjobs  = 0
-    badpackages = {}
-    for p in packages:
-        minders = p.getElementsByTagName('minder')
-        minders = [m for m in minders if len(Evaluate('errors/text()',m))>0]
-        if not minders: continue
-        packageName = Evaluate('packageName/text()', p)
-        
-        mdata = [(Evaluate('identifiedName/text()', m), Evaluate('errors/text()', m)) for m in minders]
-
-        npackages += 1
-        nbadjobs += len(mdata)
-        badpackages[packageName] = len(mdata)
-        
-        print
-        print 'package -------------------', packageName 
-        print
-        for d in mdata: print '   minder: %s\n      %s', d
-
-
-    sortedItems = badpackages.items()
-    sortedItems.sort()
-
-    sortedItems = [(i.strip(), j) for (i,j) in sortedItems]
-    print
-    print 'No of packages with bad jobs:', npackages
-    print 'No of bad jobs:              ', nbadjobs
-    print
-    print 'No of bad jobs by package:   '
-    print
-    for i in sortedItems:
-        print '%20s %10d' % i
-
-def usage():
-    print
-    print 'BadJobsFromRTTSummary -- checks for jobs with error messages in RTTSummary'
-    print
-    print 'usage: python BadJobsFromRTTSummary [options]'
-    print
-    print 'Options'
-    print '   -b branch - one of ["dev", "bugfix", "val"]'
-    print '   -c cmtconfig - one of ["i686-slc4-gcc34-opt", "i686-slc3-gcc323-opt"]'
-    print '   -r release - one of [0-6]'
-    print '   -h print this message and exit'
-    return
-    
-        
-
-if __name__ == '__main__':
-
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'b:c:r:h')
-        print opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    branch    = None
-    config    = None
-    release   = None
-
-    configDict = {'slc4': 'i686-slc4-gcc34-opt',
-                  'slc3': 'i686-slc3-gcc323-opt'}
-
-    print opts 
-    for o, a in opts:
-        if o == '-b':
-            
-            branch = a
-            if branch not in ['bugfix', 'dev', 'val']:
-
-                print 'bad branch:', b
-                
-                usage()
-                sys.exit(2)
-                
-        if o == '-r':
-            
-            release = a
-            if release not in ['0', '1', '2', '3', '4', '5', '6']:
-
-                print 'bad release:', release
-                
-                usage()
-                sys.exit(2)
-                
-        if o == '-c':
-            
-            config = a
-            if config not in configDict:
-
-                print 'bad cmtconfig:', config
-                
-                usage()
-                sys.exit(2)
-            config = configDict[config]
-
-    
-    import sys
-    # if sys.argv[2] == 'slc4' : conf = 'i686-slc4-gcc34-opt'
-    # if sys.argv[2] == 'slc3' : conf = 'i686-slc3-gcc323-opt'
-    # root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/dev/build/i686-slc3-gcc323-opt' % sys.argv[1]
-    root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/%s/build/%s' % (release, branch, config)
-    fn = os.path.join(root, 'RTTSummary.xml')
-    print fn
-    dom = xml.dom.minidom.parse(fn)
-    print 'parsed ',fn
-
-    
-    print badJobsFromRTTSummary(dom)
diff --git a/Tools/RunTimeTester/releaseTools/ExceptionStrings.dat b/Tools/RunTimeTester/releaseTools/ExceptionStrings.dat
deleted file mode 100755
index 72fabf7a02eca5aaa72947752d15468fb977222e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/ExceptionStrings.dat
+++ /dev/null
@@ -1,88 +0,0 @@
-/afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_3sep_10:15
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC3/logs/logfiles_29aug_22:30
-# /afs/cern.ch/atlas/project/RTT/prod/logs/pcache/SLC3/logs/logfiles_29aug_23:01
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_28aug_09:00
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_29aug_03:31
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_27aug_09:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_23aug_12:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_23aug_09:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_19aug_09:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_19aug_12:01
-# /afs/cern.ch/atlas/project/RTT/prod/logs/pcache/SLC3/logs/logfiles_15aug_23:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_15aug_12:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/pcache/SLC3/logs/logfiles_9aug_22:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_8aug_13:02
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_1aug_02:01
-#/afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_5aug_09:31
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_3aug_03:30
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_2aug_14:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_3aug_09:30
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_1aug_02:01
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_23jul_12:00
-#/afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_23jul_09:31
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_24jul_03:31
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_23jul_16:50
-# /afs/cern.ch/atlas/project/RTT/prod/logs/pcache/SLC3/logs/logfiles_12jul_22:01
-# /afs/cern.ch/atlas/project/RTT/prod/logs/dev/SLC4_32/logs/logfiles_12jul_16:00
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_12jul_13:30
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_12jul_13:30
-# /afs/cern.ch/atlas/project/RTT/prod/logs/val/SLC4_32/logs/logfiles_12jul_11:21
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_10jul_07:35
-#/afs/cern.ch/atlas/project/RTT/data/peter/Results/
-#/afs/cern.ch/atlas/project/RTT/data/peter/Results/rel_5/pcache/build/i686-slc3-gcc323-opt/offline
-#/afs/cern.ch/atlas/project/RTT/Results/rel_2/bugfix/build/i686-slc4-gcc34-opt/offline
-#/afs/cern.ch/atlas/project/RTT/data/peter/logs/SLC4_32/logs/logfiles_22jun_16:46
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_23apr_14:01
-# /afs/cern.ch/atlas/project/RTT/prod/logs/bugfix/SLC4_32/logs/logfiles_23apr_14:01
-# /afs/cern.ch/atlas/project/RTT/data/brinick/logs/SLC4_32/logs/logfiles_28feb_22:30
-#/afs/cern.ch/atlas/project/RTT/prod/SLC3/logs/logfiles_23feb_14:51
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_22feb_19:27
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_22feb_21:19
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_15feb_20:45
-RTT.log;*.log
-#*_log
-RTTSException
-SystemExit
-StopIteration
-StandardError
-KeyboardInterrupt
-ImportError
-EnvironmentError
-IOError
-OSError
-WindowsError
-EOFError
-RuntimeError
-NotImplementedError
-NameError
-UnboundLocalError
-AttributeError
-SyntaxError
-IndentationError
-TabError
-TypeError
-AssertionError
-LookupError
-IndexError
-KeyError
-ArithmeticError
-OverflowError
-ZeroDivisionError
-FloatingPointError
-ValueError
-UnicodeError
-UnicodeEncodeError
-UnicodeDecodeError
-UnicodeTranslateError
-ReferenceError
-SystemError
-MemoryError
- Warning.
-UserWarning
-DeprecationWarning
-PendingDeprecationWarning
-SyntaxWarning
-OverflowWarning
-RuntimeWarning
-FutureWarning
-
diff --git a/Tools/RunTimeTester/releaseTools/KitsInstaller.py b/Tools/RunTimeTester/releaseTools/KitsInstaller.py
deleted file mode 100755
index 1e5cb122b7ed63c908406d1bf5c56e20f532890b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/KitsInstaller.py
+++ /dev/null
@@ -1,246 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#This scripts installs pacman releases
-#Please verify that you are using the recommended
-#pacman version as this may affect the installation
-#process.
-
-# Configuration
-# =============
-# The Script takes a configuration file
-# as command line argument. Here are the
-# contents of a sample config
-#*******************************************************
-# pacmanArea = /unix/atlas1/software/pacman/pacman-3.19
-# release    = rel_3
-# dest       = /unix/atlasrtt/eric
-# branch     = bugfix
-# cmtconfig  = i686-slc4-gcc34-opt
-# todayNightly = True
-#*******************************************************
-#
-
-from popen2 import Popen4
-import os, os.path, time, shutil,sys                
-
-def writeToDir(script, name):
-    #name = os.path.join(os.getcwd(), "dev_kitInstaller.sh")
-    print 'Writing script to run dir'
-    open(name ,"w").write(script)
-    os.chmod(name, 0777)
-    
-
-class ReleaseInstaller:
-    def __init__(self, dest, pacmanArea, branch='dev', cmtconfig='i686-slc3-gcc323-opt', release=''):
-        self.release        = release
-        self.relInstallDest = dest
-        self.distBaseDir    = dest
-        self.pacmanArea     = pacmanArea
-        self.branch         = branch
-        self.cmtconfig      = cmtconfig
-        self.name           = os.path.join(os.getcwd(), "temp_kitInstaller.sh")
-        self.nightlies      = ['rel_0','rel_1','rel_2', 'rel_3',
-                               'rel_4', 'rel_5', 'rel_6']
-        
-    def install(self):
-        print "making run script..."
-        self.makeInstallScript()
-        #print "Setting up pacman..."
-        #setupPacman = Popen4(self.makePacmanCommand())
-        #print str(setupPacman.fromchild.readline())
-        #setupPacman.wait()
-        print "installing kit :"+str(self.release)
-        #install = Popen4(self.makeReleaseInstallCommand())
-        #print str(install.fromchild.readline())
-        #install.wait()
-        #os.system('source kitInstaller.sh')
-        #install = Popen4('source relInstaller.sh')
-        #print str(install.fromchild.readline())
-
-    def installNightlyKit(self):
-        print 'Beginning to install nightly kit'
-        cwd = os.getcwd()
-        print "making run script"
-        self.makeNightlyKitInstallScript()
-        print " nightly kit to install "+self.release
-        time.sleep(30)
-        if not os.path.exists(self.name):
-            print "install shell script not found, exiting ..."
-            sys.exit(0)
-        else:
-            print 'Found install script in %s'%str(cwd)
-            os.system(self.name)
-        
-        
-    def makePacmanCommand(self):
-        print 'making pacman set up command'
-        cwd = os.getcwd()
-        pmc ='cd '+self.pacmanArea+';'
-        pmc +='source setup.sh;'
-        pmc +='cd '+cwd
-        #pmc +='pacman -allow trust-all-caches tar-overwrite'
-        print 'Pacman setup command...'
-        print pmc
-        return pmc
-
-    def makeNightlyKitInstallCommand(self):
-        cwd = os.getcwd()
-        ric = 'cd '+self.relInstallDest+';'
-        ric += 'pacman -allow trust-all-caches tar-overwrite;'
-        ric += 'pacman -get http://cern.ch/atlas-computing'
-        ric += '/links/buildDirectory/kitrel/nightlies/'+self.branch+'/cache:'
-        ric += 'AtlasProduction_'
-        ric += self.makeRelease(self.release)
-        ric += '_'+self.cmtconfig.replace('-','_')+';'
-        ric += 'cd '+cwd
-        print 'Pacman kit install command...'
-        print ric
-        return ric
-        
-    def makeRelease(self, release):
-        if  release in self.nightlies:
-            return release
-        print release
-        parts = release.split('.')
-        rel =  parts[0]
-        rel += '_'
-        rel += parts[1]
-        rel += '_'
-        rel += parts[2]
-
-        return rel
-
-    def makeInstallScript(self):
-        #fh = open('runner.sh', 'w')
-        
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'source $HOME/.bash_profile\n'
-        pmc = self.makePacmanCommand()
-        parts = pmc.split(';')
-        for part in parts:script += part+'\n'
-
-        ric = self.makeReleaseInstallCommand()
-        parts = ric.split(';')
-        for part in parts:script += part+'\n'
-        
-        writeToDir(script, self.name)
-
-    def makeNightlyKitInstallScript(self):
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'source $HOME/.bash_profile\n'
-        script +='cd '+self.distBaseDir+'\n'
-        #script += 'rm -rf cmtsite\n'
-        pmc = self.makePacmanCommand()
-        parts = pmc.split(';')
-        for part in parts:script += part+'\n'
-
-        ric = self.makeNightlyKitInstallCommand()
-        parts = ric.split(';')
-        for part in parts:script += part+'\n'
-        
-        script +='cd '+self.distBaseDir+'\n'
-        script += 'rm -f kitInstallComplete\n'
-        script += 'echo " ">kitInstallComplete\n'
-        script += 'rm -f latest\n'
-        script += 'ln -s '+self.release +' latest\n'
-        #script += '/usr/local/bin/python2.3 /unix/atlasrtt/RTT/sendBrinickMail.py'
-        writeToDir(script, self.name)
-
-    def makeReleaseInstallCommand(self):
-        ric = 'cd '+self.relInstallDest+';'
-        ric += 'pacman -allow trust-all-caches tar-overwrite;'
-        ric += 'pacman -get http://cern.ch/atlas-computing'
-        ric += '/links/kitsDirectory/projects/cache:'
-        ric += 'AtlasProduction_'
-        ric += self.makeRelease(self.release)
-        #ric += '_i686_slc3_gcc323_opt'
-        ric +='_'+self.cmtconfig.replace('-', '_')
-        print 'Pacman kit install command...'
-        print ric
-        return ric
-     
-    def findKitToInstall(self):
-        print 'looking for release to install'
-        lcr  ='/afs/cern.ch/atlas/software/builds/kitrel/nightlies/'+self.branch+'/latest'
-        self.release= os.readlink(lcr)
-        print 'nightly Tag is '+str(os.readlink(lcr))
-        return self.release
-    
-    def makeInstallDir(self):
-        cwd = os.getcwd()
-        #rel = self.findKitToInstall()
-        rel = self.release
-        os.chdir(self.relInstallDest)
-        self.dest = os.path.join(self.relInstallDest, rel)
-        self.relInstallDest= self.dest
-        if os.path.exists(self.dest):
-            print 'removing '+rel+' from '+self.relInstallDest
-            shutil.rmtree(rel)
-        os.mkdir(rel)
-        if os.path.exists(self.dest):
-            print'install directory made: ' +self.dest
-
-        os.chdir(cwd)
-        
-    def stampsForTest(self):
-        print 'looking for nightly stamp'
-        stampFile  = '/afs/cern.ch/atlas/software'
-        stampFile += '/builds/kitrel/nightlies/dev/stamp_for_tests_'+self.cmtconfig
-        self.stampDate  = time.gmtime(os.stat(stampFile).st_atime)[:3]
-        self.today = time.gmtime()[:3]
-
-if __name__ =='__main__':
-    dict = {}
-    if len(sys.argv) < 2:
-        print 'Usage: KitsInstaller.py <config>'
-        sys.exit(1)
-    lines = open(sys.argv[1], 'r').readlines()
-    lines = [line.strip() for line in lines]
-    lines = [line for line in lines if len(line)>0]
-    lines = [line for line in lines if not line.startswith('#')]
-    for line in lines:
-        parts = line.split('=')
-        dict[parts[0].strip()] = parts[1].strip()
-
-    print 'Configuration parameters'
-    for key in dict.keys():
-        print key +' = '+dict[key]
-
-    pacmanArea   = dict['pacmanArea']
-    release      = dict['release']
-    dest         = dict['dest']
-    cmtconfig    = dict['cmtconfig']
-    branch       = dict['branch']
-    todayNightly = dict['todayNightly']
-    
-    RI = ReleaseInstaller(dest, pacmanArea, branch, cmtconfig)
-    RI.stampsForTest()
-    if todayNightly == 'True':
-        while RI.stampDate != RI.today:
-            print "Today\'s stamp not available, will check again in 5 minutes"
-            time.sleep(300)
-            RI.stampsForTest()
-        
-        rel = RI.findKitToInstall()
-        print 'Will now install latest nightly :%s' %rel
-    else:
-        RI.release = release
-        print 'Installing the nightly from config'
-    #print "Found nightly stamp, kit to install is: " +rel
-    #print 'Using hard coded release %s'%release
-    #RI.release = release
-    RI.makeInstallDir()
-    if RI.release in RI.nightlies:
-        print 'now installing...'
-        RI.installNightlyKit()
-        time.sleep(30)
-        #os.system('/home/rtt/kitInstaller.sh')
-        #run=Popen4('source /home/rtt/kitInstaller.sh')
-        #run.wait()
-        #print 'installation done...'
-    else:
-        print 'found '+RI.release
-        RI.install()
-        
diff --git a/Tools/RunTimeTester/releaseTools/PeterGrepStrings.dat b/Tools/RunTimeTester/releaseTools/PeterGrepStrings.dat
deleted file mode 100755
index a9166bcc92a5c3efbf0b5076f7959dfa73feb8c5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/PeterGrepStrings.dat
+++ /dev/null
@@ -1,6 +0,0 @@
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_15feb_20:45
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_16feb_00:35
-/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_16feb_17:00
-*error.log
-#found in
-Memory usage
diff --git a/Tools/RunTimeTester/releaseTools/PeterLogStrings.dat b/Tools/RunTimeTester/releaseTools/PeterLogStrings.dat
deleted file mode 100755
index 9f46dbc20cf16885e275c601eb8ded934c35a478..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/PeterLogStrings.dat
+++ /dev/null
@@ -1,11 +0,0 @@
-/afs/cern.ch/atlas/project/RTT/prod/SLC3/logs/logfiles_20feb_15:43
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_15feb_20:45
-#RTT.log;*.log
-*
-exceptions.MemoryError
-MemFree:
-leaving TestRun run
-operator close
-transition detected
-exception
-Fatal error
\ No newline at end of file
diff --git a/Tools/RunTimeTester/releaseTools/RTTLogStrings.dat b/Tools/RunTimeTester/releaseTools/RTTLogStrings.dat
deleted file mode 100755
index bc7bc5c0a53ee22dfb89b52ebe29eb02b0166287..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/RTTLogStrings.dat
+++ /dev/null
@@ -1,14 +0,0 @@
-/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_16feb_01:44
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_16feb_09:46
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_15feb_14:21
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_15feb_00:11
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_14feb_22:26
-#/afs/cern.ch/atlas/project/RTT/prod/Peter/logs/logfiles_14feb_18:57
-#/afs/cern.ch/atlas/project/RTT/prod/SLC4_32/logs/logfiles_14feb_11:18
-#/afs/cern.ch/atlas/project/RTT/prod/SLC4_32/logs/logfiles_13feb_19:25
-#/afs/cern.ch/atlas/project/RTT/prod/SLC4_32/logs/logfiles_13feb_22:36
-#/afs/cern.ch/atlas/project/RTT/prod/SLC4_32/logs/logfiles_14feb_07:38
-RTT.log;*error.log
-exceptions.MemoryError
-MemFree:
-
diff --git a/Tools/RunTimeTester/releaseTools/RTTSummaryDumper.py b/Tools/RunTimeTester/releaseTools/RTTSummaryDumper.py
deleted file mode 100755
index 49551be3dd2833e1eb4039b70bf3bb3e1082a2c2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/RTTSummaryDumper.py
+++ /dev/null
@@ -1,78 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from xml.dom.minidom import parse
-import os.path
-
-root = '/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc3-gcc323-opt'
-import xml.dom, time
-import xml.dom.minidom
-
-import sys
-sys.path.append('../src')
-from Evaluate import Evaluate
-#from xml.xpath import Evaluate
-fn = os.path.join(root, 'RTTSummary.xml')
-dom = xml.dom.minidom.parse(fn)
-print 'parsed file:', fn
-
-def makeDict(parentNode):
-    els = [e for e in parentNode.childNodes if e.nodeType == e.ELEMENT_NODE]
-    
-    dict = {}
-    def filldict(e):
-        dict[e.tagName.strip()] = Evaluate('text()', e).strip()#
-        
-    [filldict(e) for e in els]
-    return dict
-
-
-
-de = dom.documentElement
-
-
-dict = makeDict(Evaluate('overview', de)[0])
-print    
-print '---------overview-----------------'
-print
-for i in dict.items():
-    print '%30s    %s'% i
-
-
-
-packageNodes = dom.getElementsByTagName('package')
-for n in packageNodes:
-    name = Evaluate('packageName/text()', n).strip()
-    print '--------------- %s ------------' % name
-
-    dict = makeDict(n)
-    print
-    items = (dict.items())
-    items.sort()
-    for i in items:
-        print '%30s    %s'% i
-
-
-    minders = Evaluate('minder',n)
-    items = []
-    mindermat = {}
-    for m in minders:
-        minderdat = {}
-        minderdat['name']      = (Evaluate('identifiedName/text()',m)).strip()
-        id                     = (Evaluate('jobID/text()', m)).strip()
-        minderdat['id']        = id
-        minderdat['history']   = (Evaluate('stateHistory/text()', m)).strip()
-        minderdat['state']     = (Evaluate('state/text()', m)).strip()
-        minderdat['status']    = (Evaluate('status/text()', m)).strip()
-        minderdat['results']   = (Evaluate('resultsPath/text()', m)).strip()
-        mindermat[id] = minderdat
-
-    ids = mindermat.keys()
-    ids.sort()
-    print 
-    for id in ids:
-        md = mindermat[id]
-        print '\n   %s:' % md.pop('name')
-        items = mindermat[id].items()
-        items.sort()
-        for i in items:
-            print '     %10s:   %s' % i
diff --git a/Tools/RunTimeTester/releaseTools/RTTSummaryDumper2.py b/Tools/RunTimeTester/releaseTools/RTTSummaryDumper2.py
deleted file mode 100755
index bf5f3316405ccd7ac20b81c6e53995dfc80c3ead..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/RTTSummaryDumper2.py
+++ /dev/null
@@ -1,112 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path, string, time, xml.dom.minidom, sys
-sys.path.append('../src')
-from Evaluate import Evaluate
-
-def makeDict(parentNode):
-    els = [e for e in parentNode.childNodes if e.nodeType == e.ELEMENT_NODE]
-    
-    dict = {}
-    def filldict(e):
-        dict[string.ljust(e.tagName.strip(), 30)] = Evaluate('text()', e).strip()#
-        
-    [filldict(e) for e in els]
-    return dict
-
-
-def RTTSummaryToText(dom):
-    
-    de = dom.documentElement
-
-
-    dict = makeDict(Evaluate('overview', de)[0])
-
-    lines = []
-    
-    line  = '|'+'-'*60 +'|'
-    vert  = '|'+' '*60 +'|'
-    title = '|'+string.center('RTT Run Summary', 60)+'|'
-    tim   = '|'+string.center(time.strftime('%d %B %Y  %H:%M  %Z', time.gmtime()), 60) +'|'
-    blank = ' '
-    lines.extend([line, vert, title, tim, vert, line, blank, blank])
-
-    [lines.append('%30s    %s'% i) for i in dict.items()]
-    lines.extend([blank, blank])
-
-    
-    tags = ['nPPSuccessInPackage',
-            'nRetriesInPackage',
-            'nTimeOutsInPackage',
-            'packageName',
-            'nTestsSuccessInPackage',
-            'nJobsOpCloseInPackage',
-            'nTestsFailureInPackage',
-            'containerPackage',
-            'nJobsDoneInPackage',
-            'nJobsSuccessInPackage',
-            'nTestsInPackage',
-            'packageTag',
-            'nPPFailureInPackage',
-            'nJobsFailureInPackage'
-            ]
-
-    packageNodes = dom.getElementsByTagName('package')
-    precords = []
-    total = [string.ljust('totRTT', 30), '-', 0, 0, 0, 0, 0, 0, 0, 0, 0]
-    for n in packageNodes:
-        record = ( (Evaluate('packageName/text()', n).strip()).ljust(30),
-                   Evaluate('phase/text()', n).strip(),
-                   #int(Evaluate('nJobsTotalPackage/text()', n).strip()),
-                   int(Evaluate('nJobsInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsSuccessInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsFailureInPackage/text()', n).strip()),
-                   int(Evaluate('nTimeOutsInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsOpCloseInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsSuccessInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsFailureInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsDoneInPackage/text()', n).strip()))
-
-        for i in range(len(total[2:])):
-            j = i+2
-            total[j] += record[j]
-                  
-
-        precords.append(record)
-
-    precords.sort()
-    records = [tuple(total)]
-    records.extend(precords)
-
-    print
-    labels = (' ', 'pkge', 'jobs', 'jobs', 'jobs', 'jobs', 'oper', 'tests','tests', 'tests', 'process')
-    lines.append( '%30s %5s %5s %5s %5s %5s %5s %5s %5s %5s %8s' % labels)
-    labels = (' ', 'phase', 'submt','succ','fail', 't_out', 'abort', 'total', 'succ', 'fail', 'complete')
-    lines.append('%30s %5s %5s %5s %5s %5s %5s  %5s %5s %5s %8s' % labels)
-
-    for r in records:
-        lines.append( '%s %5s %5d %5d %5d %5d %5d  %5d %5d %5d %8d' % r)
-
-    result = ''
-    for l in lines:
-        result += l  + '\n'
-
-    return result
-
-
-if __name__ == '__main__':
-
-
-    import sys
-    if sys.argv[2] == 'slc4' : conf = 'i686-slc4-gcc34-opt'
-    if sys.argv[2] == 'slc3' : conf = 'i686-slc3-gcc323-opt'
-    # root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/dev/build/i686-slc3-gcc323-opt' % sys.argv[1]
-    root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/dev/build/%s' % (sys.argv[1], conf)
-    fn = os.path.join(root, 'RTTSummary.xml')
-    print fn
-    dom = xml.dom.minidom.parse(fn)
-    print 'parsed ',fn
-
-    
-    print RTTSummaryToText(dom)
diff --git a/Tools/RunTimeTester/releaseTools/RTTSummaryTagDumper.py b/Tools/RunTimeTester/releaseTools/RTTSummaryTagDumper.py
deleted file mode 100755
index 0b8c2fa9489ef36901f92c4ce63573791b47df85..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/RTTSummaryTagDumper.py
+++ /dev/null
@@ -1,144 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-dump minder information
-"""
-
-
-
-import os.path, string, time, xml.dom.minidom, sys, getopt
-sys.path.append('../src')
-from Evaluate import Evaluate
-
-
-def dumpTag(p, de):
-    print 
-
-def dumpMinderTags(fn, paths):
-
-    dom = xml.dom.minidom.parse(fn)
-    de = dom.documentElement
-    pnodes = de.getElementsByTagName('package')
-
-    for p in pnodes:
-        print '\n------- %s --------' % Evaluate('packageName/text()', p)
-        mnodes = p.getElementsByTagName('minder')
-        for m in mnodes:
-            print '   %s' % Evaluate('identifiedName', m)
-            for p in paths:
-                print '      tag: %s   value %s' %  (p, Evaluate(p+'/text()', m))
-
-
-if __name__ == '__main__':
-
-    def usage():
-        print
-        print 'RTTSummaryTagDumper -- dumps minder tags'
-        print
-        print 'usage: python RTTSummaryTagDumper [options]'
-        print
-        print 'Options'
-        print '   -l use the most recent version of the log files. must supply arg branch or release:branch:os'
-        print '      posible release: rel_<n> n in [0,6], possible branch = ("bugfix", "dev", "val")'
-        print '      possible os = ("slc3", "slc4") example: -l rel_0:dev:slc3'
-        print '   -f triggers writing to a file. Filename is written to stdout'
-        print '   -u set the user, eg  brinick, eric, peter, prod, Default prod'
-        print '   -t colon seprated list of tags'
-        print '   -h print this message and exit'
-        return
-    
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 't:b:l:fu:h')
-        print opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    useLatest = False
-    tags    = []
-    slc       = None
-    output    = None
-    branch    = None
-    user      = 'prod'
-    for o, a in opts:
-        if o == '-l':
-            tokens = a.split(':')
-            if len(tokens)!= 3:
-                print 'bad -l'
-                usage()
-                sys.exit(2)
-            
-            release = tokens[0]
-            if release not in ['rel_0', 'rel_1', 'rel_2', 'rel_3', 'rel_4', 'rel_5', 'rel_6']:
-
-                print 'bad release'
-                usage()
-                sys.exit(2)
-                
-            branch = tokens[1]
-            if branch not in ['bugfix', 'dev', 'val']:
-
-                print 'bad branch'
-                usage()
-                sys.exit(2)
-                
-            opsys = tokens[2]
-            if opsys not in ['slc3', 'slc4']:
-                usage()
-                sys.exit(2)
-                
-            if opsys == 'slc3':
-                opsys = 'i686-slc3-gcc323-opt'
-            elif opsys == 'slc4':
-                opsys = 'i686-slc4-gcc34-opt'
-
-
-        elif o == '-u':
-            user = a
-            if user != 'prod': user = 'data/'+user
-        elif o == '-h':
-            usage()
-            sys.exit(0)
-        if o == '-t':
-            tags = a.split(':')
-        if o == '-f':
-            output = '/afs/cern.ch/user/r/rtt/Scratch'
-
-    if not tags:
-        print '-t option required'
-        usage()
-        sys.exit(2)
-
-
-    if not tags:
-        print '-b option required'
-        usage()
-        sys.exit(2)
-
-
-
-    if user == 'prod':
-        root = '/afs/cern.ch/atlas/project/RTT/%s/Results/%s/%s/build/%s' % (user, release, branch, opsys)
-    else:
-        root = '/afs/cern.ch/atlas/project/RTT/%s/Results/%s/%s/build/%s' % (user, release, branch, opsys)
-
-    fn = os.path.join(root, 'RTTSummary.xml')
-    print 'looking for summaryFile %s' %  fn
-
-
-    if not os.path.exists(fn):
-        print 'Balking as RTTSummary file does not exist:\n%s' % root
-        sys.exit(0)
-
-
-    if output:
-        cname = 'RTTSummaryTagDump'
-        cname += '_'+os.path.basename(root)+'.log'
-        output = os.path.join(output, cname)
-        of = open(output,'w')
-        print 'writing results to ', output
-    else:
-        of = sys.stdout
-
-
-    of.write(dumpMinderTags(fn, tags))
diff --git a/Tools/RunTimeTester/releaseTools/RTTwatcher.py b/Tools/RunTimeTester/releaseTools/RTTwatcher.py
deleted file mode 100644
index f09d73a0253e91d1652517a4b0920567a0fb2cc9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/RTTwatcher.py
+++ /dev/null
@@ -1,594 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-import sys
-import os.path
-import time
-from popen2 import Popen4
-
-sys.path.append(os.environ['SCRIPTS'])
-from utils import SYSPATH, RTT, ATLAS
-SYS  = SYSPATH()
-SYS += ['prod_src', 'prod_share']
-  
-from Evaluate import Evaluate
-
-from  xml.dom.minidom import getDOMImplementation
-
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-outputPath = ''
-try:
-    outputPath = sys.argv[1].strip()
-except:
-    print 'Please give the directory in which results should be output.'
-    sys.exit(1)
-    
-if not os.path.exists(outputPath):
-    print '%s: inexistant' % outputPath
-    sys.exit(1)
-
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-
-def makeDOM(docname):
-    impl            = getDOMImplementation()
-    docType         = impl.createDocumentType(docname, '', '')
-    namespace       = None
-    document        = impl.createDocument(namespace, docname, docType)
-    
-    return document
-
-# global dom
-theDom = makeDOM('rttStatus')
-
-# useful shortcuts
-xmlElementNode = theDom.createElement
-xmlTextNode    = theDom.createTextNode
-
-_currentruns = RTT.Current.runs
-
-# ------------------------------------------------------------------------------------
- 
-def createTextNode(tagName, tagTextContent):
-    tag  = xmlElementNode(tagName)
-    text = xmlTextNode(tagTextContent)
-    tag.appendChild(text)
-    return tag
-
-# ------------------------------------------------------------------------------------
-
-def getPopen4Output(cmd, max_lines=-1):
-    po = Popen4(cmd)
-    out = []
-    while po.poll()==-1:
-        out.extend(po.fromchild.readlines())
-        if max_lines>0 and length(out)>=max_lines: break
-    del po
-    return out
-
-# ------------------------------------------------------------------------------------
-# =====================================================================================
-# =====================================================================================
-# =====================================================================================
-
-class ClusterHealth:
-    def __init__(self):
-        self.urlBase = 'http://lemon-gateway.cern.ch/lemon-xml/xml_gateway.php?'
-        self.nodes   = self.getnodes()
-        self.metrics = {'30045':'exception.no_contact',
-                        '20002':'LoadAvg',
-                        '6404': 'HostPsNjobs'}
-
-    # ----------------------------------------------------
-
-    def getnodes(self):
-        url = self.urlBase + 'clusters=lxbatch/atlasrtt&metrics=20002'
-        docEl = self.getLemonXMLResponse(url)
-        return [d.getAttribute('node').strip() for d in Evaluate('metric/data', docEl)]
-        
-    # ----------------------------------------------------
-
-    def makeEntityURLfrag(self):
-        entitiesString = 'entities='
-        for node in self.nodes: entitiesString += '%s,' % node
-        entitiesString = entitiesString[:-1] # remove trailing comma
-        return entitiesString
-
-    # ----------------------------------------------------
-    
-    def makeMetricsURLfrag(self):
-        metricsString = 'metrics='
-        for metric in self.metrics.values(): metricsString += '%s,' % metric
-        metricsString = metricsString[:-1] # remove trailing comma
-        return metricsString
-
-    # ----------------------------------------------------
-    
-    def makeURL(self):
-        return self.urlBase + self.makeEntityURLfrag() + '&' + self.makeMetricsURLfrag()
-        
-    # ----------------------------------------------------
-
-    def getLemonXMLResponse(self, url):
-        import urllib
-        from xml.dom.minidom import parseString
-        handle = urllib.urlopen(url)
-        xml_file = handle.read()
-        handle.close()
-        xml_dom  = parseString(xml_file)
-        return xml_dom.documentElement        
-        
-    # ----------------------------------------------------
-    
-    def handleNoContactMetric(self, metricEl):
-        nodeContactInfo = {}
-        for node in self.nodes: nodeContactInfo[node] = 'DOWN' # initialise to "all machines down"
-        
-        dataEls = metricEl.getElementsByTagName('data')
-        
-        for dataEl in dataEls:
-            nodeName = dataEl.getAttribute('node').strip()
-            nodeContactInfo[nodeName] = 'UP'
-        return nodeContactInfo    
-
-    # ----------------------------------------------------
-
-    def handleLoadAvg(self, metricEl):
-        loadAvgInfo = {}        
-        for node in self.nodes: loadAvgInfo[node] = '-1' # initialise to no load found
-
-        dataEls = metricEl.getElementsByTagName('data')
-        for dataEl in dataEls:
-            # node name
-            nodeName = dataEl.getAttribute('node').strip()
-            # when load was measured on this node last
-            timestamp = Evaluate('r', dataEl)[0].getAttribute('ts').strip()
-
-            # what the load was
-            load = Evaluate('r/d/text()', dataEl)
-            loadAvgInfo[nodeName] = (load, timestamp)
-        return loadAvgInfo
-
-    # ----------------------------------------------------
-
-    def handleNumbBatchJobs(self, metricEl):
-        numbJobsInfo = {}
-        for node in self.nodes: numbJobsInfo[node] = '-1'
-        dataEls = metricEl.getElementsByTagName('data')
-        for dataEl in dataEls:
-            nodeName = dataEl.getAttribute('node').strip()
-            timestamp = Evaluate('r', dataEl)[0].getAttribute('ts').strip()
-            # number of jobs running on this node
-            numbJobs = Evaluate('r/d/text()', dataEl)
-            numbJobsInfo[nodeName] = (numbJobs, timestamp)
-        return numbJobsInfo
-
-    # ----------------------------------------------------
-    
-    def handle(self, metric_name, metricEl):
-        dispatcher = {'exception.no_contact': self.handleNoContactMetric,
-                      'LoadAvg':self.handleLoadAvg,
-                      'HostPsNjobs': self.handleNumbBatchJobs
-                      }
-        return dispatcher[metric_name](metricEl)
-    
-    # ----------------------------------------------------
-            
-    def parse(self, parentEl):
-        info = {}
-        metricEls = parentEl.getElementsByTagName('metric')
-        for metricEl in metricEls:
-            metric_name = self.metrics.get(metricEl.getAttribute('id'), None)
-            if not metric_name: continue
-            info[metric_name] = self.handle(metric_name, metricEl)
-        return info
-    
-    # ----------------------------------------------------
-
-    def constructXMLOutput(self, info):
-        lemonEl = xmlElementNode('lemon_info')
-        lemonEl.appendChild(createTextNode('numbNodes', str(len(self.nodes))))
-        
-        for node_name in self.nodes:
-            nodeEl = xmlElementNode('node')
-            nodeEl.appendChild(createTextNode('nodeName', node_name))
-            # contact info i.e. is machine up
-            nodeEl.appendChild(createTextNode('up', info['exception.no_contact'][node_name]))
-            # load avg info
-            load, timestamp = info['LoadAvg'][node_name]
-            nodeEl.appendChild(createTextNode('loadAvg', load))
-            # number of batch jobs running on the node
-            jobs, timestamp = info['HostPsNjobs'][node_name]
-            nodeEl.appendChild(createTextNode('numbJobs', jobs))
-            lemonEl.appendChild(nodeEl)
-        return lemonEl
-
-    def run(self):
-        parentNode = self.getLemonXMLResponse(self.makeURL())
-        result = self.parse(parentNode)
-        return self.constructXMLOutput(result)
-        
-# =====================================================================================
-# =====================================================================================
-# =====================================================================================
-
-class JobsInBatch:
-    def __init__(self):
-        self.cmd = 'python /afs/cern.ch/atlas/project/RTT/scripts/prod/list_jobs.py'
-    def run(self):
-        lines = [l for l in getPopen4Output(self.cmd)[:-2] if l.find('Brinick')==-1] #ignore Brinick jobs
-        jobsEl = xmlElementNode('jobsInBatch')
-        runs = {}
-        for line in lines:
-            lsf_info,run_info,job_info = line.strip().split(']')[:-1]
-            lsf_info = lsf_info.strip()[1:] # remove '('
-            run_info = run_info.strip()[1:] # remove '('
-            job_info = job_info.strip()[1:] # remove '('
-            if run_info not in runs.keys(): runs[run_info] = []            
-            runs[run_info].append((job_info,lsf_info))
-
-        running = pending = 0
-        for k,v in runs.items():
-            for job_info, lsf_info in v:
-                if lsf_info.find('RUN') != -1: running += 1
-                if lsf_info.find('PEND') != -1: pending += 1
-
-        runs2 = {}
-        for k,jobs in runs.items():
-            runs2[k] = {}
-            for job_info, lsf_info in jobs:
-                batchNode = lsf_info.split(',')[1].strip()
-                if batchNode not in runs2[k].keys(): runs2[k][batchNode] = []
-                runs2[k][batchNode].append((job_info, lsf_info))
-                
-        overEl = xmlElementNode('overview')
-        overEl.appendChild(createTextNode('jobsTotal',str(len(lines))))
-        overEl.appendChild(createTextNode('jobsRunning',str(running)))
-        overEl.appendChild(createTextNode('jobsPending',str(pending)))
-        jobsEl.appendChild(overEl)
-
-        runsEl = xmlElementNode('runs')
-        for run,dict in runs2.items():
-            runEl = xmlElementNode('run')
-            runEl.appendChild(createTextNode('name',str(run)))
-            # v.sort()
-
-            for batch_node, jobs in dict.items():
-                batchEl = xmlElementNode('batchNode')
-                batchEl.appendChild(createTextNode('name', batch_node))
-                for job_info, lsf_info in jobs:
-                    jobEl = xmlElementNode('job')
-                    jobEl.appendChild(createTextNode('lsf_info',str(lsf_info)))
-                    jobEl.appendChild(createTextNode('job_info',str(job_info)))
-                    batchEl.appendChild(jobEl)
-                runEl.appendChild(batchEl)
-            runsEl.appendChild(runEl)
-        jobsEl.appendChild(runsEl)
-        return jobsEl
-        
-# =====================================================================================
-# =====================================================================================
-# =====================================================================================
-
-class ExtractCronInfo:
-    def __init__(self):
-        self.copier_jobs, (self.rtt_jobs, self.tct_jobs, self.fct_jobs), self.other_jobs = self.load(os.path.join(os.environ['RES'], 'rtt.acrontable'))
-
-    def load(self, fpath):
-        h = open(fpath)
-        jobs = h.readlines()
-        h.close()
-        jobs = [j for j in jobs if not j.strip().startswith('#')] # remove comment lines
-        
-        copier_jobs = [j for j in jobs if j.lower().find('datacopier')!=-1]
-
-        rtt_jobs   = [j for j in jobs if j.find('/afs/cern.ch/atlas/project/RTT/prod/launch/run.rtt.prod.sh')!=-1 and j.find('log/rtt')!=-1]
-        fct_jobs   = [j for j in jobs if j.find('/afs/cern.ch/atlas/project/RTT/prod/launch/run.rtt.prod.sh')!=-1 and j.find('log/fct')!=-1]
-        tct_jobs   = [j for j in jobs if j.find('/afs/cern.ch/atlas/project/RTT/prod/launch/run.rtt.prod.sh')!=-1 and j.find('log/tct')!=-1]
-
-        all_jobs = copier_jobs + rtt_jobs + tct_jobs + fct_jobs
-        other_jobs = [j for j in jobs if j not in all_jobs]
-        return (copier_jobs, (rtt_jobs, tct_jobs, fct_jobs), other_jobs)
-
-    def run(self):
-        cronJobsEl = xmlElementNode('cronJobs')    
-        copierEl = xmlElementNode('copierCronJobs')
-
-        rttEl    = xmlElementNode('rttCronJobs')
-        tctEl    = xmlElementNode('tctCronJobs')
-        fctEl    = xmlElementNode('fctCronJobs')
-        
-        otherEl    = xmlElementNode('otherCronJobs')
-
-        for job in self.copier_jobs:
-            copierEl.appendChild(createTextNode('job', job.strip()))
-
-        for job in self.rtt_jobs:
-            rttEl.appendChild(createTextNode('job', job.strip()))
-
-        for job in self.tct_jobs:
-            tctEl.appendChild(createTextNode('job', job.strip()))
-
-        for job in self.fct_jobs:
-            fctEl.appendChild(createTextNode('job', job.strip()))
-
-        for job in self.other_jobs:
-            otherEl.appendChild(createTextNode('job', job.strip()))
-
-        cronJobsEl.appendChild(copierEl)
-        cronJobsEl.appendChild(rttEl)
-        cronJobsEl.appendChild(tctEl)
-        cronJobsEl.appendChild(fctEl)                        
-        cronJobsEl.appendChild(otherEl)
-        return cronJobsEl
-
-# ------------------------------------------------------------------------------------
-
-class NightlyFlagAvailability:
-    def __init__(self):
-        self.todaysDate = self.getTodayDate()
-
-    def buildFlag(self, branch, platform, project):
-        projects = {'offline':'AtlasProduction', 'p1hlt':'AtlasP1HLT', 'tier0':'AtlasTier0', 'hlt':'AtlasHLT'}
-        base = {'hlt':'/afs/cern.ch/atlas/project/hlt/builds/nightlies'}.get(project, '/afs/cern.ch/atlas/software/builds/nightlies')
-        arch, OS, compiler, build = platform.split('-')
-        OS     = OS[0].upper()+OS[-1] # e.g S4
-        bits   = {'i686':32,'x86_64':64}.get(arch)
-        suffix = '%sB%sAll%s' % (bits, OS, build[0].upper()+build[1:])
-        return os.path.join(base, branch, projects[project], 'latest_copied_release%s' % suffix)
-
-    def getTodayDate(self):
-        today = time.ctime().split()
-        todaysDate = {}
-        todaysDate['month'] = today[1]
-        todaysDate['day']   = today[2]
-        todaysDate['hour']  = today[3].split(':')[0]
-        todaysDate['mins']  = today[3].split(':')[1]
-        return todaysDate
-
-    def run(self):
-        nightliesEl = xmlElementNode('nightliesAvailable')
-
-        for _run in _currentruns(): # current runs
-            branch, releaseType, platform, project = _run.split('/')
-            flag = self.buildFlag(branch, platform, project)
-
-            if not os.path.exists(flag): continue
-            
-            cmd = 'cd %s;ls -l %s' % (os.path.dirname(flag), os.path.basename(flag))
-
-            nightlyEl = xmlElementNode('nightly')
-
-            platEl = createTextNode('platform', str(platform))
-            flagEl = createTextNode('flagLocation', flag)
-            flagRelEl = createTextNode('flagRelease', os.readlink(flag).strip())
-
-            nightlyEl.appendChild(platEl)
-            nightlyEl.appendChild(flagEl)
-            nightlyEl.appendChild(flagRelEl)
-        
-            out = getPopen4Output(cmd)
-
-            if not out:
-                nightlyEl.appendChild(createTextNode('available', 'unable to obtain any information'))
-                nightliesEl.appendChild(nightlyEl)
-                print '%s: unable to obtain any information' % str(platform)
-                continue
-
-            out   = out[0].split()
-            month = out[5]
-            day   = out[6]
-            hour  = out[7].split(':')[0]
-            mins  = out[7].split(':')[1]
-
-            theDate = '%s %s@%s:%s' % (day, month, hour, mins)
-
-            isAvailable = False
-            if month == self.todaysDate['month'] and day == self.todaysDate['day']:
-                if int(hour) < int(self.todaysDate['hour']):
-                    isAvailable = True
-                elif (int(hour) == int(self.todaysDate['hour']) and (int(mins) < int(self.todaysDate['mins']))):
-                    isAvailable = True
-
-            dateEl = createTextNode('date', theDate)
-
-            if isAvailable:
-                readyEl = createTextNode('available', 'READY')
-                status = 'READY'
-            else:
-                readyEl = createTextNode('available', 'NOT READY')            
-                status = 'NOT READY'
-            
-            nightlyEl.appendChild(dateEl)
-            nightlyEl.appendChild(readyEl)
-            nightliesEl.appendChild(nightlyEl)    
-        return nightliesEl
-
-# ------------------------------------------------------------------------------------
-
-class CheckQuotas:
-    def __init__(self):
-        self.resBase        = '/afs/cern.ch/atlas/project/RTT/Results'
-        self.bigfilesBase0  = '/afs/cern.ch/atlas/project/RTT/Results/tempfiles/0/nightly'
-        self.bigfilesBase1  = '/afs/cern.ch/atlas/project/RTT/Results/tempfiles/1/nightly'        
-        self.bigfile_quotas = []
-        self.admin_quotas   = ['/afs/cern.ch/user/r/rtt/',
-                               self.resBase,
-                               '/afs/cern.ch/atlas/project/RTT/data',
-                               '/afs/cern.ch/atlas/project/RTT/prod',
-                               '/afs/cern.ch/atlas/project/RTT/prod/log']
-
-    def getQuotaUsage(self, path):
-        out = getPopen4Output('fs lq %s' % path)
-        volume = out[1].split()[0]
-        quota  = out[1].split()[1]
-        used   = out[1].split()[2]
-        frac   = out[1].split()[3]
-        frac   = frac[:frac.find('%')]
-        return (path, volume, quota, used, frac)
-
-    def getDateStamp(self, thing):
-        from stat import ST_CTIME
-        summ = os.path.join(thing, 'RTTSummary.xml')
-        if not os.path.exists(summ): return 'n/a'        
-        return time.strftime('%d%b@%H:%M', time.localtime(os.lstat(summ)[ST_CTIME]))
-    
-    def uniquifyThenSort(self, what):
-        from sets import Set
-        what = list(Set([b for b in what if os.path.exists(b)]))
-        what.sort()
-        return what
-
-    def run(self):
-        runs_dict = {}
-
-        # big file quotas
-        for each in _currentruns(): # current runs
-            bigfiles1 = os.path.join(self.bigfilesBase0, '%s') % each
-            bigfiles2 = os.path.join(self.bigfilesBase1, '%s') % each
-            self.bigfile_quotas.extend([bigfiles1, bigfiles2])
-
-            for rel in ['rel_0','rel_1','rel_2','rel_3','rel_4','rel_5','rel_6']:
-                resDir = os.path.join(self.resBase, '%s/%s' % (rel, each))
-                runs_dict.setdefault(rel, []).append(resDir)
-
-        self.bigfile_quotas = self.uniquifyThenSort(self.bigfile_quotas)
-
-        quotasEl = xmlElementNode('quotas')
-    
-        admin_quotas = [self.getQuotaUsage(path) for path in self.admin_quotas]
-        bigfile_quotas = [self.getQuotaUsage(path) for path in self.bigfile_quotas]
-
-        for k,v in runs_dict.items():
-            runs_dict[k] = [(self.getQuotaUsage(path), self.getDateStamp(path)) for path in v if os.path.exists(path)]
-
-        def createQuotaNode(quota_results, date_stamp=None):
-            path, volName, tot, used, frac  = quota_results
-            error = None
-            try:
-                if int(tot) < 10000:
-                    return None # not in use quota
-            except:
-                error = 'Problem retrieving!'
-            
-            quotaEl = xmlElementNode('quota')
-            quotaEl.appendChild(createTextNode('dir', path))
-            quotaEl.appendChild(createTextNode('volume', volName))
-
-            if not error:
-                quotaEl.appendChild(createTextNode('total', tot))
-                quotaEl.appendChild(createTextNode('used', used))
-                quotaEl.appendChild(createTextNode('frac', frac))
-                if date_stamp:
-                    quotaEl.appendChild(createTextNode('date', date_stamp))
-            else:
-                quotaEl.appendChild(createTextNode('total', error))
-                quotaEl.appendChild(createTextNode('used', 'n/a'))
-                quotaEl.appendChild(createTextNode('frac', 'n/a'))
-                
-            return quotaEl
-
-        def wrap(parentNodeName, els):
-            parentNode = xmlElementNode(parentNodeName)
-            parentNode.appendChild(createTextNode('quota_cat_name', parentNodeName))
-            [parentNode.appendChild(qel) for qel in els if qel]
-            quotasEl.appendChild(parentNode)
-
-        wrap('admin_quotas', [createQuotaNode(qr) for qr in admin_quotas])
-        wrap('bigfile_quotas', [createQuotaNode(qr) for qr in bigfile_quotas])
-
-        # run quotas are a bit different in structure
-        parentNode = xmlElementNode('runs_quotas')
-        parentNode.appendChild(createTextNode('quota_cat_name', 'runs_quotas'))
-        for rel, runs in runs_dict.items():
-            rel_node = xmlElementNode('release')
-            rel_node.appendChild(createTextNode('release_name', rel))
-            for quota, date_stamp in runs:
-                qr_node = createQuotaNode(quota, date_stamp)
-                if qr_node:
-                    rel_node.appendChild(qr_node)
-            parentNode.appendChild(rel_node)
-
-        quotasEl.appendChild(parentNode)
-        return quotasEl
-
-# ------------------------------------------------------------------------------------
-
-class CheckRTTsummaries:
-    def __init__(self):
-        self.resBase = '/afs/cern.ch/atlas/project/RTT/Results'
-        self.summFile = os.path.join(self.resBase, 'summaryFilePaths.txt')
-
-    def readFile(self, fpath):
-        h = open(fpath)
-        cont = h.readlines()
-        h.close()
-        return cont
-
-    def getSummTextFiles(self, lines):
-        def file_exists(l):
-            return os.path.exists(os.path.join(self.resBase, os.path.dirname(l.strip()), 'RTTSummary.txt'))
-
-        from sets import Set
-        lines = [l for l in lines if l.find('rel_')!=-1]
-        lines = list(Set(lines)) # uniquify
-        lines = [l for l in lines if l.split('rel_')[1][2:].strip().split('/RTTSummary.xml')[0] in _currentruns()] # get only current runs summaries
-
-        summaries = [os.path.join(self.resBase, os.path.dirname(l.strip()), 'RTTSummary.txt') for l in lines if file_exists(l)]
-        summaries.sort()
-        return summaries
-
-    def run(self):
-        summFile  = os.path.join(self.resBase, self.summFile)
-        summaries = self.getSummTextFiles(self.readFile(summFile))
-
-        summsEl   = xmlElementNode('runSummaries')
-    
-        for s in summaries:
-            cont = self.readFile(s)
-            totRTT = [l for l in cont if l.find('totRTT')!=-1]
-            if len(totRTT)!=1:
-                continue
-
-            toks = totRTT[0].split()
-            toks = toks[2:]
-
-            summEl = xmlElementNode('summary')
-            summEl.appendChild(createTextNode('runID', s.strip()))
-            summEl.appendChild(createTextNode('jobsTot', toks[0].strip()))
-            summEl.appendChild(createTextNode('jobsSubmit', toks[1].strip()))
-            summEl.appendChild(createTextNode('jobSucc', toks[2].strip()))
-            summEl.appendChild(createTextNode('jobsFail', toks[3].strip()))
-            summEl.appendChild(createTextNode('jobsTout', toks[4].strip()))
-            summEl.appendChild(createTextNode('jobsBerr', toks[5].strip()))
-            summEl.appendChild(createTextNode('jobsOpShut', toks[6].strip()))
-            summEl.appendChild(createTextNode('jobsDone', toks[-1].strip()))
-
-            summsEl.appendChild(summEl)
-
-        return summsEl
-
-# ------------------------------------------------------------------------------------
-
-def xmlToFile():
-    xml_file = os.path.join(outputPath, 'RTTstatus.xml')
-    of = open(xml_file, 'w')    
-    of.write(theDom.toprettyxml('   ','\n'))
-    of.close
-
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-# ------------------------------------------------------------------------------------
-    
-if __name__ == '__main__':
-    toDo = [ClusterHealth, ExtractCronInfo, NightlyFlagAvailability, CheckQuotas, CheckRTTsummaries, JobsInBatch]
-    [theDom.documentElement.appendChild(xmlSnippet) for xmlSnippet in [callable().run() for callable in toDo]]
-    xmlToFile()
-
-    lastRun = os.path.join(outputPath, 'rttStatusLastRun')
-    handle  = open(lastRun, 'w')
-    handle.write(time.ctime())
-    handle.close()
diff --git a/Tools/RunTimeTester/releaseTools/areNightlyFlagsReady.py b/Tools/RunTimeTester/releaseTools/areNightlyFlagsReady.py
deleted file mode 100755
index ee91a9d5f801138eff714e25a06090c176a15580..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/areNightlyFlagsReady.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, time
-from popen2 import Popen4
-
-files = {
-    ('bug', '32bit', 'SLC4') : '/afs/cern.ch/atlas/software/builds/nightlies/bugfix/AtlasProduction/latest_copied_releaseBF32BS4AllOpt',
-    ('bug', '32bit', 'SLC3') : '/afs/cern.ch/atlas/software/builds/nightlies/bugfix/AtlasProduction/latest_copied_releaseBF32BS3AllOpt',
-    ('val', '32bit', 'SLC4') : '/afs/cern.ch/atlas/software/builds/nightlies/val/AtlasProduction/latest_copied_releaseVAL32BS4AllDbg'
-    }
-
-today = time.ctime().split()
-todaysDate = {}
-todaysDate['month'] = today[1]
-todaysDate['day']   = today[2]
-todaysDate['hour']  = today[3].split(':')[0]
-todaysDate['mins']  = today[3].split(':')[1]
-
-for platform, flag in files.items():
-    cmd = 'cd %s;ls -l %s' % (os.path.dirname(flag), os.path.basename(flag))
-    po = Popen4(cmd)
-    out = []
-    while po.poll()==-1:
-        out.extend(po.fromchild.readlines())
-
-    if not out:
-        print '%s: unable to obtain any information' % str(platform)
-        continue
-
-    out   = out[0].split()
-    month = out[5]
-    day   = out[6]
-    hour  = out[7].split(':')[0]
-    mins  = out[7].split(':')[1]
-
-    theDate = '%s %s@%s:%s' % (day, month, hour, mins)
-
-    isAvailable = False
-    if month == todaysDate['month'] and day == todaysDate['day']:
-        if int(hour) < int(todaysDate['hour']):
-            isAvailable = True
-        elif (int(hour) == int(todaysDate['hour']) and (int(mins) < int(todaysDate['mins']))):
-            isAvailable = True
-
-    if isAvailable: 
-        status = 'READY'
-    else:
-        status = 'NOT READY'    
-    
-    print '%s.....%s (%s)' % (str(platform), status, theDate)
-
diff --git a/Tools/RunTimeTester/releaseTools/closeDownKeyDumper.py b/Tools/RunTimeTester/releaseTools/closeDownKeyDumper.py
deleted file mode 100755
index f3f30d8c2f1d23abe70c5aabe9eb0ced2a0403f7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/closeDownKeyDumper.py
+++ /dev/null
@@ -1,37 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import xml.dom, time
-import xml.dom.minidom
-
-import sys
-sys.path.append('../src')
-from Evaluate import Evaluate
-fn = '/afs/cern.ch/atlas/project/RTT/Results/closeDownFile.xml'
-dom = xml.dom.minidom.parse(fn)
-
-de = dom.documentElement
-keys = dom.getElementsByTagName('key')
-for k in keys:
-    els = [e for e in k.childNodes if e.nodeType == e.ELEMENT_NODE]
-
-    dict = {}
-    def filldict(e):
-        dict[e.tagName] = Evaluate('text()', e)#
-        
-    [filldict(e) for e in els]
-
-    key = u'startTime'
-    # convert to human readable format
-    try:
-        dict[key] = time.strftime('%y/%m/%d %H:%M',
-                                  time.localtime(float(dict[key])))
-    except Exception.KeyError, e:
-        print 'no time to convert'
-    except:
-        print 'unknown error converting time', e.__class__.__name__
-
-    print    
-    print '--------------------------'
-    print
-    for i in dict.items():
-        print '%30s    %s'% i
diff --git a/Tools/RunTimeTester/releaseTools/dailyRTT.py b/Tools/RunTimeTester/releaseTools/dailyRTT.py
deleted file mode 100755
index cde9e5d841ce5e7ce747b8779d34a43418694bf7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/dailyRTT.py
+++ /dev/null
@@ -1,349 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time, xml.dom.minidom, os.path, getopt, sys
-sys.path.append('/afs/cern.ch/atlas/project/RTT/prod/Tools/RunTimeTester/src')
-from Evaluate    import Evaluate
-from exc2string2 import exc2string2
-from makeRTTDOM  import makeRTTDOM
-
-
-# ----- global stuff -----
-
-daysecs = 24*3600
-# value of each entry is the date of the first run (DD/MM/YYYY)
-
-runIdentifiers = [('i686-slc4-gcc34-opt', 'bugfix'),
-                  ('i686-slc4-gcc34-opt', 'dev'),
-                  ('i686-slc4-gcc34-opt', 'val'),
-                  ('i686-slc4-gcc34-dbg', 'bugfix'),
-                  ('i686-slc3-gcc323-opt', 'pcache'),
-                  ('i686-slc3-gcc323-opt', 'bugfix'),
-                  ]
-
-runTypeIDtags = ['originalBranch', 'targetCMTCONFIG']
-
-numberOfDaysToGoBackInTime = 6
-
-prodResultsBase = '/afs/cern.ch/atlas/project/RTT/Results'
-
-# --------------------------------------------------------------
-
-def getReports(fn):
-    sortedReports = {}
-    reportsDoc = None
-    reports = []
-
-    if os.path.exists(fn):
-        reportsDoc = xml.dom.minidom.parse(fn)
-        reports = Evaluate('report', reportsDoc.documentElement)
-    else:
-        print 'no reports file', fn
-        return sortedReports
-
-    for r in reports:
-        
-        rtime     = Evaluate('startTime1/text()', r).strip()
-        platform  = Evaluate('CMTconfig/text()', r).strip()
-        branch    = Evaluate('branch/text()', r).strip()
-        dayOfWeek = Evaluate('humanDate/dayOfWeek/text()', r).strip()
-        date      = Evaluate('humanDate/date/text()', r).strip()
-        month     = Evaluate('humanDate/month/text()', r).strip()
-        year      = Evaluate('humanDate/year/text()', r).strip()
-
-            
-        if date and int(date) < 10 and len(date)==1:
-            date = '0' + date
-        day = '%s %s %s %s' % (dayOfWeek, date, month, year)
-        sortedReports[(branch, platform, day, rtime)] = r
-
-    return sortedReports
-
-# --------------------------------------------------------------
-
-def getEarliestTime():
-    startsecs = time.time()-numberOfDaysToGoBackInTime*daysecs
-    
-    start = time.strftime('%y:%m:%d', time.localtime(startsecs))
-    start = time.strptime(start+':00:00:00', '%y:%m:%d:%H:%M:%S')
-    startsecs = time.mktime(start)
-    return startsecs
-
-# --------------------------------------------------------------
-
-def makeByDayTable():
-    '''Return an empty grid of all possible runs. Later we fill
-    in the runs that did occur.'''
-    def includeEntry(loopDate, entryDate):
-        eday,emonth,eyear = entryDate.split('/')
-        dayName, lday, lmonth, lyear = loopDate.split()
-        lmonth = {'January':1,'February':2,'March':3,'April':4,
-                  'May':5,'June':6,'July':7,'August':8,
-                  'September':9,'October':10,'November':11,'December':12}.get(lmonth)
-        if int(eyear) < int(lyear): return False
-        if int(eyear) > int(lyear): return True
-        if int(emonth) < int(lmonth): return False
-        if int(emonth) > int(lmonth): return True
-        if int(eday) < int(lday): return False
-        if int(eday) >= int(lday): return True
-        
-    startsecs = getEarliestTime()
-    
-    startL = [startsecs+d*daysecs for d in range(numberOfDaysToGoBackInTime+1)]
-    startD = [time.strftime('%A %d %B %Y', time.localtime(s)) for s in startL]
-
-    byDay = {}
-    for d in startD:
-        byDay[d] = {}
-        for r in runIdentifiers:
-            byDay[d][r] = []
-
-    return (startD, byDay)
-
-# --------------------------------------------------------------
-
-def sortOverViewNodesByDay(ovNodes):
-    # runsByDay is a dictionary of dictionaries.
-    orderedKeys, runsByDay = makeByDayTable()
-
-    for o in ovNodes:
-        stime  = float(Evaluate('startTime2/text()', o))
-
-        branch = Evaluate(runTypeIDtags[0] + '/text()', o).strip()
-        config = Evaluate(runTypeIDtags[1] + '/text()', o).strip()
-        
-        day = time.strftime('%A %d %B %Y', time.localtime(stime))
-        if (config, branch) in  runsByDay[day]:
-            runsByDay[day][(config,branch)].append(o)
-    return orderedKeys, runsByDay
-
-# --------------------------------------------------------------
-
-def getOverviewNodes(fn, weekDom):
-    fns = open(fn, 'r').readlines()
-    fns = [f.strip() for f in fns]
-    fns = [f for f in fns if os.path.exists(f)]
-    overviewNodes = []
-
-    for f in fns:
-        try:
-            d = xml.dom.minidom.parse(f)
-        except:
-            print 'error parsing ', f
-            print exc2string2()
-            continue # ignore unparseable files
-
-        on = Evaluate('overview', d.documentElement)[0]
-        domStart = float(Evaluate('startTime2/text()', on))
-
-        earliestTime = getEarliestTime()
-        if domStart > earliestTime:    
-            overviewNodes.append(weekDom.importNode(on, True))
-            
-        d.unlink()
-    return overviewNodes
-
-# --------------------------------------------------------------    
-
-def getStartTimeFromOverViewNode(on):
-    return Evaluate('startTime1/text()', on)
-
-# --------------------------------------------------------------
-
-def sortByStartDate(overV1, overV2):
-    startTime1 = getStartTimeFromOverViewNode(overV1)
-    startTime2 = getStartTimeFromOverViewNode(overV2)
-    if startTime1 < startTime2: return -1
-    if startTime1 > startTime2: return 1
-    return 0
-
-# --------------------------------------------------------------
-
-def makeWeeklyViewDom(overviewNodes, reportTimes, weekDom):
-    dtags  = ['release', 'startTime1', 'startTime2', 'endTime1','statusText','topProject','nPackages', 'nPackagesAllJobsOK']
-
-    weekDom = makeRTTDOM('weeklyReport')
-
-    orderedDays, runsByDay = sortOverViewNodesByDay(overviewNodes)
-    orderedDays.reverse()
-
-    # day is of format: Tuesday 05 May 2007
-    for day in orderedDays:
-
-        el = weekDom.createElement('day')
-        el2 = weekDom.createElement('name')
-        el2.appendChild(weekDom.createTextNode(day))
-        el.appendChild(el2)
-
-        runsEl = weekDom.createElement('runs')
-        
-        for runID in runsByDay[day]:
-            # is there an actual run for this platform/branch ?
-            on = runsByDay[day][runID]
-            
-            if not on:
-                runEl = weekDom.createElement('run')
-                
-                config = runID[0]
-                branch = runID[1]
-                branchEl  = weekDom.createElement('originalBranch')
-                branchEl.appendChild(weekDom.createTextNode(branch))
-                
-                configEl  = weekDom.createElement('targetCMTCONFIG')
-                configEl.appendChild(weekDom.createTextNode(config))
-                
-                # now append these two to the <run> parent element
-                runEl.appendChild(branchEl)
-                runEl.appendChild(configEl)
-
-                # is there a report for this?
-                startTime = '---'
-                possibleReportKey = (branch, config, day, startTime)
-                r = reportTimes.get(possibleReportKey, None)
-                if r: runEl.appendChild(r)
-
-                runsEl.appendChild(runEl)
-                continue
-
-
-            for o in on:
-                runEl = weekDom.createElement('run')
-
-                config = runID[0]
-                branch = runID[1]
-                branchEl  = weekDom.createElement('originalBranch')
-                branchEl.appendChild(weekDom.createTextNode(branch))
-                
-                configEl  = weekDom.createElement('targetCMTCONFIG')
-                configEl.appendChild(weekDom.createTextNode(config))
-                
-                # now append these two to the <run> parent element
-                runEl.appendChild(branchEl)
-                runEl.appendChild(configEl)
-                
-                startTime = Evaluate('startTime1/text()', o)
-                startTime = startTime.strip()
-            
-                nodes = []
-                for t in dtags:
-                    e = Evaluate(t, o)
-                    if e:
-                        e = e[0]
-                        nodes.append(e)
-                nodes = [n.cloneNode(n) for n in nodes] # allow the on to be unlinked
-            
-                o.unlink()
-                
-                # add a report if there is one
-                possibleReportKey = (branch, config, day, startTime)
-                report = reportTimes.get(possibleReportKey, None)
-                if report: nodes.append(report)
-            
-                # append all these extra nodes to the <run> parent element
-                [runEl.appendChild(n) for n in nodes]
-        
-                runsEl.appendChild(runEl)
-
-        el.appendChild(runsEl)
-        weekDom.documentElement.appendChild(el)
-
-    return weekDom
-
-# --------------------------------------------------------------
-
-def display(weeklyDom, longReport):
-    orderedKeys, runsByDay = ns(weeklyDom)
-
-    tags  = ['releaseName', 'targetCMTCONFIG', 'startTime1', 'endTime1','statusText']
-    rtags = ['reportPerson', 'afsReport', 'buildReport','runStatus', 'clusterStatus', 'commentReport']
-
-    format = '%14s %20s %14s %14s %s'
-    for d in orderedKeys:
-        print '\n------------%s-------------\n\n' % d
-        rpd = runsByDay[d]
-        if rpd:
-            print format %  tuple(tags)
-            print
-        for r in rpd:
-            vals = []
-            for t in tags:
-                val = Evaluate(t+'/text()', r)
-                vals.append(val.strip())
-            print format % tuple(vals)
-
-            if longReport:
-                reports = Evaluate('report', r)
-                if reports:
-                    report = reports[0]
-                    print
-                    for t in rtags: 
-                        val = Evaluate(t+'/text()', report)
-                        val = val.strip()
-                        print '   %20s  %s' % (t, val)
-
-# --------------------------------------------------------------
-
-def doit(longReport, writeOut, disp):
-    fn = '/afs/cern.ch/atlas/project/RTT/prod/webPageCollector/src/summaryFilePaths.txt'
-
-    weeklyViewDom = makeRTTDOM('weeklyReport')
-
-    overviewNodes = getOverviewNodes(fn, weeklyViewDom)
-    # print 'found %d doms' % len(doms)
-
-    fn = 'RTTreports.xml'
-    sortedReports = getReports(fn)
-    
-    weeklyViewDom = makeWeeklyViewDom(overviewNodes, sortedReports, weeklyViewDom)
-    # print weeklyViewDom.toxml()        
-
-    # print weekDom.toprettyxml()
-    # if disp: display(weeklyViewDom, longReport)
-
-    if writeOut: open('weekly.xml', 'w').write(weeklyViewDom.toxml())
-    print 'Done.'
-
-# --------------------------------------------------------------    
-
-def usage():
-    print
-    print 'dailyRTT -- displays summaries of all RTT runs for the last week and optionally the checker reports'
-    print
-    print 'usage: python dailyRTT [options]'
-    print
-    print 'Options to *turn off* features'
-    print '   -w do not write out summary file'
-    print '   -l do not display checker reports  (condensed output)'
-    print '   -d do not display on standard out'
-    print '   -h print this message and exit'
-    print
-    print 'The script should be run in a dictory containg the files summaryFilePaths.txt (a list of paths to'
-    print 'the RTT summary files, typically created y webpagecollector) and RTTreports.xml, a file containing'
-    print ' the checkers reports'
-    
-    return
-
-# --------------------------------------------------------------
-
-if __name__ == '__main__':
-    
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'wldh')
-        print 'opts',opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    disp     = True
-    longDisp = True
-    writeOut = True
-
-    for o, a in opts:
-        if o == '-l': longDisp = False
-        if o == '-w': writeOut = False
-        if o == '-d': disp     = False
-        if o == '-h':
-            usage()
-            sys.exit(0)
-
-
-    doit(longDisp, writeOut, disp)
diff --git a/Tools/RunTimeTester/releaseTools/dataCopierWatcher.py b/Tools/RunTimeTester/releaseTools/dataCopierWatcher.py
deleted file mode 100644
index cb4ef200d7679f6bbc88acff6eb163b2e53e9893..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/dataCopierWatcher.py
+++ /dev/null
@@ -1,135 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os, os.path, time
-from datetime import datetime
-
-# *********************************************************************************
-# Author  : B.Simmons
-# Date    : 29th August 2007
-# Purpose : Checks two aspects of the copier:
-#              a) that the last copier report is recent (indicates the copier is running)
-#              b) that the proxy time left is great enough (else sends mail)
-#           Reads in the datacopier report to figure both these out.
-# Usage   : python dataCopierWatcher.py
-# *********************************************************************************
-
-# --- Global variables
-
-copierReport = '/afs/cern.ch/atlas/project/RTT/Results/DatasetManager.report'
-copierReportMaxAge = 1 # copier report should be no more than 1 day old
-proxyMinTimeLimit = 24*5 # 5 days advanced warning of proxy expiry
-
-# ----------------------------------------------------------------
-
-def sendEmail(emailTitle, emailContent):
-    """A simple mailer"""
-    if not emailContent.strip():
-        print 'All OK, no email was sent.'
-        return
-    
-    def at(): return '@'
-    def dot(): return '.'
-    
-    to = 'brinick' + dot() + 'simm' + 'ons' + at() + 'ce' + 'rn' + dot() + 'c' + 'h'
-    
-    subject = emailTitle
-    command = 'mail ' + str(to) + ' -s ' + '"' + str(subject) + '"'
-    os.popen(command,'w').write(emailContent)
-    print 'An email has been sent to %s' % to
-    print '------------------'
-    
-# ----------------------------------------------------------------
-
-def readCopierReport():
-    try:
-        handle = open(copierReport)
-        conts = handle.readlines()
-        handle.close()
-    except Exception, e:
-        title   = 'WARNING: unable to read in the data copier report!'
-        message = str(e)
-        sendEmail(title, message)
-        print 'Unable to read in the data copier report.'
-        print str(e)
-        sys.exit(0)
-
-    return conts
-
-# ----------------------------------------------------------------
-
-def grabProxyInfo(conts):
-    date = grabDateOfReport(conts)
-    timeLeftInHours, timeLeftInDays = grabProxyTimeLeft(conts)
-    return (date, timeLeftInHours, timeLeftInDays)
-
-# ----------------------------------------------------------------
-
-def grabDateOfReport(conts):
-    for line in conts:
-        if line.find('DataSetManager Report')!=-1:
-            date = line[line.find('DataSetManager Report')+len('DataSetManager Report '):].strip()
-            return date
-    return ''
-
-# ----------------------------------------------------------------
-
-def grabProxyTimeLeft(conts):
-    for line in conts:
-        if line.find('Proxy time left in Hours')!=-1:
-            timeLeftInHours  = int(line.split(':')[2])
-            timeLeftInDays = str(timeLeftInHours/24.0)
-            timeLeftInDays = timeLeftInDays[:timeLeftInDays.find('.')+2]            
-            return (timeLeftInHours, timeLeftInDays)
-    return ('', '')
-
-# ----------------------------------------------------------------
-
-def isDateTooOld(theDate):
-    months = {'Jan':1,'Feb':2,'Mar':3,'Apr':4,'May':5,'Jun':6,'Jul':7, 'Aug':8, 'Sep':9, 'Oct':10, 'Nov':11, 'Dec':12}
-    dDayName, dMonth, dDay, dTime, dYear = theDate.split()
-    ddMonth = months[dMonth]
-    dateInst = datetime(int(dYear),ddMonth,int(dDay))
-
-    nowTuple = time.gmtime()
-    nYear  = nowTuple[0]
-    nMonth = nowTuple[1]
-    nDay   = nowTuple[2]
-    nowInst = datetime(nYear,nMonth,nDay)
-
-    timeDeltaInst = nowInst - dateInst
-    return timeDeltaInst.days > copierReportMaxAge
-
-# ----------------------------------------------------------------
-
-def checkDataCopierReport():
-    conts = readCopierReport()
-    date = grabDateOfReport(conts)
-    tooOld = isDateTooOld(date)
-    if tooOld:
-        title   = 'WARNING: data copier manager report out of date!'
-        message = 'Data copier manager report seems out of date (last stamp: %s). Is copier working?' % date
-        print 'Checking data copier report age.....FAILED (last report from: %s)' % date
-        sendEmail(title, message)
-    else:
-        print 'Checking data copier report age.....PASSED OK' 
-        
-# ----------------------------------------------------------------
-
-def checkDataCopierProxy():
-    conts = readCopierReport()
-    dateOfReport, proxyTimeHours, proxyTimeDays = grabProxyInfo(conts)
-    if proxyTimeHours < proxyMinTimeLimit:
-        title   = 'WARNING: data copier proxy expires soon!'
-        message = 'At %s, proxy had only %s days left (%d hours)' % (dateOfReport, proxyTimeDays, proxyTimeHours)
-        print 'Checking data copier proxy time left.....FAILED (only %d hours left)' % proxyTimeHours
-        sendEmail(title, message)
-    else:
-        print 'Checking data copier proxy time left.....PASSED OK'
-
-# ----------------------------------------------------------------
-# ----------------------------------------------------------------
-# ----------------------------------------------------------------
-
-if __name__ == '__main__':
-    checkDataCopierReport()
-    checkDataCopierProxy()
diff --git a/Tools/RunTimeTester/releaseTools/dumpLastLine.py b/Tools/RunTimeTester/releaseTools/dumpLastLine.py
deleted file mode 100755
index 8e69c1ee10fd10b553f43a88a9f24fc14e971420..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/dumpLastLine.py
+++ /dev/null
@@ -1,102 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, sys, fnmatch, getopt
-from time import strftime, localtime
-
-from findMostRecentLogDir import findMostRecentLogDir
-
-def dumpLastLine(root, packages):
-    try:
-        recentLogDir =  findMostRecentLogDir(root)
-    except:
-        print 'cannot find log files below ', root
-        sys.exit(0)
-
-
-    print 'latest log directory',recentLogDir
-    dirs = os.listdir(recentLogDir)
-    logdirs = []
-    [logdirs.extend(fnmatch.filter(dirs,p)) for p in packages]
-    logdirs = [os.path.join(recentLogDir, l) for l in logdirs]
-    logdirs = [l for l in logdirs if os.path.isdir(l)]
-
-    nlines = []
-
-    for logdir in logdirs:
-        print 'examining ',logdir
-
-        logs = os.listdir(logdir)
-        for f in logs :
-            fn = os.path.join(logdir,f)
-            ifile = open(fn, 'r')
-            lines = ifile.readlines()
-            ifile.close()
-            nlines.append( (os.stat(fn).st_atime, lines[-1], fn) )
-        
-    print 'found %d lines' % len(nlines)
-
-    nlines.sort()
-    for l in nlines:
-        print '\n%s\n%s: %s\n' % (l[2], strftime("%d/%m %H:%M:%S", localtime(l[0])), l[1])
-
-                                  
-
-def usage():
-    print
-    print 'dumpLastLine -- dumps the time ordered last line for all log files in an RTT run.'
-    print
-    print 'usage: python dumpLastLine [options]'
-    print
-    print 'Options'
-    print '   -p colon separated list of packages to examine defaults to "*"'
-    print '   -l use the most recent version of the log files. must supply arg slc3 or slc4'
-    print '      if ommitted, the path to the log files will be taken from the config file.'
-    print '   -f triggers writing to a file. Filename is written to stdout'
-    print '   -u set the user, eg  brinick, eric, peter, prod, Default prod'
-    print '   -h print this message and exit'
-    return
-
-if __name__ == '__main__':
-
-
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'p:l:fu:h')
-        print opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    
-    useLatest =  False
-    packages  =  '*'
-    slc       =  None
-    output    =  None
-    user      =  'prod'
-    
-    for o, a in opts:
-        if o == '-l':
-            if a == 'slc3':
-                slc = 'SLC3'
-            elif a == 'slc4':
-                slc = 'SLC4_32'
-        elif o == '-u':
-            user = a
-        elif o == '-p':
-            packages = a.split(':')
-        elif o == '-h':
-            usage()
-            sys.exit(0)
-        if o == '-f':
-            output  = '/afs/cern.ch/user/r/rtt/Scratch'
-
-    frag = ''
-    if user == 'prod':
-        frag = os.path.join(user, slc, 'logs')
-    else:
-        frag = os.path.join('data', user, 'logs', slc, 'logs')
-    
-    root = '/afs/cern.ch/atlas/project/RTT/%s' %frag
-    # packages = ['TestAtlfast', 'BPhysValidation']
-    #packages = ['*']
-
-    dumpLastLine(root, packages)
diff --git a/Tools/RunTimeTester/releaseTools/dumpMinderTag.py b/Tools/RunTimeTester/releaseTools/dumpMinderTag.py
deleted file mode 100755
index 76317034a3a9b410ce166bc60254f8cb7d10d46c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/dumpMinderTag.py
+++ /dev/null
@@ -1,129 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import xml.dom.minidom, sys, getopt, os
-#import xml.dom
-
-
-"""
-Replace xpath version Evaluate because I (PS)cant get xpath to work at cern
-for the moment (14/6/05).
-"""
-
-from xml.dom.minidom     import parse
-import os.path
-sys.path.append('../src')
-from Evaluate import Evaluate
-from exc2string2 import exc2string2
-
-#def Evaluate(path, element):
-#    paths = path.split('/')
-#/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc3-gcc323-opt/RTTSummary.xml
-#    curGenEls = [element]
-#    for p in paths:
-#        nextGenEls = []
-#        if p == 'text()':
-#            texts = []
-#            [texts.extend(getText(e)) for e in curGenEls]
-#            text = reduce(lambda x,y:x+y, texts,'')
-#            return text
-#        else:
-#            # [nextGenEls.extend(e.getElementsByTagName(p)) for e in curGenEls]
-#            [nextGenEls.extend(e.childNodes) for e in curGenEls]
-#            nextGenEls = [n for n in nextGenEls if n.nodeName == p]
-#            # print p, len(nextGenEls), str([n.nodeName for n in nextGenEls])
-#            
-#        curGenEls = nextGenEls
-#    return curGenEls
-#def getText(element):
-#    texts = [node.data for node in element.childNodes if node.nodeType == node.TEXT_NODE]
-#    text = reduce(lambda x,y:x+y, texts, '')
-#    return text
-  
-
-def processPackage(node, tagnames):
-    packagename =  Evaluate('packageName/text()', node).strip()
-    jobs =  Evaluate('minder', node)
-
-    res = '-------- '+packagename.ljust(30)+'-------------\n\n'
-    for j in jobs:
-        res += '   '+Evaluate('identifiedName/text()', j).strip()+'\n\n'
-        texts =  [(tagname.ljust(20),Evaluate(tagname+'/text()', j).strip('\n\n')) for tagname in tagnames]
-        for t in texts:
-            res +=  '      %20s %s\n' % t
-    return res
-#print len(states)
-
-
-def usage():
-    print 'Name - dumpMinderTag.py  -- dump minder text for user provided xml tags'
-    print 'Synopsis python dumpMinderTag -r release -s slc version -t tags [-f] [-h]'
-    print '  -r Mandatory. release number in range [0,6]'
-    print '  -s Mandatory. slc3 version (slc3 or slc4)'
-    print '  -t mandatory. colon separated list of tags tag1:tag2:...'
-    print '  -f write to file to program provided location'
-    print '  -h print  this help message and exit'
-if __name__ =='__main__':
-    
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'r:s:t:fh')
-    except:
-        usage()
-        sys.exit(2)
-        
-    cmtTarget = None
-    release   = None
-    tags      = []
-    of        = False
-    
-    for o,a in opts:
-        if o == '-r':
-            if a in ['0', '1', '2', '3', '4', '5', '6']:
-                release = 'rel_'+a
-        if o == '-s':
-            if a == 'slc3': cmtTarget = 'i686-slc3-gcc323-opt'
-            if a == 'slc4': cmtTarget = 'i686-slc4-gcc34-opt'
-
-        if o == '-t':
-            tags = a.split(':')
-
-        if o == '-f':
-            of = True
-        if o == '-h':
-            of = usage()
-            sys.exit(0)
-
-    
-    if (not cmtTarget or not release or not tags):
-        usage()
-        sys.exit(2)
-
-    fn='/afs/cern.ch/atlas/project/RTT/Results/%s/dev/build/%s/RTTSummary.xml' % (release, cmtTarget)
-
-    print '\nexamining %s \n' % fn
-
-    if not os.path.exists(fn):
-        print 'no summary file %s, exiting' % fn
-        sys.exit(0)
-
-    print 'Will look for the tags:'
-    for t in tags: print '   ',t
-    
-    if of:
-        of = '/afs/cern.ch/user/r/rtt/Scratch/dumpMinderTag_%s_%s.log' % (release, cmtTarget)
-        print 'Output will be written to ', of
-        
-    try:
-        dom = xml.dom.minidom.parse(fn)
-    except:
-        print '\n\nCould not parse file', fn
-        print exc2string2()
-        sys.exit(2)
-        
-    packages = dom.getElementsByTagName('package')
-
-    if of:
-        outf = open(of, 'w')
-    else:
-        outf = sys.stdout
-        
-    [outf.write(processPackage(p, tags)) for p in packages]
diff --git a/Tools/RunTimeTester/releaseTools/findMostRecentLogDir.py b/Tools/RunTimeTester/releaseTools/findMostRecentLogDir.py
deleted file mode 100755
index 244a957acd5460847d58a1e4a77d1a0f6517c12b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/findMostRecentLogDir.py
+++ /dev/null
@@ -1,12 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-
-# example root: '/afs/cern.ch/atlas/project/RTT/prod/Peter/logs'
-
-def findMostRecentLogDir(root):
-
-    logdirs = [os.path.join(root,l) for l in os.listdir(root)]
-    logdirs = [(os.stat(l).st_atime, l) for l in logdirs]
-    logdirs.sort()
-    return logdirs[-1][1]
diff --git a/Tools/RunTimeTester/releaseTools/kit_install_config.cfg b/Tools/RunTimeTester/releaseTools/kit_install_config.cfg
deleted file mode 100644
index fbee139fd7c71d605036fee6d1c67884cac655c5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/kit_install_config.cfg
+++ /dev/null
@@ -1,7 +0,0 @@
-pacmanArea = /unix/atlas1/software/pacman/pacman-3.19
-release    = rel_3	
-dest       = /unix/atlasrtt/eric
-branch     = bugfix
-cmtconfig  = i686-slc4-gcc34-opt
-todayNightly = True
-
diff --git a/Tools/RunTimeTester/releaseTools/listFiles.py b/Tools/RunTimeTester/releaseTools/listFiles.py
deleted file mode 100755
index 98640a4ed0c6b5ab0d792965dc0a88bb3c2ec25c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/listFiles.py
+++ /dev/null
@@ -1,33 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#Python cookbook 4.18
-
-import os.path, fnmatch
-
-def listFiles(root, patterns='*', recurse=True, return_folders=False):
-    pattern_list = patterns.split(';')
-
-    class Bunch:
-        def __init__(self, **kwds): self.__dict__.update(kwds)
-
-    arg = Bunch(recurse=recurse, pattern_list=pattern_list, return_folders=return_folders, results = [])
-
-    def visit(arg, dirname, files):
-
-        for name in files:
-            
-            fullname= os.path.normpath(os.path.join(dirname, name))
-            #print fullname
-            if arg.return_folders or os.path.isfile(fullname):
-                for pattern in arg.pattern_list:
-                    if fnmatch.fnmatch(name, pattern):
-                        arg.results.append(fullname)
-                        break
-                    
-        if not arg.recurse: files[:] = []
-
-    #print root
-    #print pattern_list
-    os.path.walk(root, visit, arg)
-            
-    return arg.results
diff --git a/Tools/RunTimeTester/releaseTools/logFileChecker.py b/Tools/RunTimeTester/releaseTools/logFileChecker.py
deleted file mode 100755
index 518595ae64153f0d48f160683eac9ebb04a522ac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/logFileChecker.py
+++ /dev/null
@@ -1,191 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-import os.path, sys, getopt
-from findMostRecentLogDir import findMostRecentLogDir
-
-def exceptionTypes():
-    import exceptions
-    return [e for e in dir(exceptions) if e.find('__')==-1]
-
-def stripstr(s):
-    return s.strip()[:-1]
-
-class Checker:
-    def __init__(self, strings):
-        self.strings = strings
-        self.dict = {}
-        self.fdict = {}
-        self.sdict = {}
-        for s in self.strings:
-            self.dict[s]=0
-            self.fdict[s]=[]
-
-        self.ncalls = 0
-        
-    def checkForStrings(self,f):
-        self.ncalls += 1
-        lines = open(f,'r').readlines()
-        self.sdict[f] = []
-        for s in self.strings:
-            for l in lines:
-                # print "\n looking for:\n%s\n in line\n %s" %(s,l)
-                if l.find(s)>= 0:
-                    self.dict[s]+=1
-                    if f not in self.fdict[s]: self.fdict[s].append(f)
-                    self.sdict[f].append(s)
-
-
-
-def logFileChecker(root, globpar, strings):
-
-    res =  'looking for files below %s matching %s\n' % (root, globpar)
-    logfiles = listFiles(root, globpar)
-    res +=  'Found %d log files\n' % len(logfiles)
-
-    checker = Checker(strings)
-
-    [checker.checkForStrings(f) for f in logfiles]
-
-    res +=  '\nerror strings:\n'
-    dict = checker.dict
-
-    res += 'Occurences   text\n'
-
-    for s in dict:
-        if dict[s]>0: res += '%10d   %s\n' % (dict[s], s)
-
-    res += '\n Strings searched for:\n'
-    for s in dict.keys(): res += s+'\n'
-
-    for s in checker.fdict.keys():
-        if len(checker.fdict[s]):
-            res +=  '\n\n"%s" was found in %d files:\n' % (s, len(checker.fdict[s]))
-            for f in checker.fdict[s]: res +=  f+'\n'
-        else:
-            res +=  '\n"%s" was found in no files\n\n' % s
-
-    res +=  '\n\n%d files with no strings:\n\n' % (checker.sdict.values()).count([])
-
-    for f in checker.sdict.keys():
-        if not checker.sdict[f]: res += f+'\n'
-
-    return res
-
-def readConfig(config):
-
-    lines  = open(config, 'r').readlines()
-
-    toRemove = [l for l in lines if l.startswith('#')]
-    [lines.remove(l) for l in toRemove]
-    lines = [stripstr(l) for l in lines]
-    toRemove = [s for s in lines if len(s)==0]
-    [lines.remove(l) for l in toRemove]
-
-    return lines
-
-if __name__ == '__main__':
-
-    def usage():
-        print
-        print 'logFileChecker -- checks for strings in files'
-        print
-        print 'usage: python logFileChecker [options]'
-        print
-        print 'Options'
-        print '   -c filename name of file with dir, glob parameters and strings. Mandatory'
-        print '   -l use the most recent version of the log files. must supply arg branch or branch:os'
-        print '      possible branch = ("bugfix", "dev", "val"), possibe os = ("slc3", "slc4")'
-        print '      example: -l dev or -l dev:slc3'
-        print '      if ommitted, the path to the log files will be taken from the config file.'
-        print '   -f triggers writing to a file. Filename is written to stdout'
-        print '   -u set the user, eg  brinick, eric, peter, prod, Default prod'
-        print '   -h print this message and exit'
-        return
-    
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'c:b:l:fu:h')
-        print opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    useLatest = False
-    config    = None
-    slc       = None
-    output    = None
-    branch    = None
-    user      = 'prod'
-    for o, a in opts:
-        if o == '-l':
-            tokens = a.split(':')
-            
-            branch = tokens[0]
-            if branch not in ['bugfix', 'dev', 'val']:
-                
-                usage()
-                sys.exit(2)
-                
-            if len(tokens)>1:
-                opsys = tokens[1]
-            if opsys not in ['slc3', 'slc4']:
-                usage()
-                sys.exit(2)
-                
-            if opsys == 'slc3':
-                opsys = 'SLC3'
-            elif opsys == 'slc4':
-                opsys = 'SLC4_32'
-
-
-        elif o == '-u':
-            user = a+'/logs'
-            if user != 'prod': user = 'data/'+user
-        elif o == '-h':
-            usage()
-            sys.exit(0)
-        if o == '-c':
-            config = a
-        if o == '-f':
-            output = '/afs/cern.ch/user/r/rtt/Scratch'
-
-    if not config:
-        print '-c option required'
-        usage()
-        sys.exit(2)
-    if not os.path.exists(config):
-        print
-        print 'unknown config file',config,' exiting'
-        sys.exit(2)
-
-    # config = open(sys.argv[1], 'r')
-    lines = readConfig(config)
-    if branch:
-        root = '/afs/cern.ch/atlas/project/RTT/%s/logs/%s/%s/logs' % (user, branch, opsys)
-
-        print 'looking for most recent log file directory below %s' %  root
-        root = findMostRecentLogDir(root)
-        print '  ... found %s' %  root
-    else:
-        root = lines[0]
-
-
-    globpar = lines[1]
-    strings = lines[2:]
-
-    if not os.path.exists(root):
-        print 'Balking as root directory does not exist:\n%s' % root
-        sys.exit(0)
-
-
-    if output:
-        cname = config.split('.')[0]
-        cname += '_'+os.path.basename(root)+'.log'
-        output = os.path.join(output, cname)
-        of = open(output,'w')
-        print 'writing results to ', output
-    else:
-        of = sys.stdout
-
-
-    of.write(logFileChecker(root, globpar, strings))
diff --git a/Tools/RunTimeTester/releaseTools/logFileGrepper.py b/Tools/RunTimeTester/releaseTools/logFileGrepper.py
deleted file mode 100755
index 1eb1a3ebaa3f4793b8b052da9aceae98ccaa1ff4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/logFileGrepper.py
+++ /dev/null
@@ -1,80 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-import os.path, sys
-
-def stripstr(s):
-    s.strip()
-    s = s[:-1]
-    return s
-class Checker:
-    def __init__(self, f, strings):
-        self.strings = strings
-        self.f = f
-        self.dict = {}
-        self.fdict = {}
-        for s in self.strings:
-            self.dict[s]=[]
-
-        self.ncalls = 0
-        self.checkForStrings()
-        
-        
-    def checkForStrings(self):
-        self.ncalls += 1
-        lines = open(self.f,'r').readlines()
-        for s in self.strings:
-            for l in lines:
-                # print "\n looking for:\n%s\n in line\n %s" %(s,l)
-                if l.find(s)>= 0:
-                    self.dict[s].append(l)
-
-    def __str__(self):
-        m = '\n File: %s' % self.f
-
-        nostrings = True
-        for s in self.strings:
-            if self.dict[s]:
-                nostrings = False
-
-        if nostrings:
-            m = 'File: %s: no strings found' % self.f
-            return m
-        
-
-        m = '\n File: %s\n' % self.f
-            
-        
-        
-        for s in self.strings:
-            if self.dict[s]:
-                m += '\n   string: %s\n' %  s
-                for l in self.dict[s]:
-                    m += '      %s' % l
-
-        return m
-
-    def dump(self):
-        print self
-
-config = open(sys.argv[1], 'r')
-lines  = config.readlines()
-config.close()
-
-toRemove = [l for l in lines if l.startswith('#')]
-[lines.remove(l) for l in toRemove]
-lines = [stripstr(l) for l in lines]
-toRemove = [s for s in lines if len(s)==0]
-[lines.remove(l) for l in toRemove]
-root = lines[0]
-globpar = lines[1]
-strings = lines[2:]
-
-
-print 'looking for files below', root,'matching',globpar
-logfiles = listFiles(root, globpar)
-print 'Found %d files' % len(logfiles)
-
-
-checkers = [Checker(f, strings).dump() for f in logfiles]
-
diff --git a/Tools/RunTimeTester/releaseTools/logFileSizeChecker.py b/Tools/RunTimeTester/releaseTools/logFileSizeChecker.py
deleted file mode 100755
index 3e9f331fc9cae86fb4cde87cfb96342f5bc13540..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/logFileSizeChecker.py
+++ /dev/null
@@ -1,38 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-import os
-
-class Checker:
-    def __init__(self):
-        self.dict    = {}
-        self.sumsize = 0
-        self.ncalls  = 0
-        
-    def checkSize(self,f):
-        self.ncalls += 1
-        size = int((os.stat(f)).st_size)
-        self.dict[os.path.basename(f)] = size
-        self.sumsize += size
-
-root = '/afs/cern.ch/atlas/project/RTT/Work/rel_4/dev/build/i686-slc4-gcc34-opt'
-
-print 'looking for files in %s' % root
-logfiles = listFiles(root, '*_log')
-print 'Found %d log files' % len(logfiles)
-
-checker = Checker()
-
-[checker.checkSize(f) for f in logfiles]
-
-print 'Root directory = ', root
-print '\nlog file sizes:'
-
-for s in checker.dict.items():
-    print '%10s   %d' % s
-
-print 'log files below:', root
-print 'No of files:',len(checker.dict.keys())
-print 'Total size (kbytes):',checker.sumsize/1024.
-
-
diff --git a/Tools/RunTimeTester/releaseTools/printDBkeys.py b/Tools/RunTimeTester/releaseTools/printDBkeys.py
deleted file mode 100755
index bdca8c82fa7fbf9c8b16e42e2eab8f4e09c6d9be..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/printDBkeys.py
+++ /dev/null
@@ -1,65 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os, os.path
-import anydbm
-
-class DBKeyPrinter:
-    def __init__(self):
-        self.handleArgs()
-        self.dbKeys = self.getAllDBkeys()
-
-    def usage(self):
-        print
-        print 'printDBkeys: prints all DB keys that match the given criteria'
-        print
-        print 'Usage:'
-        print '     python printDBkeys.py <pathToDB> "selectionKey1,selectionKey2,..."'
-        print 'where:'
-        print '   <pathToDB>: obligatory'
-        print '   "selectionKey1,selectionKey2,...": optional. If missing = print all keys.'
-        print
-        
-    def getDBpath(self):
-        try:
-            self.dbPath = sys.argv[1]
-        except:
-            self.usage()
-            sys.exit(2)
-        else:
-            if not os.path.exists(self.dbPath):
-                print 'DB %s does not exist.'
-                sys.exit(2)
-
-    def getDBSelectorKeys(self):
-        try:
-            keys = sys.argv[2]
-        except:
-            self.selectorKeys = []
-        else:
-            self.selectorKeys = [t.strip() for t in keys.split(',')]
-            
-    def handleArgs(self):
-        self.getDBpath()
-        self.getDBSelectorKeys()
-
-    def getAllDBkeys(self):
-        h = anydbm.open(self.dbPath,'r')
-        keys = h.keys()
-        h.close()
-        return keys
-
-    def select(self):
-        matches = 0
-        for k in self.dbKeys:
-            doNotMatchSelectorKey = [s for s in self.selectorKeys if str(k).find(s)==-1]
-            if len(doNotMatchSelectorKey) != len(self.selectorKeys) or len(self.selectorKeys)==0:
-                matches += 1
-                print k
-                print
-        print '----------'
-        print '%d matching DB keys.' % matches
-
-if __name__ == '__main__':
-    dbkp = DBKeyPrinter()
-    dbkp.select()
-    
diff --git a/Tools/RunTimeTester/releaseTools/scavenger.py b/Tools/RunTimeTester/releaseTools/scavenger.py
deleted file mode 100755
index 521ae1ee6b4c76dfd577052032947ce2777b2efc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/scavenger.py
+++ /dev/null
@@ -1,93 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, os.path
-from popen2 import Popen4
-
-def getPopen4Output(cmd):
-    po = Popen4(cmd)
-    out = []
-    while po.poll()==-1:
-        out.extend(po.fromchild.readlines())
-    del po
-    return out
-
-# ==========================================================================
-
-def sendEmail(emailContent):
-    """A simple mailer"""
-    if not emailContent.strip():
-        print 'All OK, no email was sent.'
-        return
-
-    def at(): return '@'
-    
-    to = 'rtt' + at() + 'he' + 'p.uc' + 'l.ac' + '.uk'
-
-    subject = 'WARNING: scavenger found problems with RTT cluster'
-    command = 'mail ' + str(to) + ' -s ' + '"' + str(subject) + '"'
-    os.popen(command,'w').write(emailContent)
-    print '------------------'
-    print 'An email has been sent to %s' % to
-
-# ==========================================================================
-
-def checkWPC():
-    print 'Running checkWPC()'
-    procs = getPopen4Output('ps -elf | grep python | grep WebPageCollector.py')
-    if len(procs) == 1:
-        machine = getPopen4Output('uname -n')[0]
-        print 'Problems found.'
-        return 'WebPageCollector process not ongoing on machine %s' % machine
-        
-    return ''
-
-# ==========================================================================
-
-def checkQuotas(reportAtGreaterThan):
-    print 'Running checkQuotas()'
-    def getQuotaUsage(path):
-        out = getPopen4Output('fs lq %s' % path)
-        volume = out[1].split()[0]
-        quota  = out[1].split()[1]
-        used   = out[1].split()[2]
-        frac   = out[1].split()[3]
-        frac   = frac[:frac.find('%')]
-        return (path, volume, quota, used, frac)
-
-    message = ''
-    resBase = '/afs/cern.ch/atlas/project/RTT/Results'
-    workBase = '/afs/cern.ch/atlas/project/RTT/Work'
-    pathsToCheck = ['/afs/cern.ch/user/r/rtt/',
-                    '/afs/cern.ch/atlas/project/RTT/prod',
-                    resBase,
-                    workBase,
-                    '/afs/cern.ch/atlas/project/RTT/data',
-                    '/afs/cern.ch/atlas/project/RTT/source']
-
-    # now check run quotas (each on a separate AFS volume)
-    for rel in ['rel_0','rel_1','rel_2','rel_3','rel_4','rel_5','rel_6',]:
-        for branch in ['dev','bugfix','val', 'pcache', 'devval']:
-            for config in ['i686-slc4-gcc34-opt','i686-slc4-gcc34-dbg','i686-slc3-gcc323-opt']:
-                resDir = os.path.join(resBase, '%s/%s/build/%s' % (rel, branch, config))
-                wrkDir = os.path.join(workBase, '%s/%s/build/%s' % (rel, branch, config))
-                if os.path.exists(resDir):
-                    pathsToCheck.append(resDir)
-                if os.path.exists(wrkDir):
-                    pathsToCheck.append(wrkDir)
-
-    results = [getQuotaUsage(path) for path in pathsToCheck]
-    for path, volName, tot, used, frac in results:
-        if int(frac) >= reportAtGreaterThan:
-            message += 'Volume %s (%s) is %s%% used\n' % (volName, path, frac)
-            print 'Problems found.'
-    return message
-
-# ==========================================================================
-
-if __name__ == '__main__':
-    message = ''
-    message += checkQuotas(90) # % usage over which raise alarm
-    message += checkWPC() # check WPC process is ongoing, if not raise alarm
-    sendEmail(message)
-
-    
diff --git a/Tools/RunTimeTester/releaseTools/vmstatLines.py b/Tools/RunTimeTester/releaseTools/vmstatLines.py
deleted file mode 100755
index a8720ae1ed7fa0600e36ab651f233f1413c18eeb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/releaseTools/vmstatLines.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-
-root='/afs/cern.ch/atlas/project/RTT/prod/SLC4_32/logs/logfiles_20feb_08:38'
-fn = os.path.join(root, 'RTT.log')
-ilines = open(fn, 'r').readlines()
-olines1 = []
-olines2 = []
-ind = 0
-for i in ilines:
-    if i.find(' active ')>-1:
-        olines1.extend(ilines[ind-1:ind+2])
-    if i.find(' cache ')>-1:
-        olines2.extend(ilines[ind-1:ind+2])
-    ind +=1
-
-olines1 = [ i[:17]+i[43:] for i in olines1]
-olines2 = [ i[:17]+i[43:] for i in olines2]
-of = open('vmstatLines1.txt', 'w')
-[of.write(l) for l in olines1]
-of = open('vmstatLines2.txt', 'w')
-[of.write(l) for l in olines2]
diff --git a/Tools/RunTimeTester/share/AtlasEnvSetup.py b/Tools/RunTimeTester/share/AtlasEnvSetup.py
deleted file mode 100644
index b651e37756684baae2825203be203dbc4256eeea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/AtlasEnvSetup.py
+++ /dev/null
@@ -1,52 +0,0 @@
-from Logger            import Logger
-from ShellCommand import ShellCommand
-# from Factory_CmtLinesForScripts import CmtLinesForScriptsFactory
-import os
-import os.path
-
-class AtlasEnvSetup:
-    def __init__(self, argDict):
-        self.desc          = argDict['JobDescriptor']
-        self.RTTShare      = self.desc.paths.RTTLibDir
-        self.RTTSrc        = self.desc.paths.RTTSrcDir
-        self.userScript    = argDict['userScript']
-        self.userScriptLoc = argDict['userScriptLoc']
-
-        self.logger = Logger()
-
-        picklefile = os.path.join(self.desc.runPath, 'rtt.argdict.cpickle')
-
-
-        # self.cmds = CmtLinesForScriptsFactory(self.logger).create(self.desc).lines()
-        self.cmds = argDict['cmtLinesCmds']
-        self.cmds.append('cd %s;python UserScriptRunner.py %s' % (self.desc.runPath, picklefile))
-
-        __module__ = self.userScript+'.py'
-        __name__ = self.userScript
-
-    def run(self):
-        self.createRunnerScript() # dump user runner script into runpath
-        sc = ShellCommand(self.cmds, self.logger)
-        self.logger.debug(sc.getRawReply())
-
-    def createRunnerScript(self):
-        def readTemplate():
-            f = os.path.join(self.RTTShare, 'UserScriptRunnerTemplate.py')
-            h = open(f)
-            conts = h.read()
-            h.close()
-            return conts
-        
-        def outputFile(conts):
-            f = os.path.join(self.desc.runPath, 'UserScriptRunner.py')
-            h = open(f, 'w')
-            h.write(conts)
-            h.close()
-            self.logger.debug('Output file: %s' % f)
-
-        conts = readTemplate()
-        conts = conts.replace('USER_MODULE_NAME', self.userScript)
-        conts = conts.replace('%RTT_SHARE_DIR%', self.RTTShare)
-        conts = conts.replace('%RTT_SRC_DIR%', self.RTTSrc)
-        conts = conts.replace('%USER_MODULE_LOCATION%', self.userScriptLoc)
-        outputFile(conts)
diff --git a/Tools/RunTimeTester/share/CheckFileRunner.py b/Tools/RunTimeTester/share/CheckFileRunner.py
deleted file mode 100644
index 56f771e675522b89e0062bb50fd4810c89c9f296..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/CheckFileRunner.py
+++ /dev/null
@@ -1,60 +0,0 @@
-from Logger import Logger
-from ShellCommand import ShellCommand
-import os
-import os.path
-
-class CheckFileRunner:
-    def __init__(self, argDict):
-        desc  = argDict.get('JobDescriptor')
-        paths = desc.paths
-        self.runPath = desc.runPath
-        self.logger  = Logger()
-        self.envSetupCmds = argDict['cmtLinesCmds']
-
-    def checkFileStatus(self, lines):
-        for line in lines:
-            if line.find('RTT determined that checkFile exit status was:') != -1:
-                return int(line.split(':')[1].strip())
-        return 1
-            
-    def run(self):
-        poolFiles = [f for f in os.listdir(self.runPath) if f.endswith('.pool.root')]
-
-        statusCode = 0
-
-        for poolFile in poolFiles:
-            pathToPoolFile = os.path.join(self.runPath, poolFile)
-
-            cmds = []
-            cmds.extend(self.envSetupCmds)
-            cmds.append('checkFile.py -f %s' % pathToPoolFile)
-            cmds.append('echo "RTT determined that checkFile exit status was: " $?')
-
-            # launch the commands
-            sc    = ShellCommand(cmds, self.logger)            
-            reply = sc.getRawReply()
-            [self.logger.debug(str(l)) for l in reply]
-
-            handle = None
-            outFile = poolFile + '.checkFile'
-            outPath = os.path.join(self.runPath, outFile)
-            
-            try:
-                handle = open(outPath, 'w')
-                handle.writelines(reply)
-                handle.close()
-                self.logger.debug('Successfully wrote out: %s' % outPath)
-                status = self.checkFileStatus(reply)
-                if status!=0:
-                    self.logger.warning('CheckFile returned error, non-zero status (%d) for %s' % (status, poolFile))
-                statusCode += status
-            except Exception, e:
-                m  = 'Unable to output results of checkFile to file: %s\n' % outPath
-                m += str(e)
-                self.logger.error(m)
-                if handle: handle.close()
-                return 1
-        return statusCode
-            
-
-
diff --git a/Tools/RunTimeTester/share/DCubeRunner.py b/Tools/RunTimeTester/share/DCubeRunner.py
deleted file mode 100644
index 0aee5a0295ab029f5aa0dae637ed288e899c1fe3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/DCubeRunner.py
+++ /dev/null
@@ -1,376 +0,0 @@
-## 
-# @module DCubeRunner
-# @author Krzysztof Daniel Ciba
-# @date 4 Sep 2007
-# @brief Glue module to plugin DCube to Atlas RTT. 
-#
-# * Feb 4, 2007 - switch to DCubeClient-00-00-13
-#                 and DCubeServer-00-00-08
-# * Apr 10, 2008 - switch to DCubeServer-00-00-10
-# * Nov 7, 2008 - switch to new DCubeServer-00-00-11
-# * Nov 11, 2008 - switch to new DCubeClient-00-00-19 and DCubeServer-00-00-12 
-# 
-
-## some imports 
-import xml.dom.minidom as minidom
-import os, sys
-import os.path 
-import re
-
-## @class DCubeRunner
-#  @brief prepares and runs shell script with DCube command 
-class DCubeRunner:
-
-      ## DCubeClient API classes
-      __reAPIClasses = [ re.compile("DCubeApp"), 
-                         re.compile("DCubeConfig"), 
-                         re.compile("DCubeTester"), 
-                         re.compile("DCubePlotter"), 
-                         re.compile("DCubePHPWriter") ]
-      
-      ## base directory for references
-      # baseRef = "/afs/cern.ch/atlas/project/RTT/Results/reference/"
-
-      ## DCubeServer directory
-      # dcubeServer = "/afs/cern.ch/atlas/project/RTT/Results/dcube/DCubeServer-00-00-12"
-
-      ## DCubeClient directory
-      # dcubeClient = "/afs/cern.ch/atlas/project/RTT/DCubeClient/DCubeClient-00-00-19/python"
-
-
-      ## c'tor
-      # @param self "Me, myself and Irene"
-      # @param argDict RTT action runner dictionary 
-      def __init__( self, argDict={} ):
-
-            try:
-                  from Logger import Logger
-                  self.logger = Logger()
-            except ImportError:
-                  self.logger = False
-
-            self.error = None 
-
-            self.debug("DCubeRunner constructor is getting args from dictionary...")
-
-
-            self.cfgFile = None 
-            if ( "DCubeCfg" in argDict ):
-                  self.cfgFile = argDict["DCubeCfg"]
-            else:
-                  self.info( "*** No DCubeCfg in DCubeRunner action dictionary ***" )
-                  self.error = "No DCubeCfg in DCubeRunner argDict"
-
-            self.refFile = None 
-            if ( "DCubeRef" in argDict ): 
-                  self.refFile = argDict["DCubeRef"]
-            else:
-                  self.info( "*** No DCubeRef in DCubeRunner action dictionay, will use value from config XML file ***" )
-
-            self.monFile = None 
-            if ( "DCubeMon" in argDict ):
-                  self.monFile = argDict["DCubeMon"]
-            else:
-                  self.info( "*** No DCubeMon in DCubeRunner action dictionary ***" )
-                  self.error = "No DCubeMon in DCubeRunner argDict"
-
-            if ( "DCubeJobId" in argDict ):
-                  self.jobId = argDict["DCubeJobId"]
-            else:
-                  self.jobId = None
-
-            jobDtor = argDict.get('JobDescriptor', None )
-
-            if ( jobDtor != None ):
-                  self.DCubeForRTTCfg = jobDtor.paths.dCubeCfgFile
-                  self.dcubeClient    = os.path.dirname(self.DCubeForRTTCfg)
-                  self.outputPath = jobDtor.runPath
-                  self.installArea = jobDtor.paths.installArea
-                  self.cvsPath = 'offline/' + jobDtor.paths.containerPackage
-                  self.cmtPath = jobDtor.paths.cmtPath
-                  self.nightly = jobDtor.paths.branch # branch
-                  self.install = jobDtor.paths.runType #  install
-                  self.cmtconfig = jobDtor.paths.cmtConfig # CMTCONFIG
-                  self.project = jobDtor.paths.topProject # Atlas project name
-            else:
-                  ## dummy values for testing
-                  self.DCubeForRTTCfg = "./DCubeForRTTCfg.xml"
-                  self.outputPath  = "/DUMMY/RUN/PATH/FROM/JOB/DESCRIPTOR"
-                  self.installArea = "/DUMMY/INSTALL/AREA/FROM/JOB/DESCRIPTOR"
-                  self.cvsPath     = "/DUMMY/CONTAINER/PACKAGE/FROM/JOB/DESCRIPTOR"
-                  self.cmtPath     = "/DUMMY/CMT/PATH/FROM/JOB/DESCRIPTOR"
-                  self.nightly     = "dev"
-                  self.install     = "build"
-                  self.cmtconfig   = "i686-slc4-gcc34-opt"
-                  self.project     = "AtlasProduction"
-
-            self.debug("[01] will use '%s' as DCubeClient configuration file for RTT" % self.DCubeForRTTCfg )
-            self.debug("[02] will use '%s' as output path" % self.outputPath )
-            self.debug("[03] will use '%s' as install area path" % self.installArea )
-            self.debug("[04] will use '%s' as package CVS path" % self.cvsPath )
-            self.debug("[05] will use '%s' as cmt path" % self.cmtPath )
-            self.debug("[06] will use DCubeClient monitored file '%s'" % self.monFile )
-            self.debug("[07] will use DCubeClient reference file '%s'" % self.refFile )
-            self.debug("[08] will use DCubeClient configuration file '%s'" % self.cfgFile )
-            self.debug("[09] will use '%s' as DCubeClient branch name" % self.nightly )
-            self.debug("[10] will use '%s' as DCubeClient install name" % self.install )
-            self.debug("[11] will use '%s' as DCubeClient cmtconfig name" % self.cmtconfig )
-            self.debug("[12] will use '%s' as DCubeClient project name" % self.project  )
-            if ( self.jobId ):
-                  self.debug("[13] will use '%s' as DCubeClient jobId" % self.jobId )
-            else:
-                  self.debug("[13] DCubeClient jobId not set!")
-                  
-            self.command = [ ]
-            if ( self.error == None ):
-                  for line in self.__parseDCubeRTTCfg():
-                        self.command.append( line ) 
-            msg = "command to run is:\n"+20*"-"+"\n"
-            for line in self.command:
-                  msg += line
-            msg += "\n"+20*"-"+"\n"
-            self.debug( msg )
-            self.debug( "DCubeRunner is ready to work...")
-            
-      ## run method
-      # @param self "Me, myself and Irene"
-      def run( self ):
-            if ( self.error == None ):
-                  try:
-                        from ShellCommand import ShellCommand
-                        # FIXME - self.logger could be just python False...
-                        sc = ShellCommand( self.command,
-                                           self.logger ) 
-                        replay = sc.getReply()
-                        reLine = re.compile("OVERALL STATISTICS STATUS")
-                        status = 1
-                        for line in replay:
-                              for apiClass in self.__reAPIClasses:
-                                    if ( apiClass.search( line ) ):
-                                          self.debug( line )
-                                          if ( reLine.search( line ) ):
-                                                if ( "OK" in line ): status = 0
-                                    
-                        return status
-                  except ImportError: 
-                        self.info("No ShellCommand in PYTHONPATH! 'No Cin-Cin, no run...'")
-                        return 1
-            else:
-                  self.info(self.error)
-                  return 1
-
-      ## XML parsing of DCubeForRTTCfg.xml 
-      # @param self "Me, myself and Irene"
-      def __parseDCubeRTTCfg( self ):
-            self.debug( "parsing DCubeRunner configuration file %s" % self.DCubeForRTTCfg )
-            try:
-                  xmldoc = minidom.parse( self.DCubeForRTTCfg )   
-
-                  self.bin = self.getText( xmldoc.getElementsByTagName("bin")[0].childNodes )                   
-                  self.debug( "[14] will use binary version %s" % self.bin )
-
-                  self.root = self.getText( xmldoc.getElementsByTagName("root")[0].childNodes)
-                  self.rootVersion = xmldoc.getElementsByTagName("root")[0].getAttribute("version")
-                  rootBasePath = self.root + os.sep + self.rootVersion + os.sep + self.bin + os.sep + "root"
-            
-                  self.debug( "[15] will use root version %s installed in %s" % ( self.rootVersion , rootBasePath ) )
-                  
-                  self.python = self.getText( xmldoc.getElementsByTagName("python")[0].childNodes )
-                  self.pythonVersion = xmldoc.getElementsByTagName("python")[0].getAttribute("version")
-                  
-                  pythonBasePath = self.python + os.sep + self.pythonVersion + os.sep + self.bin  
-                  self.debug( "[16] will use python version %s installed in %s" % ( self.pythonVersion, pythonBasePath ) )
-            
-                  self.refBaseDir = self.getText( xmldoc.getElementsByTagName("ref_dir")[0].childNodes )
-                  self.debug( "[17] will use '%s' as base reference directory" % self.refBaseDir )
-
-                  self.dcubeServer = self.getText( xmldoc.getElementsByTagName("server")[0].childNodes ) 
-                  self.debug( "[18] will use DCubeServer installed in %s" % self.dcubeServer )
-         
-            except:
-                  self.error("error when parsing DCubeRunner configuration file, no run will be preformed")
-                  return [ "" ]
-
-            self.debug("constructing command to run DCube...")
-
-            out  = [ "export ROOTSYS=" + rootBasePath + "\n" ]
-            out.append( "export LD_LIBRARY_PATH=${ROOTSYS}/lib:" + pythonBasePath + "/lib\n" )
-            out.append( "export PATH=" + pythonBasePath + "/bin:${PATH}\n" )
-            out.append( "export PYTHONPATH=" + self.dcubeClient + ":" + rootBasePath + "/lib:" + pythonBasePath +  "/lib/python2.5\n\n" )
-            
-            
-            # changing the outputPath 
-            monPath = self.outputPath 
-
-            self.outputPath = self.outputPath + os.sep +  "DCube-" + self.refFile
-
-            # creating DCube subdirectory
-            out.append("mkdir -p " + self.outputPath + "\n")
-
-            dcubePath = self.dcubeClient
-                
-            dcubeConvert = os.path.join( dcubePath, "./dcubeConvert.py")
-
-            run = os.path.join(dcubePath,"./dcube.py")
-            run += " -p " # generate plot
-            run += " --branch " + self.nightly # set branch
-            run += " --install " + self.install # set install
-            run += " --cmtconfig " + self.cmtconfig # set cmtconfig
-            run += " --project " + self.project # set project
-            # set jobId
-            if ( self.jobId ): 
-                  run += " --jobId " + self.jobId 
-      
-                  
-            run += " -l " + self.outputPath + os.sep + self.monFile + ".dcube.log" # save log to monitored.dcube.log
-            run += " -s " + self.dcubeServer  # path to the DCubeServer installation
-            run += " -x " + self.outputPath + os.sep + self.monFile + ".dcube.xml " # save output to monitored.dcube.xml
-                 
-            # dangerous hacking at the moment...
-            if ( "DUMMY-TEST" in self.cfgFile or os.path.isabs( self.cfgFile ) ):
-                  run += " -c " + self.cfgFile
-                  config = self.cfgFile 
-            else: 
-                  # checking cfg file in reference dir
-                  refCfg = "/afs/cern.ch/atlas/project/RTT/Results/reference/" + self.cvsPath + os.sep + self.cfgFile 
-                  self.debug("will check the timestamp of DCube configuration file (reference volume): " + refCfg )
-                  refCfgStat = None
-
-                  if ( os.path.exists(refCfg) ):
-                        refCfgStat = os.stat(refCfg)
-                        self.debug("File %s modified %d" %( refCfg, refCfgStat[8]) )
-                  else:
-                        self.debug("DCube configuration file %s not found on 'reference' volume" % self.cfgFile )
-                        
-                        # checking cfg file in cvs co dir 
-                        
-                  cvsCfg = self.cmtPath + os.sep + self.cvsPath.lstrip("offline") + "/test/" + self.cfgFile
-                  self.debug("will check the timestamp of DCube configuration file (CVS checkout volume):" + cvsCfg)
-                  cvsCfgStat = None 
-                  if ( os.path.exists(cvsCfg) ) :
-                        cvsCfgStat = os.stat( cvsCfg )
-                        self.debug("File %s modified %d" %( cvsCfg, cvsCfgStat[8]) )
-                  else:
-                        self.debug("DCube configuration file %s not found in CVS checkout directory" % self.cfgFile )
-
-                        
-                  # choose configuration file based on latest timestamp 
-                  config = None
-
-                  # both are present, check time stamps
-                  if ( None not in ( refCfgStat, cvsCfgStat) ):
-                        # install area newer
-                        if ( refCfgStat[8] <= cvsCfgStat[8] ):
-                              config = cvsCfg
-                        else: 
-                              config = refCfg
-
-                  # reference is present
-                  elif ( None != refCfgStat ):
-                        config = refCfg
-                  # install area is present
-                  elif ( None != cvsCfgStat ):
-                        config = cvsCfg
-                  # both are absent  
-                  else:
-                        pass
-
-                        
-                  if ( config != None ):
-                        run += " -c " + config # add configuration file 
-                  else:
-                        self.debug("DCube configuration file %s not found in ASF reference or InstallArea directories" % self.cfgFile )
-                        return [ "" ]
-                  
-            if ( self.refFile != None ):
-                  run += " -r " + os.path.join( os.path.join(self.refBaseDir , self.cvsPath) , self.refFile ) # overwrite reference
-                  run += " " + os.path.join( monPath , self.monFile)  # add monitored file at the end 
-                  
-
-            if ( "DUMMY-TEST" in self.cfgFile ):
-                  dcubeConvert += " %s " % os.path.abspath(self.cfgFile)
-            elif ( config != None  ):
-                  dcubeConvert += " %s " % os.path.abspath(config)
-
-            dcubeConvert += " %s\n"  % os.path.join( os.path.join(self.refBaseDir , self.cvsPath) , self.refFile )  
-            out.append( dcubeConvert )
-            out.append( run + "\n")
-                  
-            return out
-            
-
-      ## logger level INFO
-      # @param self "Me, myself and Irene"
-      # @param msg message
-      def info( self, msg ):
-            if ( bool(self.logger) ):
-                  self.logger.info( str(msg) )
-            else:
-                  print "DCubeRunner INFO  " + str(msg) 
-
-      ## logger level DEBUG
-      # @param self "Me, myself and Irene"
-      # @param msg message 
-      def debug( self, msg ):
-            if ( bool(self.logger) ):
-                  self.logger.debug( str( msg) )
-            else:
-                  print "DCubeRunner DEBUG " + str(msg)
-
-      
-      ## get CDATA section
-      # @param self "me, myself and Irene"
-      # @param nodelist list of XML DOM Nodes
-      def getText( self, nodelist ):
-        out = ""
-        for node in nodelist:
-            if node.nodeType == node.TEXT_NODE:
-                out = out + node.data
-        return out 
-
-      ## split path
-      # @param self "Me, myself and Irene"
-      # @param path path to split
-      # @param rest splitted tail
-      def pathSplit(self, path, rest=[] ):
-            ( head , tail ) = os.path.split( path )
-            if len(head) < 1: return [tail]+ rest
-            if len(tail) < 1: return [head]+ rest
-            return self.pathSplit( head, [tail] + rest )
-
-      ## find common names in path
-      # @param self "Me, myself and Irene"
-      # @param l1 first list with splitted path
-      # @param l2 second  lust with splitted path
-      # @param common list of common dirnames
-      def commonPath(self,  l1, l2, common=[] ):
-            if len(l1) < 1: return (common, l1, l2)
-            if len(l2) < 1: return (common, l1, l2)
-            if l1[0] != l2[0]: return (common, l1, l2)
-            return self.commonPath( l1[1:], l2[1:], common+[l1[0]])
-
-      ## finds relative path between p1 and p2
-      # @param self "Me, myself and Irene"
-      # @param p1 first path
-      # @param p2 second path
-      def relPath(self, p1, p2):
-            ( common, l1, l2 ) = self.commonPath( self.pathSplit(p1), self.pathSplit(p2))
-            p = []
-            if len(l1) > 0:
-                  p = [ '../' * len(l1) ]
-            p = p + l2
-            return os.path.join( *p )
-
-
-## dummy test running
-if __name__ == "__main__":
-      print "running DCubeRunner..."
-      testCfg = { "DCubeCfg"         : "DUMMY-TEST-DCUBE-CFG.XML",
-                  "DCubeMon"         : "monitored.root",
-                  "DCubeRef"         : "reference.root",
-                  "DCubeJobId"       : "test_01",
-                  "JobDescriptor"    : None }
-      dc = DCubeRunner( testCfg )
-      dc.run()
-
-      
diff --git a/Tools/RunTimeTester/share/DiffPoolFilesRunner.py b/Tools/RunTimeTester/share/DiffPoolFilesRunner.py
deleted file mode 100644
index 80131d09b58586d0cb3f525e5f4a55217ea1d3fd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/DiffPoolFilesRunner.py
+++ /dev/null
@@ -1,51 +0,0 @@
-from Logger import Logger
-# from Factory_CmtLines import CmtLinesFactory
-from ShellCommand import ShellCommand
-import os
-import os.path
-
-class DiffPoolFilesRunner:
-    def __init__(self, argDict):
-        desc              = argDict.get('JobDescriptor')
-        paths             = desc.paths
-        self.runPath      = desc.runPath
-        self.logger       = Logger()
-        self.theFile      = argDict['fileName']
-        self.theRefFile   = argDict['refFileName']        
-        self.envSetupCmds = argDict['cmtLinesCmds']        
-        # self.envSetupCmds = CmtLinesFactory(self.logger).create(paths).lines()
-        
-    def outcome(self, lines):
-        diffPoolSuccessStr = '## Comparison : [OK]'
-        for line in lines:
-            if line.find(diffPoolSuccessStr)!=-1: return 0
-        return 1
-
-    def run(self):
-        cmds = []
-        cmds.extend(self.envSetupCmds)
-        cmds.append('cd %s' % self.runPath)
-        cmds.append('diffPoolFiles.py -f %s -r %s' % (self.theFile, self.theRefFile))
-
-        # launch the commands
-        sc    = ShellCommand(cmds, self.logger)            
-        reply = sc.getRawReply()
-        [self.logger.debug(str(l)) for l in reply]
-
-        handle = None
-        outFile = '%s__%s.diffPool' % (os.path.basename(self.theFile), os.path.basename(self.theRefFile))
-        outPath = os.path.join(self.runPath, outFile)
-            
-        try:
-            handle = open(outPath, 'w')
-            handle.writelines(reply)
-            handle.close()
-            self.logger.debug('Successfully wrote out: %s' % outPath)
-            return self.outcome(reply)
-        except Exception, e:
-            m  = 'Unable to output results of diffPoolFiles to file: %s\n' % outPath
-            m += str(e)
-            self.logger.error(m)
-            if handle: handle.close()
-            return 1
-        
diff --git a/Tools/RunTimeTester/share/EchoFile.py b/Tools/RunTimeTester/share/EchoFile.py
deleted file mode 100755
index 50f8d1dd440ce9690c55b43daefb93590d4fedef..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/EchoFile.py
+++ /dev/null
@@ -1,28 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-from Logger            import Logger
-
-class EchoFile:
-    def __init__(self, argDict):
-
-        self.inFile   = argDict.get('datafile')
-        # self.logger = params['logger']
-        self.logger   = Logger()
-
-    def run(self):
-        inF  = open(self.inFile, 'r').readlines()
-        outF = open('echoedFile.txt', 'w')
-
-        for line in inF:
-            outF.write(line)
-
-        outF.close()
-        
diff --git a/Tools/RunTimeTester/share/ExeRunner.py b/Tools/RunTimeTester/share/ExeRunner.py
deleted file mode 100644
index b381af480f4399a74aa70f096e80f4412e04d299..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/ExeRunner.py
+++ /dev/null
@@ -1,44 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in from    #
-# more complicated classes in __init__               #
-#                                                    #
-######################################################
-
-from Logger   import Logger
-from commands import getstatusoutput
-import os.path
-
-class ExeRunner:    
-    def __init__(self, paramDict={}):
-        jDescriptor  = paramDict['JobDescriptor']
-        self.exeName  = paramDict['exeName']
-        self.exeParamString = paramDict.get('exeParamString','')
-        self.logger  = Logger()
-        self.runPath = jDescriptor.runPath
-
-    def stdOutToFile(self, output, fName):
-        handle = None
-        outPath = os.path.join(self.runPath, fName)
-        try:
-            handle = open(outPath, 'w')
-            handle.write(output)
-            handle.close()
-            self.logger.debug('Successfully wrote out: %s' % outPath)
-        except Exception, e:
-            m  = 'Unable to output results of ExeRunner %s to file: %s\n' % (self.exeName, outPath)
-            m += str(e)
-            self.logger.error(m)
-            if handle: handle.close()
-
-    def run(self):
-        cmd = '%s %s' % (self.exeName, self.exeParamString)
-        stat, output = getstatusoutput(cmd)
-        header = 'RTT ran: %s\n\n' % cmd
-        output = header + output
-        {0:self.logger.debug}.get(stat, self.logger.error)(output)
-        self.stdOutToFile(output, 'ExeRunner.%s.stdout' % self.exeName)
-        return {0:0}.get(stat, 1)
diff --git a/Tools/RunTimeTester/share/FileComparator.py b/Tools/RunTimeTester/share/FileComparator.py
deleted file mode 100755
index 37c6e55b707a1dc0df90b99c0891ea36a0575fec..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/FileComparator.py
+++ /dev/null
@@ -1,124 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-import sys, os, logging
-
-from Logger            import Logger
-
-class FileComparator:
-
-    def __init__(self, paramDict):
-        self.logger          = Logger()
-        self.logger.debug('instantiating file comparator')
-        jDescriptor          = paramDict.get('JobDescriptor')
-        #location             = paramDict.get('refFileLocation','rttDB')
-        location             = 'install'
-        #self.diffFile        = "differ.out"
-        self.testFile        = paramDict['testFile']
-        self.refFile         = paramDict['refFile']
-        self.diffFile        = paramDict['diffFile']
-        self.shareArea       = jDescriptor.paths.shareArea
-        
-        #self.success  =  0
-        #self.error    = -1
-        self.error     = 'error'
-        self.success   = 'success'
-        # ................
-        
-        srcDir = jDescriptor.paths.RTTSrcDir
-        if srcDir not in sys.path: sys.path.append(srcDir)
-
-        #location=self.getLocation(jDescriptor)
-        if location == 'install':
-            self.refFile = os.path.join(self.shareArea, self.refFile)
-            self.diffFile = os.path.join(jDescriptor.runPath, self.diffFile)
-            self.testFile  = os.path.join(jDescriptor.runPath, self.testFile)
-        elif location == 'rttDB':
-            pass
-            #self.refFile = dbFileRetriever.writeRefFile(self.refFile)
-            #dbFileRetriever.storeFileInDB(self.testFile)            
-        else:
-            self.logger.error("Ref File location unknown "+str(location))
-            
-            
-    def getLocation(self, descriptor, whereFrom ='refFileSystem'):
-        from formatCollection import formatCollection
-        path, file = os.path.split(descriptor.runPath)
-        from FileLocationKeyGenerator import KeyGeneratorFactory
-        
-        lo = KeyGeneratorFactory().create(whereFrom, path)
-        return lo
-
-    def run(self):
-        self.logger.debug('Running simple file comparator')
-        if not os.path.exists(self.refFile):
-            self.logger.error('reference file %s does not exist or not found '%str(self.refFile))
-            return 'error'
-
-        cwd   =  os.getcwd()
-        testF = ""
-        refF  = ""
-        testFile = self.testFile
-        #reference files will be installed or will be in RTT DB
-        if not os.path.isabs(self.testFile):
-            testFile = os.path.join(cwd,self.testFile)
-
-        try:
-            testF = open(testFile, "rb").read()
-            refF  = open(self.refFile, "rb").read()            
-        except Exception, e:
-            self.logger.error("Could not read files to compare ")
-            self.logger.error(str(e))
-            msg = 'File Comparator failed : ' +str(e)
-            self.logger.error(msg)
-            return self.error
-        
-        if testF != refF:
-            print 'files are different'
-            self.logger.debug("The following two files are different")
-            self.logger.debug("File 1 :"+ str(testFile))
-            self.logger.debug("File 2 :"+ str(self.refFile))
-            diffCommand = ["diff "+testFile +" "+self.refFile+" >"+self.diffFile]
-            
-            from ShellCommand      import ShellCommand
-            runDiff = ShellCommand(diffCommand, self.logger)
-            #runDiff.wait()
-            
-            msg =  'File Comparator failed. \n'
-            msg += 'The following files are different : \n%s\n%s' % (
-                str(testFile), str(self.refFile))
-
-            self.logger.error(msg)
-            return self.error   
-        else:
-            self.logger.info("The following two files are the same ")
-            self.logger.info("File 1: "+str(testFile))
-            self.logger.info("File 2: "+str(self.refFile))
-            #logger.info("File 2: "+str(self.refFile))
-            return self.success
-            
-        
-
-    def usage(self):
-        self.logger.info("Usage : FileComparator(<dict>)")
-        self.logger.info("success ==> Files same; error ==> Files differ or exception")
-
-
-if __name__ == '__main__':
-
-    print 'running file comparator'
-    dict = {'testFile':'/home/en/RTT/test.txt',
-            #'refFile':'HelloWorldOptions.txt',
-            'refFile':'/home/en/RTT/test.txt',
-            'diffFile': 'differ.out'}
-
-    fc = FileComparator(dict)
-    print str(fc.__dict__)
-    fc.run()
diff --git a/Tools/RunTimeTester/share/FileGrepper.py b/Tools/RunTimeTester/share/FileGrepper.py
deleted file mode 100755
index 960ece51327a0f1b8c5736dfc98d13ad9ad7518b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/FileGrepper.py
+++ /dev/null
@@ -1,218 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannot shelve the objects                #
-# Instead, unload the necessary variables in from    #
-# more complicated classes in __init__               #
-#                                                    #
-######################################################
-
-
-
-import sys, os, os.path, glob, re, types, string
-from Logger import Logger
-from sets import Set
-
-class FileGrepper:
-    
-    def __init__(self, paramDict={}):
-        jDescriptor = paramDict['JobDescriptor']
-
-        self.runPath = jDescriptor.runPath
-        self.srcDir  = jDescriptor.paths.RTTSrcDir
-        if self.srcDir not in sys.path: sys.path.append(self.srcDir)
-
-        self.success   =  0
-        self.error     = -1
-        self.logger    = Logger()
-                
-        self.iFilePat = paramDict.get('inputFile', '').strip()
-        if not self.iFilePat: return        
-
-        self.illegalOutputFileName = False
-        self.outFile       = paramDict.get('outputFile', '')
-        if self.outFile:
-            ok_chars = string.digits + string.ascii_letters + '_-.+'
-            if [t for t in self.outFile if t not in ok_chars]:
-                self.illegalOutputFileName = True
-                return
-            self.outFile = os.path.join(self.runPath, self.outFile)
-
-        search = paramDict.get('searchList', [])
-        if search: search = [s.strip() for s in search.split(',')]
-
-        veto = paramDict.get('vetoList', [])
-        if veto: veto = [v.strip() for v in veto.split(',')]
-
-        search.extend(paramDict.get('searchStrings', []))
-        veto.extend(paramDict.get('vetoStrings', []))
-
-        self.searchDict    = self.dictify(search)
-        self.vetoDict      = self.dictify(veto)
-        self.findAllSearch = self.boolify(paramDict.get('mustFindAllSearchTerms', True))
-        self.sendToFile    = paramDict.get('outputLinesMatching', '')
-        self.verbose       = self.boolify(paramDict.get('verbose', True))
-        
-    # ========================================================================
-    # ========================================================================
-    # ========================================================================
-
-    def dictify(self, list_):
-        if type(list_)==type(''):
-            list_ = list(list_)
-        d = {}
-        [d.setdefault(l, []) for l in list_] # uniquify search/veto strings (dict keys)
-        return d
-
-    def boolify(self, string_):
-        if type(string_)!=type(''):
-            string_ = str(string_)
-        return {'False':False}.get(string_, True)
-        
-    def run(self):
-        if not self.iFilePat:
-            m = '%s ==> no input file pattern to match was given. Abandoning.' % self.__class__.__name__
-            self.logger.error(m)
-            return self.error
-
-        if not self.outFile:
-            m = "%s ==> no output file was given. Abandoning." % self.__class__.__name__
-            self.logger.error(m)
-            return self.error
-
-        if self.illegalOutputFileName:
-            m = '%s ==> Illegal output file name: %s' % (self.__class__.__name__, self.outFile)
-            self.logger.error(m)
-            return self.error
-        
-        def print_(file_, fileBad, failureReason, dict, type_):
-            matches = [(n,m) for n,m in dict.items() if m]
-            if not matches: return ''
-            output = ''
-            if self.verbose:
-                output  = '--------------------------------------------------------------------\n'
-                output += 'File: %s\n' % file_
-                output += '--------------------------------------------------------------------\n'
-                output += '%s\n' % {False: 'STATUS: PASS', True: 'STATUS: FAIL'}[fileBad]
-                if fileBad:
-                    output += 'This file FAILed for the following reason:\n'
-                    output += '%s\n\n' % failureReason
-                output += '%d lines match a %s term. '  % (reduce(lambda x,y:x+y, [len(m) for n,m in matches],0), type_)
-                output += '%d (of %d total) %s terms were matched.\n' % (len(matches), len(dict.keys()), type_)
-                # output += 'All search terms must be matched: %s\n' % self.findAllSearch
-                output += '--------------------------------------------------------------------\n'
-            for n, m in matches:
-                if self.verbose:
-                    output += '%s term %s matched the following %d lines:\n' % (type_, n, len(m))
-                    for mm in m: output += '\t%s\n' % mm.strip()
-                else:
-                    for mm in m: output += '%s\n' % mm.strip()
-                        
-            return output
-                
-        self.outFile = os.path.join(self.runPath, self.outFile)
-        inFiles = self.getInFiles()
-        if not inFiles: return self.error
-
-        success = True
-        output  = ''        
-        
-        for file_, ((fileBad, failureReason), searchDict, vetoDict) in [(inf, self.processInfile(inf)) for inf in inFiles]:
-
-            success = success and not fileBad
-            if self.sendToFile == 'veto':
-                output += print_(file_, fileBad, failureReason, vetoDict, 'veto')
-            elif self.sendToFile == 'search':
-                output += print_(file_, fileBad, failureReason, searchDict, 'search')
-            else: # send both veto and search matches to file
-                output += print_(file_, fileBad, failureReason, vetoDict, 'veto')
-                output += print_(file_, fileBad, failureReason, searchDict, 'search')
-
-        h = open(self.outFile, 'w')
-        h.write(output)
-        h.close()
-
-        return {True:self.success, False:self.error}[success]
-
-    # ========================================================================
-    # ========================================================================
-    # ========================================================================
-        
-    def getInFiles(self):
-        pattern = os.path.join(self.runPath, self.iFilePat)
-        matches = [m for m in glob.glob(pattern) if os.path.isfile(m)]
-        if not matches:
-            m = '%s => No matches found for the in-file pattern: %s\n' % (self.__class__.__name__, self.iFilePat)
-            self.logger.warning(m)
-        return matches
-
-    # ========================================================================
-    # ========================================================================
-    # ========================================================================
-
-    def processInfile(self, inFile): 
-        """Return list of matching lines from the in file."""
-        def readfile():
-            handle = open(inFile)
-            for line in handle.readlines(): yield line
-            handle.close()
-
-        def matchline(sdict, vdict, line):
-            line = line.strip()
-            slist = [s for s in sdict.keys() if re.search(s, line)]
-            vlist = [v for v in vdict.keys() if re.search(v, line)]
-            [sdict[s].append(line) for s in slist]
-            [vdict[v].append(line) for v in vlist]            
-            return slist or vlist
-        def foundAllSearchTerms(sdict):
-            return len([sv for sv in sdict.values() if sv])==len(sdict.keys())
-        def foundSomeSearchTerms(sdict):
-            return len([sv for sv in sdict.values() if sv])>0
-        def fileBad(sdict,vdict):
-            vetofound = [vv for vv in vdict.values() if vv]
-            if len(vetofound)>0:
-                return (True, '%d veto terms were found' % len(vetofound))
-            if self.findAllSearch and not foundAllSearchTerms(sdict):
-                return (True, 'Not all search terms matched, and they should have been.')
-            if not foundSomeSearchTerms(sdict):
-                return (True, 'No search terms matched at all.')
-            return (False, '')
-            
-        s = {}
-        v = {}
-        s.update(self.searchDict)
-        v.update(self.vetoDict)
-
-        # fill the dicts
-        [matchline(s,v,line) for line in readfile()]
-        return (fileBad(s,v), s, v)
-
-    
-if __name__ == '__main__':
-    class Paths:
-        def __init__(self):
-            self.RTTSrcDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Tools.CURRENT/RunTimeTester/src'
-
-    class Desc:
-        def __init__(self):
-            self.paths = Paths()
-            self.runPath = '/afs/cern.ch/atlas/project/RTT/data/brinick/Tools.CURRENT/RunTimeTester/share'
-            
-    d = {'JobDescriptor': Desc(),
-         'inputFile' : 'a ',# 'bl?h/log.log',
-         'outputFile': 'log.log.output', 
-         'searchStrings': ['leaving with code 0', 'ErrorCode=0 \(OK\)', 'successful run'],
-         'searchList'   : 'ErrorCode=0 \(OK\),    leaving with code 0',
-         'vetoStrings'  : ['FATAL', 'Core dump', 'stack trace', 'Shortened traceback'],
-         'mustFindAllSearchTerms': 'bugger',
-         #'outputLinesMatching': 'veto',
-         # 'verbose': 'True'
-         }
-    
-    fg = FileGrepper(d)
-    print fg.run()
-    print fg.logger.errorL
-    print fg.logger.debugL
-    print fg.logger.infoL
-    print fg.logger.warningL
diff --git a/Tools/RunTimeTester/share/FileToDict.py b/Tools/RunTimeTester/share/FileToDict.py
deleted file mode 100755
index b52282ac6a119f7fc21331d0538a4ecd0f21b34d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/FileToDict.py
+++ /dev/null
@@ -1,56 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-import string
-from Logger            import Logger
-
-class FileToDict:
-    def __init__(self, argDict):
-
-        self.inFile   = argDict.get('datafile')
-        self.logger   = Logger()
-        # self.logger = params['logger']
-
-    def run(self):
-        self.DataDict      = self.makeDict(self.inFile)
-
-        return self.DataDict
-        
-
-    def retrieveFile(self, key, db):
-        pass
-    
-    
-    def linesCleanUp(self, lines):
-        ls =[]
-        for line in lines:
-            line =string.strip(line)
-            if len(line)>0:
-                ls.append(line)
-
-        return ls
-
-
-    def makeDict(self,file):
-        lines         = open(file, 'r').readlines()
-        lines         = self.linesCleanUp(lines)
-        
-        self.dict={}
-        for item in lines:
-            keyValuePairs     = string.split(item, ":")
-            for m in  keyValuePairs:
-                keyValuePairs = self.linesCleanUp(keyValuePairs)
-            self.dict[keyValuePairs[0]] = int( keyValuePairs[1])
-            
-        return self.dict
-
-
-
-
diff --git a/Tools/RunTimeTester/share/FilesWithStringsFinder.py b/Tools/RunTimeTester/share/FilesWithStringsFinder.py
deleted file mode 100755
index 7d17da8dfb5e37fdba2977faaed5b7bd78381a4c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/FilesWithStringsFinder.py
+++ /dev/null
@@ -1,236 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-from Logger            import Logger
-
-import os
-
-class FilesWithStringsFinder:
-    
-    def __init__(self, paramDict):
-
-        self.logger            = Logger()
-        self.inputFilePatterns = paramDict.get('inputFilePatterns', [])
-        self.searchPatterns    = paramDict.get('searchPatterns',    [])
-        self.dirWithFiles      = paramDict.get('dirWithFiles', 'resultsDir')
-        self.writeTxt          = paramDict.get('writeTxt', 'yes')
-        self.writeHTML         = paramDict.get('writeHTML','yes')
-        jDescriptor            = paramDict.get('JobDescriptor', None)
-        self.runPath           = jDescriptor.runPath
-        outputFile             = paramDict.get('outputFile','filesWithStrings')
-        self.outputFile        = os.path.join(self.runPath,outputFile)
-        self.resultsDirs       = jDescriptor.paths.resultsDirs
-        self.writers           =  []
-
-
-        dict = {
-            'runPath':           self.runPath,
-            'outputFile':        self.outputFile,
-            'inputFilePatterns': self.inputFilePatterns,
-            'searchPatterns':    self.searchPatterns,
-            'logger':            self.logger
-            }
-
-        class Writer:
-            def __init__(self, dict):
-                self.ofile  = None
-                self.outfn = os.path.join( dict['runPath'],
-                                           dict['outputFile'])
-
-                
-                fpats = dict['inputFilePatterns'] 
-                spats = dict['searchPatterns']
-
-                header  = ['Files to match:']
-                header.extend([('  %s ' % i ) for i in fpats])
-                header.append(' ')
-                header.append('\n\nStrings to match:')
-                header.extend([('  %s ' % s ) for s in spats])
-                header.append(' ')
-                self.header = [self.terminateLine(line) for line in header]
-                self.logger   = dict['logger']
-
-            
-            def openFile(self):
-                """File is opened when there is a need to write (is not
-                possible to open the file at contruction). On opening
-                write out (once!) the header. This part of the action writing
-                an item is common to all derived sclasses - they do the rest.
-                """
-
-                if self.ofile: return
-
-                self.ofile = open(self.outfn, 'w')
-                if not self.ofile:
-                    msg = 'FilesEithStringsFinder: could not open file %s' % self.outfn
-                    self.logger.error(msg)
-                    return
-                    
-                [self.ofile.write(line) for line in self.header]
-
-        class HTMLWriter(Writer):
-            def __init__(self, dict):
-                Writer.__init__(self, dict)
-                self.outfn  += '.html'
-                self.logger   = dict['logger']
-
-        
-                htmlHeader = []
-                htmlHeader.append('<html>\n')
-                htmlHeader.append('<head>\n')
-                htmlHeader.append('   <title>FilesWithStringsFinder</title>\n')
-                htmlHeader.append('</head>\n')
-                htmlHeader.append('<body>\n')
-                htmlHeader.extend(self.header)
-                htmlHeader.append('<table>\n')
-                self.header = htmlHeader
-
-            def terminateLine(self, line):
-                line+='<br>\n'
-                return line
-
-            def outputItem(self, item):
-
-                if not self.ofile: return
-                
-                fn  = os.path.basename(item[0])
-                href ='<a href="%s">  %s </a>' % (fn, fn)
-                htext =  'contains "%s"'% (item[1])
-
-
-
-                line  = '<tr>\n'
-                line += '<td>'+href+'</td><td>'+htext+'</td>\n'
-                line += '</tr>\n'
-                
-                self.ofile.write(line)
-
-                
-            def closeFile(self):
-                if not self.ofile: return
-
-                self.ofile.write('</table>\n')
-                self.ofile.write('</body>\n')
-                self.ofile.write('</html>\n')
-                self.ofile.close()
-            
-        class TxtWriter(Writer):
-            def __init__(self, dict):
-                Writer.__init__(self, dict)
-                self.outfn  += '.txt'
-                self.logger   = dict['logger']
-                    
-            def terminateLine(self, line):
-                line+='\n'
-                return line
-
-
-            def outputItem(self, item):
-                # output to the html file needs to be htmlified
-                # delay opening the files until now, otherwise there is
-                # the danger the output directories have not been constructed.
-
-                if not self.ofile: return
-
-                width = 60
-                fn = os.path.basename(item[0])
-                line = 'file: %s contains "%s" \n' % (string.ljust(fn, width),
-                                                      item[1]
-                                                      )
-                self.ofile.write(line)
-
-            def closeFile(self):
-                if not self.ofile: return
-                self.ofile.close()
-            
-        if self.writeTxt == 'yes':
-            self.writers.append(TxtWriter(dict))
-        if self.writeHTML == 'yes':
-            self.writers.append(HTMLWriter(dict))
-
-        self.logger.debug('FilesWithStringsFinder Writers: %s' % (
-            str(self.writers)))
-    def run(self):
-
-        files = self.filesInDirs()
-        self.logger.debug('FilesWithStringsFinder: filesInDirs %s' % str(files))
-
-        fPatterns     = self.inputFilePatterns
-        filesToSearch = []
-        
-        [filesToSearch.extend(fnmatch.filter(files,
-                                             ipfp)) for ipfp in fPatterns]
-
-        strs         = self.searchPatterns
-
-        m = """FilesWithStringsFinder: files to search %s\nsearchPatterns:
-        %s""" % (str(filesToSearch), str(strs))
-        self.logger.debug(m)
-
-        class StringInFile:
-            def __init__(self, base, file):
-                ifn = os.path.join(base,file)
-                print 'checking file ',ifn,'... '
-                self.contents = open(ifn, 'r').read()
-            def check(self, string):
-                print '    ... for string', string
-                if self.contents.find(string) > 0: return 0
-                return -1
-            
-        filesWithString = []
-        for f in filesToSearch:
-            sFile = StringInFile(dwf, f)
-            [filesWithString.append((f,s))  for s in strs if sFile.check(s)]
-
-        self.logger.debug('filesWithString %s' % str(filesWithString))
-
-        [writer.openFile() for writer in self.writers]
-
-        for writer in self.writers:
-            [writer.outputItem(item) for item in filesWithString]
-
-        [writer.closeFile() for writer in self.writers]
-
-        return 0
-
-    def filesInDirs(self):
-        """
-        return a list of files found using the descriptor Paths object scan
-        the descriptor is passed as an argument, rather than getting
-        it from the dictionary asthe descriptor passed may not be the local
-        on: watchers pass the descriptors of their confederation minders, not
-        their own descriptors.
-        """
-
-        msgHead ='FilesWithStringsFinder.filesInDirs: '
-
-        logicalDir = self.dirWithFiles
-        files = []
-        
-        self.logger.debug('%s files to search %s' % (msgHead, str(logicalDir)))
-
-        if logicalDir == 'resultsDir':
-            key = 'resBase/build/branch/runType/release/jobGroup'
-            if key not in self.resultsDirs:
-                self.logger.error('results directory not in Paths %s' % key)
-                return files
-            dwf = self.resultsDirs[key]
-            self.logger.debug('FilesWithStringsFinder violfile directory is %s' % dwf)
-            files = os.listdir(dwf)
-            return files
-        elif logicalDir == 'runDir':
-            files = os.listdir(self.runPath)
-            self.logger.debug('FilesWithStringsFinder violfile directory is %s' % self.runPath)
-            return files
-        else:
-            msg = 'FilesWithStringFinder: Unknown directory to scan %s' % dwf
-            self.logger.error(msg)
-            return files
-
diff --git a/Tools/RunTimeTester/share/JunkTest.py b/Tools/RunTimeTester/share/JunkTest.py
deleted file mode 100755
index 2fec5fa6d62a8c6e57f7b3a3aabfbf7a3bde9d42..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/JunkTest.py
+++ /dev/null
@@ -1,28 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-from Logger import Logger
-
-class JunkTest:
-    def __init__(self, argDict):
-        self.int = 1
-        self.logger   = Logger()
-
-    def run(self):
-        # self.logger.error('JunkTest RTT test run method')
-        print 'JunkTest RTT test run method'
-        self.logger.critical('JunkTest: critical message')
-        self.logger.error('JunkTest: error message')
-        self.logger.warning('JunkTest: warning message')
-        self.logger.info('JunkTest: info message')
-        self.logger.debug('JunkTest: debug message')
-        return 0
-        
-
diff --git a/Tools/RunTimeTester/share/LogFileGrepper.py b/Tools/RunTimeTester/share/LogFileGrepper.py
deleted file mode 100644
index ca263f97e9ca98309613665097a66eafa9d6ab32..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/LogFileGrepper.py
+++ /dev/null
@@ -1,29 +0,0 @@
-from FileGrepper import FileGrepper
-import os.path
-
-class LogFileGrepper(FileGrepper):
-    def __init__(self, argdict={}):
-        argdict['inputFilePattern']  = argdict['JobDescriptor'].log
-        argdict['outputFile'] = '%s.grep_output' % argdict['JobDescriptor'].log
-        FileGrepper.__init__(self, argdict)
-
-    def run(self):
-        return FileGrepper.run(self)
-
-if __name__ == '__main__':
-    class Paths:
-        def __init__(self):
-            self.RTTSrcDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Tools.CURRENT/RunTimeTester/src'
-
-    class Desc:
-        def __init__(self):
-            self.paths = Paths()
-            self.log  = 'log.log'
-            self.runPath = '/afs/cern.ch/atlas/project/RTT/data/brinick/Tools.CURRENT/RunTimeTester/share'
-            
-    d = {'JobDescriptor': Desc(),
-         'searchList': ['leaving with code 0', 'ErrorCode=0 \(OK\)', 'successful run'],
-         'vetoList'  : ['FATAL', 'Core dump', 'stack trace', 'Shortened traceback']}
-    
-    fg = LogFileGrepper(d)
-    print fg.run()
diff --git a/Tools/RunTimeTester/share/Logger.py b/Tools/RunTimeTester/share/Logger.py
deleted file mode 100755
index 70488dbb93ffd4d4730e9e9332cff4e80da76a34..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/Logger.py
+++ /dev/null
@@ -1,13 +0,0 @@
-class Logger:
-    def __init__(self):
-        self.debugL    = []
-        self.infoL     = []
-        self.warningL  = []
-        self.errorL    = []
-        self.criticalL = []
-
-    def debug(self, line):    self.debugL.append(line)
-    def info(self, line):     self.infoL.append(line)
-    def warning(self, line):  self.debugL.append(line)
-    def error(self, line):    self.errorL.append(line)
-    def critical(self, line): self.criticalL.append(line)
diff --git a/Tools/RunTimeTester/share/PerfMonDozer.py b/Tools/RunTimeTester/share/PerfMonDozer.py
deleted file mode 100644
index a72af2ae18d352d58e2c618c141912811ae714c4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/PerfMonDozer.py
+++ /dev/null
@@ -1,231 +0,0 @@
-#!/bin/env python
-
-import sys,time
-sys.argv.append("-b")
-
-
-dozerPath =  "/afs/cern.ch/atlas/project/RTT/Results/dozer/api" 
-sys.path = [ dozerPath ] + sys.path
-sys.path.append( dozerPath )
-
-from CoolDozer import *
-import RTTDozerUtils
-import random
-import os.path
-
-class PerfMonDozer( object ):
-
-    @RTTDozerUtils.init
-    def __init__( self , argDict = {} ):
-        jd = argDict.get("JobDescriptor", None )
-        self.jobLog     = jd.log
-        self.package    = jd.package
-        self.release    = jd.paths.release
-        self.nightly    = jd.paths.branch
-        self.install    = jd.paths.runType
-        self.cmtconfig  = jd.paths.cmtConfig
-        self.project    = jd.paths.topProject
-        self.jobId      = jd.trendId+'_perfMonDozer'
-        self.runMe      = True
-        
-        if jd.trendId is None or not jd.trendId.strip(): 
-            self.log.panic("Job has no trendID. Cannot perform PerfMonDozer magic.")
-            self.runMe = False
-            self.channelKey = None
-            return
-
-        self.log.info("Job has a trendID (%s), proceeding..." % jd.trendId)
-        self.channelKey = self.jobId
-
-    @RTTDozerUtils.run
-    def run(self):
-        if not self.runMe: return
-        ## write latest to db
-        self.write() 
-            
-        ## read data back and make some plots
-        self.read()
-
-    def createSchema(self):
-        self.rConfig = DozerConfig("RTT",  self.log)
-        
-        rChannel  = DozerChannel(self.channelKey, "PerfMonDozer run")       
-        rChannel += DozerData("CPU per evt", "Float", "ms")
-        rChannel += DozerData("CPU per evt;error", "Float", "ms")
-        
-        rChannel += DozerData("Real time per evt", "Float", "ms")
-        rChannel += DozerData("Real time per evt;error", "Float", "ms")
-
-        rChannel += DozerData("Vmem per evt", "Float", "MB")
-        rChannel += DozerData("Vmem per evt;error", "Float", "MB")
-
-        rChannel += DozerData("Malloc per evt", "Float", "ms")
-        rChannel += DozerData("Malloc per evt;error", "Float", "ms")
-
-        rChannel += DozerData("NormCPU per evt", "Float", "kSi2k")
-        rChannel += DozerData("Si2k constant", "Float")
-
-        rChannel += DozerData("MemLeak per evt", "Float", "kB")
-        rChannel += DozerData("MemLeak per evt:Chi2", "Float")
-        rChannel += DozerData("MemLeak per evt:NDF", "Int")
-
-        rSchema = DozerSchema("*", "*", "*", "*",  "*")
-        rSchema += rChannel
-        
-        self.rConfig += rSchema 
-        return self.rConfig.get(self.dbkey)
-            
-    def extractFromLog(self):
-        class NotYetException(Exception):
-            def __init__(self, msg):
-                Exception.__init__(self, msg)
-        class NoPerfMonException(Exception):
-            def __init__(self, msg):
-                Exception.__init__(self, msg)
-        def readLog():
-            h = open(self.jobLog)
-            lines = h.readlines()
-            h.close()
-            return lines
-        def extract(line):
-            values, unit = [t.strip() for t in line.split('(')[1].split(')')]
-            value, error = [t.strip() for t in values.split('+/-')]
-            return float(value), float(error), unit
-        def extractSi2k(line):
-            si2kConst, rest = [t.strip() for t in line.split('Si2k:')[1].split('=>')]
-            normCPU, normCPUunits = [t.strip() for t in rest.split(':')[1].strip().split()]
-            return float(si2kConst), float(normCPU)
-        def extractVmemFit(line):
-            toks      = [t.strip() for t in line.split('INFO 000: [vmem-fit]')[1].strip().split()]
-            memleak   = toks[2]
-            chi2, ndf = toks[-1].split('/')
-            return float(memleak), float(chi2), int(ndf)
-            
-        # grab the log contents
-        lines = readLog()
-
-        start = ["Py:PerfMonSvc", "Statistics for 'evt':"]
-        stop  = ["Py:PerfMonSvc", "Statistics for 'fin':"]
-
-        # reject jobs that don't have PerfMon on, or that
-        # are chained jobs (for now, we deal with only single jobs)
-        found = []
-        for i,l in enumerate(lines):
-            found_it = True
-            for s in start:
-                if s not in l:
-                    found_it = False
-                    break
-            if found_it: found.append((i,l))
-            
-        # reject jobs that don't have PerfMon on, or that
-        # are chained jobs (for now, we deal with only single jobs)
-        # found =  [(i,l) for i,l in enumerate(lines) if l.find(start)!=-1]
-        if len(found) > 1:
-            m = "For now: not dealing with chained jobs"
-            self.log.panic(m)
-            raise NotYetException(m)
-
-        if len(found)==0:
-            m = "PerfMon not switched on in this job. Nothing to do."
-            self.log.panic(m)
-            raise NoPerfMonException(m)
-
-        # OK, let's deal with this job
-        startlineno, startline = found[0]
-
-        searchTemplate = "Py:PerfMonSvc        INFO %s:"
-        searchTerms = [{'offset':1, 'search':'<cpu>',    'label':'CPU per evt'},
-                       {'offset':2, 'search':'<real>',   'label':'Real time per evt'},
-                       {'offset':3, 'search':'<vmem>',   'label':'Vmem per evt'},
-                       {'offset':4, 'search':'<malloc>', 'label':'Malloc per evt'}]
-
-        dict = {}
-        
-        for searchTerm in searchTerms:
-            lineno = startlineno + searchTerm['offset']
-            line = lines[lineno]
-            if line.find(searchTemplate % searchTerm['search']) !=-1:
-                value, error, unit = extract(line)
-                dict[searchTerm['label']] = value
-                dict[searchTerm['label']+';error'] = error
-
-        # Si2K
-        lineno = startlineno + 10
-        line   = lines[lineno]
-        if line.find('Si2k:')!=-1:
-            si2kConst, normCPU      = extractSi2k(line)
-            dict['Si2k constant']   = si2kConst
-            dict['NormCPU per evt'] = normCPU
-
-        # Vmem fit
-        for line in lines[lineno:]:
-            if line.find('INFO 000: [vmem-fit]')!=-1:
-                memleak, chi2, ndf = extractVmemFit(line)
-                dict['MemLeak per evt'] = memleak
-                dict['MemLeak per evt:Chi2'] = chi2
-                dict['MemLeak per evt:NDF'] = ndf
-        return dict
-
-    def write(self):
-        if self.channelKey in self.schema:
-            channel = self.schema[self.channelKey]
-            terms = self.extractFromLog()
-
-            for key, newValue in terms.items():
-                self.log.info( "setting value '%s' for DozerData '%s'" % ( str(newValue), key) )
-                channel[key].setValue(newValue)
-        
-        # add to the DB
-        self.log.debug("will add DozerSchema as payload...")
-        self.db.addSchemaPayload( self.schema, DozerIOV().today() )
-        self.log.debug("done!")
-
-    def read(self):
-        try:
-            me = self.schema.pathChunk() + "/" + self.channelKey
-            recordSet = self.db.get( me, since=DozerIOV.lastNDays(30))
-        except:
-            recordSet = None
-            
-        if not recordSet: 
-            self.log.warn("Unable to retrieve channel with key: %s" % self.channelKey)
-            return
-
-        graph = recordSet.getDozerGraph()
-
-    
-        for plotable in graph.plotables():
-            
-            plotName = plotable.replace(" ", "_").lower()
-            if ';error' in plotName: continue
-            try:
-                graph.trend( plotable, True, "perfMonDozer_%s.png" % plotName )
-            except TypeError:
-                pass
-
-        
-## dummy test running 
-if __name__ == "__main__":
-      print "Running PerfMonDozer..."
-      class Paths:
-          def __init__(self):
-              self.release = 'rel_0'
-              self.branch = 'dev'
-              self.runType = 'build'
-              self.targetCMTCONFIG='i686-slc4-gcc34-opt'
-              self.topProject = 'AtlasProduction'
-              
-      class Desc:
-          def __init__(self):
-              self.log = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results/rel_0/dev/build/i686-slc4-gcc34-opt/offline/CoolDozer/AthenaCoolDozer/CoolDozer_fakePerfMon_jobOptions/1/CoolDozer_fakePerfMon_jobOptions1_log'
-              self.package    = 'CoolDozer'
-              self.trendId    = 'ZZZ_BrinickTestJob'
-              self.paths      = Paths()
-              
-      testCfg = { "JobDescriptor"    : Desc(),
-                  "DryRun"           : True }
-      dozerApp = PerfMonDozer( testCfg )
-      print 'Initialisation done; calling run()'
-      dozerApp.run()
-      print 'Done'
diff --git a/Tools/RunTimeTester/share/PerfMonRunner.py b/Tools/RunTimeTester/share/PerfMonRunner.py
deleted file mode 100644
index 303e77bc9d7109c3a4bace0e2b57491f0f122564..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/PerfMonRunner.py
+++ /dev/null
@@ -1,44 +0,0 @@
-from Logger import Logger
-# from Factory_CmtLines import CmtLinesFactory
-from ShellCommand import ShellCommand
-import os
-import os.path
-
-class PerfMonRunner:
-    def __init__(self, argDict):
-        desc  = argDict.get('JobDescriptor')
-        paths = desc.paths
-        self.runPath  = desc.runPath
-        self.fileName = argDict['fileName']
-        self.options  = argDict.get('options', '')
-        self.logger   = Logger()
-        self.envSetupCmds = argDict['cmtLinesCmds']
-        # self.envSetupCmds = CmtLinesFactory(self.logger).create(paths).lines()
-        
-    def run(self):
-        cmds = []
-        cmds.extend(self.envSetupCmds)
-        cmds.append('cd %s;perfmon.py %s %s' % (self.runPath, self.options, self.fileName))
-        
-        # launch the commands
-        sc    = ShellCommand(cmds, self.logger)            
-        reply = sc.getRawReply()
-        [self.logger.debug(str(l)) for l in reply]
-
-        handle = None
-        outFile = 'out.perfmon.stdout.txt'
-        outPath = os.path.join(self.runPath, outFile)
-            
-        try:
-            handle = open(outPath, 'w')
-            handle.writelines(reply)
-            handle.close()
-            self.logger.debug('Successfully wrote out: %s' % outPath)
-        except Exception, e:
-            m  = 'Unable to output results of checkFile to file: %s\n' % outPath
-            m += str(e)
-            self.logger.error(m)
-            if handle: handle.close()
-            return 1
-        
-        return 0
diff --git a/Tools/RunTimeTester/share/ROOTMacroRunner.py b/Tools/RunTimeTester/share/ROOTMacroRunner.py
deleted file mode 100755
index 14e1fe7360257c7c4950adbacb5bf292322c93b6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/ROOTMacroRunner.py
+++ /dev/null
@@ -1,141 +0,0 @@
-from Logger            import Logger
-import os, os.path
-import os, sys
-
-# from Factory_CmtLinesForScripts import CmtLinesForScriptsFactory
-
-class ROOTMacroRunner:
-    def __init__(self, argDict):
-        
-        jDescriptor       = argDict['JobDescriptor']
-        self.srcDir = jDescriptor.paths.RTTSrcDir
-        if self.srcDir not in sys.path: sys.path.append(self.srcDir)
-
-        self.macroArg            = argDict.get('macroArg', None)
-        self.macro               = argDict.get('macro', None)
-        # self.logger            = argDict['logger']
-        self.logger              = Logger()
-        
-        self.identifiedName      = jDescriptor.identifiedName
-        self.runPath             = jDescriptor.runPath
-
-        paths                    = jDescriptor.paths
-        self.runType             = paths.runType
-        self.distArea            = paths.distArea
-        self.branch              = paths.branch
-        self.release             = paths.release
-        self.workDirs            = paths.workDirs
-
-        self.nightlyPathFragment = ''
-        if self.branch != 'Numbered':
-            self.nightlyPathFragment = 'nightlies/%s' % self.branch
-
-        self.fullPathToMacro     = os.path.join(jDescriptor.runPath, self.macro)
-        self.envSetupCmds        = argDict['cmtLinesCmds']
-        # self.envSetupCmds        = CmtLinesForScriptsFactory(self.logger).create(jDescriptor).lines()
-        self.aclickTweakMacro    = os.path.join(paths.RTTLibDir, 'aclicTweak.C')
-
-        try:
-            compileTheMacro = argDict['compileTheMacro']
-        except:
-            compileTheMacro = False
-
-        self.compileFlag = ''
-        if compileTheMacro: self.compileFlag = '+'
-        
-        self.maxPrintLines = 500
-        self.timeLimit     = 60*60 # secs
-
-        self.rootCmds      = self.makeRootCommand()
-
-        msg  = 'root command for job '
-        msg += str(self.identifiedName)
-        msg += ':\n' + str(self.rootCmds)
-        self.logger.debug(msg)
-
-    def macroExists(self):
-        if not os.path.exists(self.fullPathToMacro):
-            msg  = 'Path to ROOT macro does not exist: %s' % self.fullPathToMacro
-            self.logger.error(msg)
-            return False
-        return True    
-        
-    def run(self):
-        # check the macro exists. NOTE: this cannot be done in __init__ because
-        # class instantiation occurs before copying of auxFilePatterns.
-        if not self.macroExists():
-            self.logger.warning("Root macro does not exit")
-            return 1
-        
-
-        # note the current directory to reset it later
-        currCwd = os.getcwd()
-
-        # cmds = []
-        # following cd must be part of the shell command has so other threads
-        # cant mess it up
-        # cmds.append('cd ' + self.runPath) 
-        # cmds.extend(self.rootCmds)
-        self.logger.info("Launching macro "+str(self.rootCmds))
-
-        # to import, need the RTT src dir. tried to do the import in __init__.
-        # Could import there, but ShellCommand was not visible here.
-        from ShellCommand import ShellCommand
-        sc    = ShellCommand(self.rootCmds,
-                             self.logger,
-                             self.maxPrintLines,
-                             self.timeLimit)
-
-        reply = sc.getReply()
-        [self.logger.debug(str(l)) for l in reply]
-        # reset the current directory as it was before
-        os.chdir(currCwd)
-
-        # reply is a list of strings, let's turn it into a single string
-        replyString = reduce(lambda x,y: x+y+'\n', reply,'')
-
-        # output reply to the logger
-        name = str(self.identifiedName)
-        msg  = 'root command output for job %s\n %s' % (name, replyString)
-        self.logger.info(msg)
-
-        # Also output reply to file 'ROOToutput.log' in the run path
-        # This file is a keep file pattern in the TopLevelJobGroup.xml file.
-        h = open(os.path.join(self.runPath,'ROOToutput.log'),'a+')
-        h.write(replyString)
-        h.close()
-
-        return 0
-
-    def makeRootCommand(self):
-        rootCmds = self.envSetupCmds
-        # rootCmds.append('source setup.sh >& /dev/null')
-        # must cd to runpath just before root call (threading)
-        rootCmds.append('cd %s' % self.runPath)
-
-        if self.macro:
-            # aclickTweak.C: a K.Ciba macro to allow 32 bit compilation in ROOT
-            # on a 64 bit native platform
-            line = 'root -b -q %s %s%s' % (self.aclickTweakMacro, self.macro, self.compileFlag)
-            if self.macroArg:
-                self.macroArg = self.macroArg.replace(' ', '\\ ')
-                line += '(%s)' % self.macroArg
-            line = self.escapeTheCommand(line)
-            rootCmds.append(line)
-
-        return rootCmds
-     
-    def escapeTheCommand(self, cmd):
-
-        self.toBeEscaped = ['(',')','"']
-        
-        for char in self.toBeEscaped:
-            pieces = cmd.split(char)
-            
-            cmd = pieces[0]
-            for piece in pieces[1:]:
-                cmd += '\\'+char+piece
-
-        return cmd
-
-
diff --git a/Tools/RunTimeTester/share/RTTDozerUtils.py b/Tools/RunTimeTester/share/RTTDozerUtils.py
deleted file mode 100644
index 66ac06f71e37a9e82ee8fc6e5de9178f9abc9afb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTTDozerUtils.py
+++ /dev/null
@@ -1,64 +0,0 @@
-from CoolDozer import *
-import os.path
-
-def init(func):
-    def inner(self, dict={}):        
-        self.jd = dict['JobDescriptor']
-        self.log = DozerLogger("%s.log" % self.__class__.__name__, self.__class__.__name__, toConsole=True).getLogger(self)
-        return func(self, dict)
-    return inner
-
-def run(func):
-    def inner(self):
-        def openDB(): 
-            isOpened = None
-            try:
-                isOpened = self.db.openDB() 
-            except DozerError, value:
-                self.log.epanic( value )
-            return isOpened
-        def closeDB():  
-            self.db.closeDB()        
-        def getSchema():
-            if not hasattr(self, 'db'):
-                raise 
-            if not self.db.isOpened():
-                raise
-            #try:
-                
-                #self.schema = self.db.getSchema(self.dbkey)
-            #except DozerError, value:
-            #    self.log.info("Schema %s hasn't been added to the database yet!" % self.dbkey)
-                            
-            # create new schema if necessary
-            self.schema =  self.createSchema() #or self.schema
-    
-            if not self.schema:
-                self.log.panic("No schema for config = '%s' found in DozerConfig " % self.dbkey )
-                sys.exit(0)
-
-            return self.schema
-            #self.db += self.schema
-
-        self.db = DozerDB(self.jd.package, "RTT", False, self.log )    
-        self.dbkey = os.path.join(self.jd.paths.branch, self.jd.paths.runType, self.jd.paths.cmtConfig, self.jd.paths.topProject, self.jobId) # db key
-        self.log.info("dbkey is %s" % self.dbkey )
-        
-        # try and open the DB
-        if not openDB(): 
-            self.log.panic('Unable to get a connection to the DB. Abandon ship...')
-            raise
-
-        # all ok, add schema to db
-        self.db += getSchema()
-        ret = None
-        try:
-            ret = func(self)
-        except:
-            self.log.epanic("%s threw an exception!" % func.__name__)
-        finally:
-            closeDB()
-        return ret
-    return inner
-
-
diff --git a/Tools/RunTimeTester/share/RTTExplosion.py b/Tools/RunTimeTester/share/RTTExplosion.py
deleted file mode 100755
index 0fa414eb507de73692437bd5507db382cc518cea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTTExplosion.py
+++ /dev/null
@@ -1,93 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-import os
-from Logger import Logger
-
-import string, sys, traceback
-
-def exc2string2():
-    list   = sys.exc_info()
-    errmsg  = str(list[0])
-    errmsg += str(list[1])
-    errmsg += string.join(traceback.format_tb(list[2]))
-    return errmsg
-                
-
-class RTTExplosion:
-    """An RTT test to run its own TestSuite under RTT control
-    This test checks that classes can be instantitiated. It also
-    creates the databass needed to run RTTRegression"""
-    
-
-    def __init__(self, argDict):
-
-
-        self.success =  0
-        self.error   = -1
-
-        # self.logger = self.argDict['logger']
-        self.logger = Logger()
-        msg = 'Instantiating RTTExplosion, args: %s' %str(argDict)
-        self.logger.debug(msg)
-        jDescriptor = argDict['JobDescriptor']
-        # fixtureDir is set in JobsXMLReader when reading in the config file.
-        self.fixtureDir = argDict['fixtureDir']
-        # the current directory
-        self.runPath   =  jDescriptor.runPath
-        self.logDir    =  jDescriptor.runPath
-        # directory of the source code under test
-        self.rttSrcDir =  os.path.join(self.runPath, 'Tools/RunTimeTester/src')
-        fixtureDir    =  os.path.basename(self.fixtureDir)
-
-    def run(self):
-
-        # prepend the source directory undertest (must be prepended,
-        # otherwise the RTT source used to run the job will be used
-        # for importing!
-        list = [self.rttSrcDir]
-        list.extend(sys.path)
-        sys.path = list
-        self.logger.debug('sys.path = %s' % str(sys.path))
-        ## horrible hack until I figure out how to get CMT to do this
-        temp = ['/afs/cern.ch/atlas/offline/external/PyXML/0.8.4-2.3.4/PyXML-0.8.4-2.3.4']
-        temp.extend(sys.path)
-        sys.path = temp
-        os.chdir(self.rttSrcDir)
-        # sys.path.append(self.rttSrcDir)
-
-        try:
-            from TestSuiteRun2 import TestSuiteRun
-            self.logger.debug('RTTExplosion: Import of TestSuiteRun succeeded')
-        except Exception, e:
-            self.logger.error('RTTExplosion: Import of TestSuiteRun failed')
-            list    = sys.exc_info()
-            errmsg  = str(list[0])
-            errmsg += str(list[1])
-            errmsg += string.join(traceback.format_tb(list[2]))
-            self.logger.error(errmsg)
-            return self.error
-
-        try:
-            tsr = TestSuiteRun(self.rttSrcDir, self.fixtureDir, self.logDir)
-            self.logger.debug('RTTExplosion: TestSuiteRun execution complete')
-        except Exception, e:
-            m = 'RTTExplosion: TestSuiteRun execution failed \n%s' % exc2string2()
-            self.logger.error(m)
-            
-            return self.error
-
-
-        if tsr.passed:
-            self.logger.debug('RTTExplosion Test passed')
-            return self.success
-        self.logger.debug('RTTExplosion Test failed')
-        return self.error
-        
diff --git a/Tools/RunTimeTester/share/RTTRegression.py b/Tools/RunTimeTester/share/RTTRegression.py
deleted file mode 100755
index 19c13f5f565c3c5ef12dd5971cc99fc2fc4f666e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTTRegression.py
+++ /dev/null
@@ -1,147 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-import os.path, anydbm
-
-class RTTRegression:
-    """An RTT test to run its own TestSuite under RTT control
-    This test checks that classes can be instantitiated. It also
-    creates the databass needed to run RTTRegression"""
-    
-
-    def __init__(self, argDict):
-
-
-        self.success =  0
-        self.error   = -1
-
-        # self.logger = argDict['logger']
-        self.logger  = Logger()
-        msg = 'Instantiating RTTRegression, args: %s' %str(argDict)
-        self.logger.debug(msg)
-        # fixtureDir is set in JobsXMLReader when reading in the config file.
-        self.fixtureDir = argDict['fixtureDir']
-        # the current directory
-        jDescriptor    = argDict['JobDescriptor']
-        self.runPath   =  jDescriptor.runPath
-        self.logDir    =  jDescriptor.runPath
-        # directory of the source code under test
-        self.rttSrcDir =  os.path.join(self.runPath, 'Tools/RunTimeTester/src')
-        self.runPath   =  jDescriptor.runPath
-        fixture        =  os.path.basename(self.fixtureDir)
-        self.dbName    =  os.path.join(self.runPath, fixture+'.db')
-        self.refdbName =  os.path.join(self.runPath, 'refFile_'+fixture+'.db')
-        # do not open the dir does not exist yet
-        self.ofName    =  os.path.join(self.runPath, fixture+'_regression.log')
-
-        
-    def run(self):
-
-        outFile = open(self.ofName, 'w')
-
-        if not os.path.exists(self.dbName):
-            msg = 'None existant path: %s' %  self.dbName
-            self.logger.error(msg)
-            outFile.write(msg+'\n')
-            outFile.close()
-            
-            return self.error
-        if not os.path.exists(self.refdbName):
-            msg = 'None existant path: %s' %  self.refdbName
-            self.logger.error(msg)
-            outFile.write(msg+'\n')
-            outFile.close()
-            
-            return self.error
-
-        newDB = anydbm.open(self.dbName,    'r')
-        oldDB = anydbm.open(self.refdbName, 'r')
-        result = self.success
-            
-    
-        onlyInNew = [k for k in newDB.keys() if k not in oldDB.keys()]
-
-        text = 'Number of keys in reference db %d\n' % len(oldDB.keys())
-        text = 'Number of keys in new       db %d\n' % len(newDB.keys())
-        
-        
-        if onlyInNew:
-
-            result = self.error
-    
-            text += '\n'
-            text +='Reference - %s: date: %s\n' % (oldDB['fixtureDir'],
-                                                   oldDB['date'])
-            
-            text +=  'New       - %s: date: %s\n' % (newDB['fixtureDir'],
-                                                     newDB['date'])
-            
-            text += '\n'
-            text +=  ' keys in new database, but not in old\n'
-            text +=  str(onlyInNew)+'\n'
-            text += '\n'
-    
-        onlyInOld = [k for k in oldDB.keys() if k not in newDB.keys()]
-        if onlyInOld:
-
-            result = self.error
-        
-            text +=  '\n'
-            text +=  ' keys in old database, but not in new\n'
-            text +=  str(onlyInOld)+'\n'
-            text += '\n'
-    
-        keys = [k for k in oldDB.keys() if k in newDB.keys()]
-        toRemove = ['fixtureDir', 'date']
-        [keys.remove(k) for k in toRemove if k in keys]
-        if keys:
-            text +=  'differences:\n'
-            text += '\n'
-
-            for k in keys:
-                
-        
-                if oldDB[k] != newDB[k]:
-
-                    result = self.error
-                    
-                    text +=  'Key: %s\n' % k
-                    text += '\n'
-                    text +=  '   old:\n'
-                    text +=  '   ' +str(oldDB[k])+'\n'
-                    text += '\n'
-                    text +=  '   new:\n'
-                    text +=  '   '+str(newDB[k])+'\n'
-
-
-        text += '\n'
-
-        totTests = 0
-        text +=  'Number of points examined:\n'
-        for k in keys:
-            line = ''
-            line += k.ljust(30)
-            ntestOld = len(oldDB[k].split(','))
-            ntestNew = len(newDB[k].split(','))
-            # assert(ntestOld == ntestNew)
-            num = '%d' % ntestOld
-            line += num.ljust(5)
-            # print line
-            totTests += ntestOld
-
-        text +=  'No of test classes which pass: %d\n' % len(keys)
-        text +=  'Total number of tests passed:  %d\n ' %totTests
-
-        outFile.write(text)
-        outFile.close()
-        return result
-        
-        
-
diff --git a/Tools/RunTimeTester/share/RTTTimePlot.cxx b/Tools/RunTimeTester/share/RTTTimePlot.cxx
deleted file mode 100755
index 46674d9536ee869d5077d9d9a31d48b64b9ce2f1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTTTimePlot.cxx
+++ /dev/null
@@ -1,106 +0,0 @@
-#include "TPostScript.h"
-#include "TH1F.h"
-#include "TFile.h"
-#include "TCanvas.h"
-#include "TStyle.h"
-#include <iostream>
-#include <vector>
-#include <algorithm>
-#include <fstream>
-#include <string>
-#include <cstdlib>
-#include <stdlib.h>
- 
-std::vector<std::vector<int> > readData(const char* fn){
-
-  std::ifstream listIn(fn);
-  std::vector<int> values, seconds;
-  int value, second;
-  std::cout<<"reading data"<<"\n";
-  while (listIn>>second>>value) {
-    std::cout<<"dataL "<<value<<" second "<<second<<"\n";
-    values.push_back(value);
-    seconds.push_back(second);
-  }
-  std::vector<std::vector<int> > data;
-  data.push_back(values);
-  data.push_back(seconds);
-
-  return data;
-
-}
-  
-
-void printInt(int t){ std::cout<<t<<std::endl;}
-void printVec(std::vector<int> t){std::for_each(t.begin(), t.end(), printInt);}
-
-int RTTTimePlot(const char* title){
-
-  const char* fn = (std::string(title)+".dat").c_str();
-
-  std::cout<<"hello"<<std::endl;
-  std::vector<std::vector<int> > data = readData(fn);
-  std::cout<<"hello2"<<std::endl;
-
-
-  
-  std::vector<int> values    = data[0];
-  std::vector<int> seconds   = data[1];
-
-  double mintime = *std::min_element(seconds.begin(), seconds.end());
-  double maxtime = *std::max_element(seconds.begin(), seconds.end());
-  // std::cout<<"mintime"<<" "<<mintime<<std::endl;
-  // std::cout<<"maxtime"<<" "<<maxtime<<std::endl;
- 
-  mintime = mintime;
-  maxtime = maxtime;
-
-  double toff = mintime;  
-  gStyle->SetTimeOffset(toff);  
-  mintime =mintime - toff;
-  int secperday = 24*3600;
-  maxtime =maxtime - toff+secperday;
-
-  // std::cout<<"mintime"<<" "<<mintime<<std::endl;
-  // std::cout<<"maxtime"<<" "<<maxtime<<std::endl;
-  int ndays = int((maxtime - mintime)/secperday);
-
-  
-
-  TH1F* h1 = new TH1F("h1", title,   ndays, mintime, maxtime);
-
-  std::vector<int>::const_iterator iterv = values.begin();
-  std::vector<int>::const_iterator iters = seconds.begin();
-
-  for(;iterv<values.end();++iterv){
-    // std::cout<<*iterl<<" "<<*iterh<<" "<<*iters<<std::ends;
-    double v = *iterv;
-    double s = *iters - toff;
-    std::cout<<v<<" "<<s<<std::endl;
-    h1->Fill(s, v);
-    
-    ++iters;
-  }
-
-
-  h1->GetXaxis()->SetTimeDisplay(1); 
-  h1->GetXaxis()->SetTimeFormat("%d/%m/%y");
-
-  const char* title2 = (std::string(title)+std::string(".gif")).c_str();
-  // TPostScript ps(title2, 112);
-  
-  TCanvas* canvas = new TCanvas(title, title, 1);
-  canvas -> Divide(1,1);
-  canvas->cd(1);
-  h1->SetStats(kFALSE);
-  h1->GetXaxis()->SetTitle("NICOS build date");  
-  h1->GetXaxis()->SetNdivisions(-ndays);
-  h1->GetYaxis()->SetTitle("Number");  
-  h1->SetFillColor(42);
-  h1->Draw();
-
-  // ps.Close();
-  canvas->Print(title2);
-  return 0;   
-  
-}
diff --git a/Tools/RunTimeTester/share/RTTTimePlot2.cxx b/Tools/RunTimeTester/share/RTTTimePlot2.cxx
deleted file mode 100755
index 21c018a14da52a3c0143592084c5aa809f7a53c9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTTTimePlot2.cxx
+++ /dev/null
@@ -1,113 +0,0 @@
-#include "TPostScript.h"
-#include "TH1F.h"
-#include "TFile.h"
-#include "TCanvas.h"
-#include "TStyle.h"
-#include <iostream>
-#include <vector>
-#include <algorithm>
-#include <fstream>
-#include <string>
-#include <cstdlib>
-#include <stdlib.h>
- 
-const std::vector<int>* readData(const char* fn){
-
-  std::ifstream listIn(fn);
-  std::vector<int>*  dataLow = new vector<int>;
-  std::vector<int>*  dataHi  = new vector<int>;
-  std::vector<int>*  seconds = new vector<int>;
-  int dataL, dataH, second;
-  while (listIn>>second>>dataH>>dataL) {
-    //std::cout<<"dataL "<<dataL<<" dataH "<<dataH<<" second "<<second<<"\n";
-    dataLow->push_back(dataL);
-    dataHi->push_back(dataH);
-    seconds->push_back(second);
-  }
-  //std::cout<<"hello3"<<std::endl;
-  const std::vector<int>* vectors = new std::vector<int>[3] ;
-  vectors[0] = dataLow;
-  vectors[1] = dataHi;
-  vectors[2] = seconds;
-  return vectors;
-}
-  
-
-void printInt(int t){ std::cout<<t<<std::endl;}
-void printVec(std::vector<int> t){std::for_each(t.begin(), t.end(), printInt);}
-
-//int RTTTimePlot2(const char* fn, const char* title){
-int RTTTimePlot2(const char* title){
-
-  const char* fn = (std::string(title)+".dat").c_str();
-
-  std::vector<int>* data = readData(fn);
-
-
-  std::vector<int> low       = data[0];
-  std::vector<int> high      = data[1];
-  std::vector<int> seconds   = data[2];
-
-  double mintime = *std::min_element(seconds.begin(), seconds.end());
-  double maxtime = *std::max_element(seconds.begin(), seconds.end());
-
-  double toff = mintime;  
-  gStyle->SetTimeOffset(toff);  
-  mintime =mintime - toff;
-  int secperday = 24*3600;
-  maxtime =maxtime - toff+secperday;
-
-  int ndays = int((maxtime - mintime)/secperday);
-
-  
-
-  TH1F* h1 = new TH1F("h1", title,   ndays, mintime, maxtime);
-  TH1F* h2 = new TH1F("h2", title,   ndays, mintime, maxtime);
-
-  std::vector<int>::const_iterator iterl = low.begin();
-  std::vector<int>::const_iterator iterh = high.begin();
-  std::vector<int>::const_iterator iters = seconds.begin();
-
-  for(;iterl<low.end();++iterl){
-    
-    double l = *iterl;
-    double h = *iterh;
-    double s = *iters - toff;
-    h1->Fill(s, h);
-    h2->Fill(s, l);
-    
-    ++iterh;
-    ++iters;
-  }
-
-
-  h1->GetXaxis()->SetTimeDisplay(1); 
-  h1->GetXaxis()->SetTimeFormat("%d/%m/%y");
-  h1->GetXaxis()->SetNdivisions(-ndays/3);
-  
-  h2->GetXaxis()->SetTimeDisplay(1); 
-  h2->GetXaxis()->SetTimeFormat("%d/%m/%y");
-  h2->GetXaxis()->SetNdivisions(-ndays/3);
-
-  //const char* title2 = (std::string(title)+std::string(".gif")).c_str();
-  //std::cout<<"printing 0 "<< title2<<std::endl;
-
-  // TPostScript ps(title2, 113);
-  TCanvas* canvas = new TCanvas(title, title, 1);
-  canvas -> Divide(1,1);
-  canvas->cd(1);
-  h1->SetStats(kFALSE);
-  h2->SetStats(kFALSE);
-  h2->SetFillColor(142);
-  h1->GetXaxis()->SetTitle("NICOS build date");  
-  h1->GetYaxis()->SetTitle("OK/All");  
-  h1->SetFillColor(42);
-  h1->Draw();
-  h2->Draw("SAME");
-
-  // ps.Close();
-  const char* title2 = (std::string(title)+std::string(".gif")).c_str();
-  canvas->Print(title2);
-  return 0;   
-  
-}
diff --git a/Tools/RunTimeTester/share/RTT_atlfast.C b/Tools/RunTimeTester/share/RTT_atlfast.C
deleted file mode 100755
index dc2fc675a88f8095092bf72bb0fd49605b9173ed..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTT_atlfast.C
+++ /dev/null
@@ -1,426 +0,0 @@
-class AtlHists{
-public:
-  AtlHists(char* fn){
-    m_file =new TFile(fn);
-
-    this->nonIsolatedElectronHists();    
-    this->isolatedElectronHists();
-    this->nonIsolatedPhotonHists();
-    this->isolatedPhotonHists();
-    this->nonIsolatedMuonHists();
-    this->isolatedMuonHists();
-    this->nonIsolatedMuonHists();
-    this->isolatedMuonHists();
-    this->cellHists();
-    this->clusterHists();
-    this->jetHists();
-    this->jetBHists();
-    this->eventHeaderHists1();
-    this->eventHeaderHists2();
-  }
-
-  void nonIsolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/1");
-    m_h2 = (TH1F*)m_file->Get("simple/2");
-    m_h3 = (TH1F*)m_file->Get("simple/3");
-    m_h4 = (TH1F*)m_file->Get("simple/4");
-    m_h5 = (TH1F*)m_file->Get("simple/5");
-    m_h6 = (TH1F*)m_file->Get("simple/6");
-    
-    TPostScript ps("nonIsolatedElectron.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-
-  void isolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/7");
-    m_h2 = (TH1F*)m_file->Get("simple/8");
-    m_h3 = (TH1F*)m_file->Get("simple/9");
-    m_h4 = (TH1F*)m_file->Get("simple/10");
-    m_h5 = (TH1F*)m_file->Get("simple/11");
-    m_h6 = (TH1F*)m_file->Get("simple/12");
-
-    TPostScript ps("isolatedElectron.ps", 112);
-    this->Draw();
-    ps.Close();
-
-  }
-    
-
-  void nonIsolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/13");
-    m_h2 = (TH1F*)m_file->Get("simple/14");
-    m_h3 = (TH1F*)m_file->Get("simple/15");
-    m_h4 = (TH1F*)m_file->Get("simple/16");
-    m_h5 = (TH1F*)m_file->Get("simple/17");
-    m_h6 = (TH1F*)m_file->Get("simple/18");
-
-    TPostScript ps("nonIsolatedPhoton.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-    return;
-  }
-
-  void isolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/19");
-    m_h2 = (TH1F*)m_file->Get("simple/20");
-    m_h3 = (TH1F*)m_file->Get("simple/21");
-    m_h4 = (TH1F*)m_file->Get("simple/22");
-    m_h5 = (TH1F*)m_file->Get("simple/23");
-    m_h6 = (TH1F*)m_file->Get("simple/24");
-
-    TPostScript ps("isolatedPhoton.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-    
-
-  void nonIsolatedMuonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/25");
-    m_h2 = (TH1F*)m_file->Get("simple/26");
-    m_h3 = (TH1F*)m_file->Get("simple/27");
-    m_h4 = (TH1F*)m_file->Get("simple/28");
-    m_h5 = (TH1F*)m_file->Get("simple/29");
-    m_h6 = (TH1F*)m_file->Get("simple/30");
-
-
-    TPostScript ps("nonIsolatedMuon.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-    return;
-  }
-
-  void isolatedMuonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/31");
-    m_h2 = (TH1F*)m_file->Get("simple/32");
-    m_h3 = (TH1F*)m_file->Get("simple/33");
-    m_h4 = (TH1F*)m_file->Get("simple/34");
-    m_h5 = (TH1F*)m_file->Get("simple/35");
-    m_h6 = (TH1F*)m_file->Get("simple/36");
-
-
-    TPostScript ps("isolatedMuon.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-    
-
-  void cellHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/37");
-    m_h2 = (TH1F*)m_file->Get("simple/38");
-    m_h3 = (TH1F*)m_file->Get("simple/39");
-    m_h4 = (TH1F*)m_file->Get("simple/40");
-    m_h5 = (TH1F*)m_file->Get("simple/41");
-    m_h6 = (TH1F*)m_file->Get("simple/42");
-
-    TPostScript ps("cells.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-
-  void clusterHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/43");
-    m_h2 = (TH1F*)m_file->Get("simple/44");
-    m_h3 = (TH1F*)m_file->Get("simple/45");
-    m_h4 = (TH1F*)m_file->Get("simple/46");
-    m_h5 = (TH1F*)m_file->Get("simple/47");
-    m_h6 = (TH1F*)m_file->Get("simple/48");
-
-    TPostScript ps("clusters.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void jetHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/49");
-    m_h2 = (TH1F*)m_file->Get("simple/50");
-    m_h3 = (TH1F*)m_file->Get("simple/51");
-    m_h4 = (TH1F*)m_file->Get("simple/52");
-    m_h5 = (TH1F*)m_file->Get("simple/53");
-    m_h6 = (TH1F*)m_file->Get("simple/54");
-
-
-    TPostScript ps("jets.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-  void jetBHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/55");
-    m_h2 = (TH1F*)m_file->Get("simple/56");
-    m_h3 = (TH1F*)m_file->Get("simple/57");
-    m_h4 = (TH1F*)m_file->Get("simple/58");
-    m_h5 = (TH1F*)m_file->Get("simple/59");
-    m_h6 = (TH1F*)m_file->Get("simple/60");
-
-    TPostScript ps("bjets.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-    
-
-  void eventHeaderHists1(){
-    m_h1 = (TH1F*)m_file->Get("simple/55");
-    m_h2 = (TH1F*)m_file->Get("simple/56");
-    m_h3 = (TH1F*)m_file->Get("simple/57");
-    m_h4 = (TH1F*)m_file->Get("simple/58");
-    m_h5 = (TH1F*)m_file->Get("simple/59");
-    m_h6 = (TH1F*)m_file->Get("simple/60");
-
-
-    TPostScript ps("eventHeader1.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void eventHeaderHists2(){
-    m_h1 = (TH1F*)m_file->Get("simple/61");
-    m_h2 = (TH1F*)m_file->Get("simple/62");
-    m_h3 = (TH1F*)m_file->Get("simple/63");
-    m_h4 = new TH1F;
-    m_h5 = new TH1F;
-    m_h6 = new TH1F;
-
-    TPostScript ps("eventHeader2.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void Draw(){
-    TCanvas* can = new TCanvas;
-    can->Clear();
-    can->Divide(3,2);
-    can->cd(1);
-    m_h1->Draw();
-    can->cd(2);
-    m_h2->Draw();
-    can->cd(3);
-    m_h3->Draw();
-    can->cd(4);
-    m_h4->Draw();
-    can->cd(5);
-    m_h5->Draw();
-    can->cd(6);
-    m_h6->Draw();
-  }
-  void Print(TCanvas* c){
-    c->Print(fn);
-  }
- 
-private:
-  TH1F* m_h1;
-  TH1F* m_h2;
-  TH1F* m_h3;
-  TH1F* m_h4;
-  TH1F* m_h5;
-  TH1F* m_h6;
-  TFile* m_file;
-  
-};
-
-class HistoRegression{
-public:
- HistoRegression(char* file, char* refFile){
-   bool initOK = true;
-   m_file =new TFile(file);
-   if (m_file->IsZombie()) {
-     initOK = false;
-   }
-
-   m_refFile = new TFile(refFile);
-   if (m_refFile->IsZombie()) {
-     initOK = false;
-   }
-
-   cout <<"initOK "<<initOK<<endl;
-   if(initOK){
-     cout<<"performing ...."<<endl;
-     this->nonIsolatedElectronHists(); 
-     this->isolatedElectronHists();
-     this->nonIsolatedPhotonHists();
-     this->isolatedPhotonHists();
-   }
- }
-
-  void nonIsolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/1");
-    m_h2 = (TH1F*)m_file->Get("simple/2");
-    m_h3 = (TH1F*)m_file->Get("simple/3");
-    m_h4 = (TH1F*)m_file->Get("simple/4");
-    m_h5 = (TH1F*)m_file->Get("simple/5");
-    m_h6 = (TH1F*)m_file->Get("simple/6");
-
-    m_refh1 = (TH1F*)m_refFile->Get("simple/1");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/2");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/3");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/4");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/5");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/6");
-    
-    TPostScript ps("nonIsolatedElectron_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-  
-  void isolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/7");
-    m_h2 = (TH1F*)m_file->Get("simple/8");
-    m_h3 = (TH1F*)m_file->Get("simple/9");
-    m_h4 = (TH1F*)m_file->Get("simple/10");
-    m_h5 = (TH1F*)m_file->Get("simple/11");
-    m_h6 = (TH1F*)m_file->Get("simple/12");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/7");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/8");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/9");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/10");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/11");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/12");
-    TPostScript ps("isolatedElectron_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-  }
-    
-
-  void nonIsolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/13");
-    m_h2 = (TH1F*)m_file->Get("simple/14");
-    m_h3 = (TH1F*)m_file->Get("simple/15");
-    m_h4 = (TH1F*)m_file->Get("simple/16");
-    m_h5 = (TH1F*)m_file->Get("simple/17");
-    m_h6 = (TH1F*)m_file->Get("simple/18");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/13");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/14");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/15");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/16");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/17");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/18");
-    TPostScript ps("nonIsolatedPhoton_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-
-  void isolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/19");
-    m_h2 = (TH1F*)m_file->Get("simple/20");
-    m_h3 = (TH1F*)m_file->Get("simple/21");
-    m_h4 = (TH1F*)m_file->Get("simple/22");
-    m_h5 = (TH1F*)m_file->Get("simple/23");
-    m_h6 = (TH1F*)m_file->Get("simple/24");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/19");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/20");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/21");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/22");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/23");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/24");
-    TPostScript ps("isolatedPhoton_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-
-  void setLineColor(int i){
-    gStyle->SetHistLineColor(i)
-  }
-  
-  void Draw(){
-    TCanvas* can = new TCanvas;
-    can->Clear();
-    can->Divide(3,2);
-    can->cd(1);
-    setLineColor(1);
-    m_h1->Draw();
-    setLineColor(2);
-    m_refh1->Draw("same");
-    can->cd(2);
-    setLineColor(1);
-    m_h2->Draw();
-    setLineColor(2);
-    m_refh2->Draw("same");
-    can->cd(3);
-    setLineColor(1);
-    m_h3->Draw();
-    setLineColor(2);
-    m_refh3->Draw("same");
-    can->cd(4);
-    setLineColor(1);
-    m_h4->Draw();
-    setLineColor(2);
-    m_refh4->Draw("same");
-    can->cd(5);
-    setLineColor(1);
-    m_h5->Draw();
-    setLineColor(2);
-    m_refh5->Draw("same");
-    can->cd(6);
-    setLineColor(1);
-    m_h6->Draw();
-    setLineColor(2);
-    m_refh6->Draw("same");
-  }
-
-  void Print(TCanvas* c){
-    c->Print(fn);
-  }
- 
-private:
-  TH1F* m_h1;
-  TH1F* m_h2;
-  TH1F* m_h3;
-  TH1F* m_h4;
-  TH1F* m_h5;
-  TH1F* m_h6;
-  TH1F* m_refh1;
-  TH1F* m_refh2;
-  TH1F* m_refh3;
-  TH1F* m_refh4;
-  TH1F* m_refh5;
-  TH1F* m_refh6;
-  TFile* m_file;
-  TFile* m_refFile;
-
-};
-
-int RTT_atlfast(){
-   
-  //  cout<<"in RTT_atlfast.C"<<endl;
-  AtlHists a1("histo.hbook");
-
-  //  cout<<"doing regression histograms..."<<endl;
-  HistoRegression hr("histo.hbook","refHisto.hbook");
-
-  return 0;
-}
diff --git a/Tools/RunTimeTester/share/RTT_histAdder.C b/Tools/RunTimeTester/share/RTT_histAdder.C
deleted file mode 100755
index 18ba22bdc454d32d6867f2d9873b88430be759c0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTT_histAdder.C
+++ /dev/null
@@ -1,198 +0,0 @@
-//********************************
-// Root macro to read in and sum histograms from a number of files,
-// and to result the sum back out
-//
-// Needs more work to ensure character buffers do not overflow.
-// PS 10/04
-//********************************
-
-
-TObjArray* getHists(char* fname){
-  //printf("getHists: start\n");
-  TObjArray* hists = new TObjArray(0);
-  int size = hists->GetSize();
-  //printf("getHists: size on creation: %d\n",size);
-  
-  TFile* f = new TFile(fname);
-
-  TIter next(f->GetListOfKeys());
-  TKey *key;
-  while ((key=(TKey*)next())) {
-    char* name = key->GetName();
-    //printf("getHists: obj name %s\n", name);
-    TObject* obj = f->Get(name);
-    hists->Add(obj);
-  }  
-  
-  int size = calcSize(hists);
-  printf("getHists: from %s found %d hists\n", fname, size);
-  return hists;
-}
-
-int calcSize(TObjArray* col){
-  //printf("calcSize: start:");
-  int count = 0;
-  TIter iter = col->MakeIterator();
-  while(iter.Next()){++count;}
-  return count;
-}
-bool addHists(char* fname, TObjArray* hsums, char* rfname, TObjArray* rsums){
-  TObjArray* hists  = getHists(fname);
-  TObjArray* rhists = getHists(rfname);
-
-  int hsize = calcSize(hists);
-  int ssize = calcSize(hsums);
-  int rsize = calcSize(rhists);
-
-  if(hsize != ssize || hsize != rsize){
-    printf("unequal numbers of hists: new h, new r, sum: %d %d %d",
-	   hsize, rsize, ssize);
-    printf("skipping %s\n", fname);
-    return 0;
-  }
-
-
-  accumulateHists(hists, hsums);
-  accumulateHists(rhists, rsums);
-
-  return 1;
-}
-
-void accumulateHists(TObjArray* hists, TObjArray* hsums){
-  TIter h_iter(hists);
-  TIter s_iter(hsums);
-  TH1* h;
-  while( h = (TH1*)h_iter.Next()){
-    s = (TH1*)s_iter.Next();
-    s->Add(h);
-  }
-}
-
-void writeHists(char* sumfile, TObjArray* hists){
-  cout<<"writing to..."<<sumfile<<endl;
-  //printf("number of hists to write: %d\n",calcSize(hists));
-  ofile = new TFile(sumfile, "recreate");
-  //printf("number of hists to write: %d\n",calcSize(hists));
-  hists->Write();
-  ofile->Close();
-}
-
-int nameWithoutCC(char* name, int n, FILE* fp){
-  //printf("nameWithoutCC: start \n");
-  char temp[201];    //buffer size now doubled 10-03-05.
-  char* s;
-
-  if (s = fgets(&temp, n, fp)){
-    //printf("before shortening %s", &temp);
-    //////////////////////////////////
-    // Mac only!!!!
-    int nchar = strlen(s)-1;
-    s = strncpy(name, &temp, nchar);
-    //////////////////////////////////
-    //////////////////////////////////
-    // Linux
-    // s = strcpy(name, &temp);
-    //////////////////////////////////
-    //printf("%d\n", nchar);
-    //printf("\n");
-    // printf("nameWithoutCC: end1: %s\n", s);
-    return 1;
-  }
-  // printf("nameWithoutCC: end2\n");
-
-  return 0;
-}
-bool fillInFileName(char* lp, char* fn){
-  int csize = strlen(lp)+strlen(fn)+strlen("/");
-  if (csize>200){
-    printf("lp character buffer too short  need %d",csize);
-    return 0;
-  }
-  strcat(lp,"/");
-  strcat(lp,fn);
-  return 1;
-}
-
-bool fillInRefFileName(char* lp, char* fn){
-  int csize = strlen(lp)+strlen(fn)+strlen("/refFile_");
-  if (csize>200){
-    printf("lp character buffer too short  need %d",csize);
-    return 0;
-  }
-  strcat(lp,"/refFile_");
-  strcat(lp,fn);
-  return 1;
-}
-
-
-int RTT_histAdder(char* fname){
-
-  //printf("histAddr: start\n");
-  //printf("histAddr: %s\n", fname);
-    
-  FILE *fp = fopen("histLocation.txt","r");     
-
-
-  TObjArray* hists  = new TObjArray(0);
-  TObjArray* rhists = new TObjArray(0);
-  int fpass = 1;
-
-  char location[201];
-  char* lp = &location;
-  char rlocation[201];
-  char* rlp = &rlocation;
-
-  while (nameWithoutCC(lp, 200, fp)) {
-
-    if (strlen(lp) >200){printf("Beware: string buffers are full!");}
-
-    strcpy(rlp, lp);
-
-    if (!(fillInFileName(lp, fname) && fillInRefFileName(rlp, fname))){
-      printf("problem with name lengths, skipping %s\n", lp);
-      continue;
-    }
-    printf("RTT_histAdder hist name %s\n",lp);
-    printf("RTT_histAdder ref hist name %s\n",rlp);
-    
-    if(fpass){
-      fpass =0;
-
-      hists  = getHists(lp);
-      rhists = getHists(rlp);
-
-      if(calcSize(hists) != calcSize(rhists)){
-	printf("no of hists != in ref & new hist files, skipping %s\n",lp);
-	fpass = 1;
-      }
-
-      if(calcSize(hists) == 0 ){
-	printf("no of hists in files = 0, skipping %s\n",lp);
-	fpass = 1;
-      }
-      
-      //printf("RTT_histAdder: fpass %d\n",fpass);
-    }else{
-      if(!addHists(lp, hists, rlp, rhists)){
-	printf("Problem accumulating hists %s\n",lp);
-      }
-    }
-    for(int i=0; i<201; ++i){location[i]  = '\0';}
-    for(int i=0; i<201; ++i){rlocation[i] = '\0';}
-  }
-  fclose(fp);
-
-  writeHists(fname, hists);
-
-  char refName[100]="refFile_";
-  char* rfname = &refName;
-  strcat(rfname, fname);
-
-  writeHists(rfname, rhists);
-
-  //int size = calcSize(hists);
-  //printf("RTT_histAdder size: %d\n", size);
-  hists->Delete();
-  delete hists;
-  return 0;
-}
diff --git a/Tools/RunTimeTester/share/RTT_histComparison.C b/Tools/RunTimeTester/share/RTT_histComparison.C
deleted file mode 100755
index a9a91b9680f2882952b306f2fcf1e5f6db66f019..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RTT_histComparison.C
+++ /dev/null
@@ -1,176 +0,0 @@
-//********************************
-// Root macro to read in and compare histograms from a hist and ref hist file.
-// Result is written to standard out which is caught by the Python popen object
-// used to launch the ROOT child process that ran this macro.
-//
-// BS 12/04
-//********************************
-
-//=========================================================================
-
-TObjArray* getHists(TFile * f){
-  //printf("getHists: start\n");
-  TObjArray* hists = new TObjArray(0);
-  int size = hists->GetSize();
-  //printf("getHists: size on creation: %d\n",size);  
-
-  TIter next(f->GetListOfKeys());
-  TKey *key;
-  while ((key=(TKey*)next())) {
-    char* name = key->GetName();
-    //printf("getHists: obj name %s\n", name);
-    TObject* obj = f->Get(name);
-    hists->Add(obj);
-  }  
-  
-  int size = calcSize(hists);
-  //  printf("getHists: from %s found %d hists\n", fname, size);
-  return hists;
-}
-
-//=========================================================================
-
-int calcSize(TObjArray* col){
-  //printf("calcSize: start:");
-  int count = 0;
-  TIter iter = col->MakeIterator();
-  while(iter.Next()){++count;}
-  return count;
-}
-
-//=========================================================================
-
-bool binByBinComparison(hist1,hist2){
-  Int_t nbins1 = hist1->GetNbinsX();
-  Int_t nbins2 = hist2->GetNbinsX();
-  if(nbins1 != nbins2){
-    return false;
-  }
-  else{
-    for(Int_t binNo=0;binNo < nbins1;binNo++){
-      if(hist1->GetBinContent(binNo) != hist2->GetBinContent(binNo)){
-	return false;
-      }
-    }
-  }
-
-  return true;
-}
-
-//=========================================================================
-
-bool statisticalComparison(hist1,hist2){
-  return true;
-  //  return (hist1->KolmogorovTest(hist2) < probValue);
-}
-
-//=========================================================================
-
-bool isSameHistogram(TH1* hist1, TH1* hist2, char * comparisonType){
-  if (comparisonType=='bin-by-bin'){
-    return binByBinComparison(hist1,hist2);
-  }
-  else{ //Kolmogorov test
-    return statisticalComparison(hist1,hist2);
-  }
-
-  return true;
-}
-
-//=========================================================================
-
-int cleanup(TObjArray * array1, TObjArray * array2, TFile * file1, TFile * file2){
-
-  array1->Delete();
-  delete array1;
-
-  array2->Delete();
-  delete array2;
-  
-  delete file1;
-  delete file2;
-
-  return 0;
-}
-
-//=========================================================================
-
-char * toString(TObjArray * histTitles){
-  TIter iter(histTitles);
-
-  TString total;
-  while( currStringObject = (TObjString*)iter.Next()){
-    TString tString = currStringObject->GetString();
-    total += tString->Data() + ", ";
-  }
-
-  return total->Data();
-}
-
-//=========================================================================
-
-int RTT_histComparison(char* newHistFile, char* refHistsFile, char * compType){
-  int successCode = 0;
-  int errorCode = 1;
-
-  //------------------------------------------------------------------
-  // Open the 2 histogram files and extract the histograms into 2 lists
-  //------------------------------------------------------------------
-  TFile* hists = new TFile(newHistFile); 
-  TFile* refHists = new TFile(refHistsFile); 
-    
-  listHists  = getHists(hists);
-  listRhists = getHists(refHists);
-  
-  //------------------------------------------------------------------
-  // Different number of histograms in the two histogram files -> EXIT
-  //------------------------------------------------------------------
-  if(calcSize(listHists) != calcSize(listRhists)){
-    cleanup(listHists,listRhists,hists,refHists);
-    cout << "RTT_histComparison.C return value: false info: the histogram files " << newHistFile << " and " << refHistsFile << " did not contain the same number of histograms" << endl;
-    return errorCode;
-  }
-
-  //------------------------------------------------------------------
-  // No histograms in the newly created histogram file -> EXIT
-  //------------------------------------------------------------------
-  if(calcSize(listHists) == 0 ){
-    cleanup(listHists,listRhists,hists,refHists);
-    cout << "RTT_histComparison.C return value: false info: the newly created histogram file" << newHistFile << " contains no histograms!" << endl;
-    return errorCode;
-  }
-  
-  //------------------------------------------------------------------
-  // Preliminary checks over, now compare each histogram in the two files
-  //------------------------------------------------------------------
-  TIter new_iter(listHists);
-  TIter ref_iter(listRhists);
-
-  TObjArray* failed = new TObjArray(0);
-  int sizeOnInit = failed->GetSize();
-
-  while( newHist = (TH1*)new_iter.Next()){
-    refHist = (TH1*)ref_iter.Next();
-    if (!isSameHistogram(newHist,refHist,compType)){
-      failed->Add(new TObjString(newHist->GetTitle()));
-    }
-  }
-
-  //------------------------------------------------------------------
-  // Some histograms were different -> EXIT
-  //------------------------------------------------------------------
-  if (failed->GetSize() > sizeOnInit){
-    cout << "failed size: " << failed->GetSize() << endl;
-    cout << "RTT_histComparison.C return value: false info: the following histograms were not identical in the two files " << newHistFile << " and " << refHistsFile << ": " << toString(failed) << endl;
-    cleanup(listHists,listRhists,hists,refHists);
-    failed->Delete();
-    delete failed;
-    return errorCode;
-  }
-  
-  //------------------------------------------------------------------
-  // All OK
-  //------------------------------------------------------------------
-  cout << "RTT_histComparison.C return value: true" << endl;
-  return successCode;
-}
diff --git a/Tools/RunTimeTester/share/RandomDozerApp.py b/Tools/RunTimeTester/share/RandomDozerApp.py
deleted file mode 100644
index 018c4ba32ea6b9aa2fac30a35e01fc092ec2e1af..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RandomDozerApp.py
+++ /dev/null
@@ -1,190 +0,0 @@
-#!/bin/env python
-
-##
-# @file RandomDozerApp.py
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief example CoolDozer application
-
-import sys
-sys.argv.append("-b")
-
-
-dozerPath =  "/afs/cern.ch/atlas/project/RTT/Results/dozer/api/test" 
-sys.path = [ dozerPath ] + sys.path
-sys.path.append( dozerPath )
-
-from CoolDozer import *
-import random
-
-##
-# @class RandomDozerApp
-# @author Krzysztof Daniel Cibe (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief example application using CoolDozer API for Atlas RTT test framework 
-class RandomDozerApp( object ):
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param argDict arguments dictionary
-    def __init__( self , argDict = {} ):
-        
-        self.log = DozerLogger( "RandomDozerApp.log", "RandomDozerApp", toConsole=True )
-        self.log = self.log.getLogger( self )
-        self.log.toConsole( True )
-        print self.log.getFileName()
- 
-        self.log.info("*** Hi ***")
-
-        self.dry = bool( argDict.get("DryRun", None ) )
-        if ( not self.dry ): self.dry = True
- 
-        self.log.info("getting RTT JobDescriptior from arguments dictionary")
-        jobDtor = argDict.get("JobDescriptor", None )
-        if ( jobDtor != None ):
-            self.log.info("JobDescriptior has been found")
-
-            self.package = jobDtor.package
-            self.outputPath = jobDtor.runPath
-
-            self.nightly = jobDtor.paths.branch
-            self.install = jobDtor.paths.runType
-            self.cmtconfig = jobDtor.paths.cmtCondig
-            self.project = jobDtor.paths.topProject
-            self.release = jobDtor.paths.release
-            self.jobId = "job_1"
-
-        else:
-            self.package = "DozerEx"
-            self.outputPath = "."
-            self.nightly = "bugfix"
-            self.install = "build"
-            self.cmtconfig = "i686-slc4-gcc34-opt"
-            self.project = "offline"
-            self.release = "rel_0"
-            self.jobId = "job_1"
-
-        self.dozerSchemaConfig = DozerSchemaConfig( self.nightly,
-                                                    self.install,
-                                                    self.cmtconfig,
-                                                    self.project, self.jobId )
-
-        self.rtSchemaPath = self.nightly + "/" + self.install + "/" + self.cmtconfig + "/" + self.project  + "/" + self.jobId
-        self.log.info("will use runtime schema path = %s" % self.rtSchemaPath )
-
-        self.db = DozerDB( self.package, "RTT", self.dry, self.log )
-
-
-    def run( self ):
-        
-        # writting 
-        try:
-            isOpened = self.db.openDB() 
-        except DozerError, value:
-            self.log.epanic( value )
-        
-        self.rtSchema = None
-        if ( isOpened ):
-
-            try:
-                self.rtSchema = self.db.getSchema( self.rtSchemaPath )
-            except DozerError, value:
-                self.log.info("schema hasn't been added to the database yet!")
-
-                            
-            # run schemator and get schema from it
-            if ( not self.rtSchema ):
-                self.schemator()
-                self.rtSchema = self.rConfig.get( self.rtSchemaPath )
-        
-            if ( not self.rtSchema ):
-                self.log.panic("no schema for config = '%s' found in db nor self.schemator()!")
-                sys.exit(0)
-            else:
-                self.db += self.rtSchema  
-            
-            ## write to db
-            self.writer() 
-        
-            ## read data back and make some plots
-            self.reader()
-
-            self.db.closeDB()
-
-        pass
-
-    ## creates dummy DozerConfig for RTT
-    # @param self "Me, myself and Irene"
-    def schemator( self ):
-        
-        self.rConfig = DozerConfig( "RTT",  self.log )
-           
-        rChannel = DozerChannel("dummyRandom", "dummy random channel", log=self.log)
-       
-        rChannel += DozerData("dummy random int", "Int", "dummy unit count", log=self.log )
-        rChannel += DozerData("dummy random float", "Float", "dummy float count", log=self.log )
-
-        rSchema = DozerSchema("*", "*", "*", "*",  "*", log=self.log)
-
-        rSchema += rChannel
-        
-        self.rConfig += rSchema 
-
-
-    def writer( self ):
-        
-        if ( "dummyRandom" in self.rtSchema ):
-            chRandom = self.rtSchema["dummyRandom"]
-            if ( "dummy random int" in chRandom ):
-                chRandom["dummy random int"].setValue( random.randrange(100, 110) )
-            if ( "dummy random float" in chRandom ):
-                chRandom["dummy random float"].setValue( random.gauss( 115.0, 3.0 ) ) 
-        
-        iov = DozerIOV( log=self.log )
-        todayAsRelease = iov.AsRelease()
-        since = iov.today()
-        until = iov.max() 
-
-        
-        if ( "rel_" in self.release ):
-
-            if ( todayAsRelease != self.release ):
-                
-                since = iov.fromRelease( self.release )
-                
-                until = since.nextDay()
-                since = since()
-                
-        
-        if ( self.db and self.db.isOpened() ):
-            self.db.addSchemaPayload( self.rtSchema, since , until )
-
-
-    ## 
-    # @param self "Me, myself and Irene"
-    def reader( self ):
-
-        recordSet = None
-        if ( "dummyRandom" in self.rtSchema ):
-            recordSet = self.db.get( self.db.pathChunk() + "/" + self.rtSchema.pathChunk() + "/" + "dummyRandom" )
-
-            
-        if ( recordSet ):
-
-            graph = recordSet.getDozerGraph() 
-
-            graph.trend( "dummy random int", False, "trend_testRandomInt.png")
-            graph.histogram( "dummy random int", False, "histo_testRandomInt.png")
-           
-
-            graph.trend( "dummy random float", False, "trend_testRandomFloat.png")
-            graph.histogram( "dummy random float", False, "histo_testRandomFloat.png")
-            
-
-
-        
-## dummy test running 
-if __name__ == "__main__":
-      print "dummy running RandomDozerApp..."
-      testCfg = { "JobDescriptor"    : None,
-                  "DryRun"           : True }
-      dozerApp = RandomDozerApp( testCfg )
-      dozerApp.run()
diff --git a/Tools/RunTimeTester/share/RttLibraryTools.py b/Tools/RunTimeTester/share/RttLibraryTools.py
deleted file mode 100755
index 968191d213797432035d1845be636ce94d640331..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RttLibraryTools.py
+++ /dev/null
@@ -1,13 +0,0 @@
-from Logger                        import Logger
-from ROOTMacroRunner               import ROOTMacroRunner
-from importName                    import importName
-from FileComparator                import FileComparator
-from FileGrepper                   import FileGrepper
-from FilesWithStringsFinder        import FilesWithStringsFinder
-from WatcherFilesWithStringsFinder import WatcherFilesWithStringsFinder
-from EchoFile                      import EchoFile
-from FileToDict                    import FileToDict
-from RTTExplosion                  import RTTExplosion
-from RTTRegression                 import RTTRegression
-from JunkTest                      import JunkTest
-
diff --git a/Tools/RunTimeTester/share/RttLibraryToolsImp.py b/Tools/RunTimeTester/share/RttLibraryToolsImp.py
deleted file mode 100755
index 5ae4753268b6336d12294e99415656c49099ca80..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RttLibraryToolsImp.py
+++ /dev/null
@@ -1,933 +0,0 @@
-"""
-Module contains tools[classes] that can be called
-from out of the RTT through configuration files
-
-"""
-
-import os, logging, sys, imp, traceback
-import os.path
-sys.path.append(os.path.join(os.getcwd(), '../src'))
-import fnmatch, string, re
-from popen2 import popen4
-
-from  popen2              import Popen4
-from shellCommand         import shellCommand
-#from siteSpecificsFactory import siteSpecificsFactory
-from Factory_SiteSpecifics import SiteSpecificsFactory
-from SiteSpecifics        import * 
-from DBFileRetriever      import DBFileRetriever
-from ShellCommand         import ShellCommand
-
-import anydbm
-
-# logger=logging.getLogger('rtt')
-class Logger:
-    def __init__(self):
-        self.debugL    = []
-        self.infoL     = []
-        self.warningL  = []
-        self.errorL    = []
-
-    def debug(self, line):    self.debugL.append(line)
-    def info(self, line):     self.infoL.append(line)
-    def warning(self, line):  self.debugL.append(line)
-    def error(self, line):    self.errorL.append(line)
-    def critical(self, line): self.criticalL.append(line)
-
-def importName(moduleName, name):
-    """From Python Cookbook
-       Import a named object from a module
-    """
-    try:
-        module = __import__(moduleName, globals(), locals(), [name])
-    except ImportError:
-        logger.error('ImportName failed to import %s %s' % (moduleName, name))
-        return None
-    return vars(module)[name]
-
-class FileComparator:
-
-    def __init__(self, paramDict):
-
-        self.diffFile        ="differ.out"
-        self.testFile        = paramDict['testFile']
-        self.refFile         = paramDict['refFile']
-        self.diffFile        = paramDict['diffFile']
-        self.refFileLocation = paramDict.get('refFileLocation')
-        
-        jDescriptor          = paramDict.get('jDescriptor')
-        self.shareArea       = jDescriptor.paths.shareArea
-        self.dbFileRetriever = DBFileRetriever(jDescriptor)
-
-        # self.logger   = paramDict['logger']
-        self.logger   = Logger()
-        self.success  =  0
-        self.error    = -1
-
-
-    def run(self):
-        
-        cwd                  =  os.getcwd()
-
-        if self.refFileLocation  == None:
-            self.refFileLocation = 'rttDB'
-            self.logger.warning("No ref file location specified. Will use default rttDB")
-        
-        testF = ""
-        refF  = ""
-        #reference files will be installed or will be in RTT DB
-        try:
-            testFile =""
-            refFile  =""
-            testFile = os.path.join(cwd,self.testFile)
-            refFile  = self.getRefFile(self.refFileLocation)
-
-            testF = open(testFile, "rb").read()
-            refF  = open(refFile, "rb").read()
-            
-        except Exception, e:
-            self.logger.error("Could not read files to compare ")
-            self.logger.error(str(e))
-            msg = 'File Comparator failed : ' +str(e)
-            # jDescriptor.reporter.failure(msg)
-            self.logger.error(msg)
-            return self.error
-        
-
-        if testF !=refF:
-            self.logger.debug("The following two files are different")
-            self.logger.debug("File 1 :"+ str(testFile))
-            self.logger.debug("File 2 :"+ str(refFile))
-            diffCommand = "diff "+testFile +" "+refFile+" >"+self.diffFile
-            runDiff =Popen4(diffCommand)
-            runDiff.wait()
-            
-            msg =  'File Comparator failed. \n'
-            msg += 'The following files are different : \n%s\n%s' % (
-                str(testFile), str(refFile))
-            # jDescriptor.reporter.failure(msg)
-            self.logger.error(msg)
-            return self.error   
-        else:
-            self.logger.info("The following two files are the same ")
-            self.logger.info("File 1: "+str(testFile))
-            self.logger.info("File 2: "+str(refFile))
-            return self.success
-
-    def getRefFile(self, location):
-
-        if location == 'install':
-            refFile = os.path.join(self.shareArea, self.refFile)
-            return refFile
-        elif location == 'rttDB':
-            refFile = self.dbFileRetriever.writeRefFile(self.refFile)
-            self.dbFileRetriever.storeFileInDB(self.testFile)
-            return refFile
-        else:
-            self.logger.error("Ref File location unknown "+str(location))
-            
-        
-
-    def usage(self):
-        self.logger.info("Usage : FileComparator(<dict>)")
-        self.logger.info("success ==> Files same; error ==> Files differ or exception")
-
-
-
-class FileGrepper:
-    
-    def __init__(self, paramDict={}):
-
-        self.success   =  0
-        self.error     = -1
-        # self.logger  = paramDict['logger']
-        self.logger    = Logger()
-        self.iFilePat  = str(paramDict.get('inputFile', None))
-        self.outFile   = paramDict.get('outputFile', None)
-
-
-        # the searchList and and vetoList entries in argDict are in fact
-        # a single string with a comma acting as a separator (!!????)
-        # undo this data structure
-        # slightly tricky: if no entry in the dictionary, init to empty list
-        # otherwise is a (comma seperated) string. Is so split into a list
-        # the user can also pass the arguments as a true list
-        # labeled as searchStrings and vetoStrings.
-        self.searchList = paramDict.get('searchList', [])
-        if self.searchList: self.searchList = self.searchList.split(',')
-        self.searchList.extend(paramDict.get('searchStrings', []))
-
-        self.vetoList = paramDict.get('vetoList', [])
-        if self.vetoList: self.vetoList = self.vetoList.split(',')
-        self.vetoList.extend(paramDict.get('vetoStrings', []))
-
-        
-        jDescriptor = paramDict['JobDescriptor']
-        self.runPath = jDescriptor.runPath
-        
-    def getInFile(self):
-        
-        # if self.argDict.get('useJobLogFile', None):
-        #    return  self.jDescriptor.log
-            
-        if not self.iFilePat:
-            m = "FileGrepper: existing - no input file pattern was given"
-            self.logger.error(m)
-            return None            
-        
-        possFiles = [f for f in os.listdir(self.runPath)
-                     if fnmatch.fnmatch(f, self.iFilePat)]
-        if not possFiles:
-            m = 'FileGrepper  No file matches the  in-file name pattern'
-            m += ' %s' % (self.iFilePat)
-            self.logger.warning(m)
-            return None
-            
-        inFile = possFiles[0]
-        if len(possFiles)>1:
-            m = 'FileGrepper  More than one file matching in file name pattern'
-            m += ' %s:  %s\n using %s' % (self.iFilePat, possFiles, inputFile)
-            self.logger.warning(m)
-
-        return inFile
-    
-
-    def run(self):
-
-
-        inFile = self.getInFile()
-        if not inFile: return self.error
-
-        if not self.outFile:
-            m = "FileGrepper: existing - no out file was given"
-            self.logger.error(m)
-            return self.error
-
-        inFile  = os.path.join(self.runPath, inFile)
-        self.outFile = os.path.join(self.runPath, self.outFile)
-
-
-        
-        print inFile
-
-        lines = open(inFile, 'r').readlines()
-        
-        def lineWanted(line):
-            for pattern in searchList:
-                if re.search(pattern, line):
-                    print 'match: pattern, line',pattern, line
-                    return True
-            return False
-        
-        lines = [l for l in lines if lineWanted(l)]
-
-        def lineVetoed(line):
-            for pattern in vetoList:
-                if re.search(pattern, line):
-                    print 'veto: pattern, line',pattern, line
-                    return True
-            return False
-            
-        lines = [l for l in lines if not lineVetoed(l)]
-
-        outF = open(self.outFile, "w")
-        [outF.write(l) for l in lines]
-        outF.close()
-
-        m  = 'FileGrepper:\n '
-        m += 'input file: %s out file %s\n' % (inFile, self.outFile)
-        m += 'Strings to match: %s\n' % str(searchList)
-        m += 'Veto strings: %s\n' % str(vetoList)
-        m += 'No of lines with a matches: %d' % len(lines)
-        self.logger.debug(m)
-        print m
-        
-        if lines: return self.success
-        return self.error
-    
-
-class FilesWithStringsFinder:
-    
-    def __init__(self, paramDict):
-
-        self.logger            = Logger()
-        self.inputFilePatterns = paramDict.get('inputFilePatterns', [])
-        self.searchPatterns    = paramDict.get('searchPatterns',    [])
-        self.dirWithFiles      = paramDict.get('dirWithFiles', 'resultsDir')
-        self.writeTxt          = paramDict.get('writeTxt', 'yes')
-        self.writeHTML         = paramDict.get('writeHTML','yes')
-        jDescriptor            = paramDict.get('JobDescriptor', None)
-        self.runPath           = jDescriptor.runPath
-        outputFile             = paramDict.get('outputFile','filesWithStrings')
-        self.outputFile        = os.path.join(self.runPath,outputFile)
-        self.resultsDirs       = jDescriptor.paths.resultsDirs
-        self.writers           =  []
-
-
-        dict = {
-            'runPath':           self.runPath,
-            'outputFile':        self.outputFile,
-            'inputFilePatterns': self.inputFilePatterns,
-            'searchPatterns':    self.searchPatterns,
-            'logger':            self.logger
-            }
-
-        class Writer:
-            def __init__(self, dict):
-                self.ofile  = None
-                self.outfn = os.path.join( dict['runPath'],
-                                           dict['outputFile'])
-
-                
-                fpats = dict['inputFilePatterns'] 
-                spats = dict['searchPatterns']
-
-                header  = ['Files to match:']
-                header.extend([('  %s ' % i ) for i in fpats])
-                header.append(' ')
-                header.append('\n\nStrings to match:')
-                header.extend([('  %s ' % s ) for s in spats])
-                header.append(' ')
-                self.header = [self.terminateLine(line) for line in header]
-                self.logger   = dict['logger']
-
-            
-            def openFile(self):
-                """File is opened when there is a need to write (is not
-                possible to open the file at contruction). On opening
-                write out (once!) the header. This part of the action writing
-                an item is common to all derived sclasses - they do the rest.
-                """
-
-                if self.ofile: return
-
-                self.ofile = open(self.outfn, 'w')
-                if not self.ofile:
-                    msg = 'FilesEithStringsFinder: could not open file %s' % self.outfn
-                    self.logger.error(msg)
-                    return
-                    
-                [self.ofile.write(line) for line in self.header]
-
-        class HTMLWriter(Writer):
-            def __init__(self, dict):
-                Writer.__init__(self, dict)
-                self.outfn  += '.html'
-                self.logger   = dict['logger']
-
-        
-                htmlHeader = []
-                htmlHeader.append('<html>\n')
-                htmlHeader.append('<head>\n')
-                htmlHeader.append('   <title>FilesWithStringsFinder</title>\n')
-                htmlHeader.append('</head>\n')
-                htmlHeader.append('<body>\n')
-                htmlHeader.extend(self.header)
-                htmlHeader.append('<table>\n')
-                self.header = htmlHeader
-
-            def terminateLine(self, line):
-                line+='<br>\n'
-                return line
-
-            def outputItem(self, item):
-
-                if not self.ofile: return
-                
-                fn  = os.path.basename(item[0])
-                href ='<a href="%s">  %s </a>' % (fn, fn)
-                htext =  'contains "%s"'% (item[1])
-
-
-
-                line  = '<tr>\n'
-                line += '<td>'+href+'</td><td>'+htext+'</td>\n'
-                line += '</tr>\n'
-                
-                self.ofile.write(line)
-
-                
-            def closeFile(self):
-                if not self.ofile: return
-
-                self.ofile.write('</table>\n')
-                self.ofile.write('</body>\n')
-                self.ofile.write('</html>\n')
-                self.ofile.close()
-            
-        class TxtWriter(Writer):
-            def __init__(self, dict):
-                Writer.__init__(self, dict)
-                self.outfn  += '.txt'
-                self.logger   = dict['logger']
-                    
-            def terminateLine(self, line):
-                line+='\n'
-                return line
-
-
-            def outputItem(self, item):
-                # output to the html file needs to be htmlified
-                # delay opening the files until now, otherwise there is
-                # the danger the output directories have not been constructed.
-
-                if not self.ofile: return
-
-                width = 60
-                fn = os.path.basename(item[0])
-                line = 'file: %s contains "%s" \n' % (string.ljust(fn, width),
-                                                      item[1]
-                                                      )
-                self.ofile.write(line)
-
-            def closeFile(self):
-                if not self.ofile: return
-                self.ofile.close()
-            
-        if self.writeTxt == 'yes':
-            self.writers.append(TxtWriter(dict))
-        if self.writeHTML == 'yes':
-            self.writers.append(HTMLWriter(dict))
-
-        self.logger.debug('FilesWithStringsFinder Writers: %s' % (
-            str(self.writers)))
-    def run(self):
-
-        files = self.filesInDirs()
-        self.logger.debug('FilesWithStringsFinder: filesInDirs %s' % str(files))
-
-        fPatterns     = self.inputFilePatterns
-        filesToSearch = []
-        
-        [filesToSearch.extend(fnmatch.filter(files,
-                                             ipfp)) for ipfp in fPatterns]
-
-        strs         = self.searchPatterns
-
-        self.logger.debug('FilesWithStringsFinder: files to search %s' % str(filesToSearch))
-        self.logger.debug('FilesWithStringsFinder: searchPatterns %s' % str(strs))
-        print 'searchpatterns',strs
-
-        class StringInFile:
-            def __init__(self, base, file):
-                ifn = os.path.join(base,file)
-                print 'checking file ',ifn,'... '
-                self.contents = open(ifn, 'r').read()
-            def check(self, string):
-                print '    ... for string', string
-                if self.contents.find(string) > 0: return 0
-                return -1
-            
-        filesWithString = []
-        for f in filesToSearch:
-            sFile = StringInFile(dwf, f)
-            [filesWithString.append((f,s))  for s in strs if sFile.check(s)]
-
-        self.logger.debug('filesWithString %s' % str(filesWithString))
-
-        [writer.openFile() for writer in self.writers]
-
-        for writer in self.writers:
-            [writer.outputItem(item) for item in filesWithString]
-
-        [writer.closeFile() for writer in self.writers]
-
-        return 0
-
-    def filesInDirs(self):
-        """
-        return a list of files found using the descriptor Paths object scan
-        the descriptor is passed as an argument, rather than getting
-        it from the dictionary asthe descriptor passed may not be the local
-        on: watchers pass the descriptors of their confederation minders, not
-        their own descriptors.
-        """
-
-        msgHead ='FilesWithStringsFinder.filesInDirs: '
-
-        logicalDir = self.dirWithFiles
-        files = []
-        
-        self.logger.debug('%s files to search %s' % (msgHead, str(logicalDir)))
-
-        if logicalDir == 'resultsDir':
-            key = 'resBase/build/branch/runType/release/jobGroup'
-            if key not in self.resultsDirs:
-                self.logger.error('results directory not in Paths %s' % key)
-                return files
-            dwf = self.resultsDirs[key]
-            self.logger.debug('FilesWithStringsFinder violfile directory is %s' % dwf)
-            files = os.listdir(dwf)
-            return files
-        elif logicalDir == 'runDir':
-            files = os.listdir(self.runPath)
-            self.logger.debug('FilesWithStringsFinder violfile directory is %s' % self.runPath)
-            return files
-        else:
-            msg = 'FilesWithStringFinder: Unknown directory to scan %s' % dwf
-            self.logger.error(msg)
-            return files
-
-class WatcherFilesWithStringsFinder(FilesWithStringsFinder):
-    def __init__(self, dict):
-        FilesWithStringsFinder.__init__(self, dict)
-        jDescriptor        = dict['JobDescriptor']
-        self.confederation = jDescriptor.confederation
-        
-    def filesInDirs(self, jDescriptor):
-        """return a list of files found using the minders of the Watcher's
-        confederation of Worker jobs"""
-
-        files = []
-        for minder in self.confederation:
-            files.extend(FilesWithStringsFinder.filesInDirs())
-
-        return files
-
-# ============ROOTMacroRunner===================
-     
-class ROOTMacroRunner:
-    def __init__(self, argDict):
-        
-        self.macroArg     = argDict.get('macroArg', None)
-        self.macro        = argDict.get('macro', None)
-        # self.logger     = argDict['logger']
-        self.logger       = Logger()
-        
-        jDescriptor       = argDict['JobDescriptor']
-        self.runPath      = jDescriptor.runPath
-        self.runType      = jDescriptor.paths.runType
-        self.distBase     = jDescriptor.paths.hardCoded.distBase
-        self.distArea     = jDescriptor.paths.distArea
-        self.branch       = jDescriptor.paths.branch
-        self.release      = jDescriptor.paths.release
-        self.workDirs     = jDescriptor.paths.workDirs
-        self.nightlyBranchFragments = jDescriptor.paths.hardCoded.nightlyPathFragments
-        
-        self.fullPathToMacro = os.path.join(self.jDescriptor.runPath,
-                                            self.macro)
-
-
-        try:
-            compileTheMacro = argDict['compileTheMacro']
-        except:
-            compileTheMacro = False
-
-
-        self.compileFlag = ''
-        if compileTheMacro: self.compileFlag = '+'
-        
-        self.maxPrintLines = 500
-        self.timeLimit     = 60*60 # secs
-        self.rootCmd       = self.makeRootCommand()
-        msg  = 'root command for job '
-        msg += str(jDescriptor.identifiedName)
-        msg += ':\n' + str(self.rootCmd)
-        self.logger.debug(msg)
-
-    def macroExists(self):
-        if not os.path.exists(self.fullPathToMacro):
-            msg  = 'Path to ROOT macro does not exist: %s' % self.fullPathToMacro
-            self.logger.error(msg)
-            return False
-        return True    
-        
-    def run(self):
-        # check the macro exists. NOTE: this cannot be done in __init__ because
-        # class instantiation occurs before copying of auxFilePatterns.
-        if not self.macroExists():
-            self.logger.warning("Root macro does not exit")
-            return 1
-        
-
-        # note the current directory to reset it later
-        currCwd = os.getcwd()
-
-        cmds = []
-        # following cd must be part of the shell command has so other threads
-        # cant mess it up
-        cmds.append('cd ' + self.runPath) 
-        cmds.extend(self.rootCmd)
-        self.logger.info("Launching macro "+str(cmds))
-        sc            = ShellCommand(cmds, self.maxPrintLines, self.timeLimit)
-        reply         = sc.getReply()
-        self.logger.debug(str(reply))
-        # reset the current directory as it was before
-        os.chdir(currCwd)
-
-        # reply is a list of strings, let's turn it into a single string
-        replyString = reduce(lambda x,y: x+y+'\n', reply,'')
-
-        # output reply to the logger
-        name = str(self.jDescriptor.identifiedName)
-        msg  = 'root command output for job %s\n %s' % (name, replyString)
-        self.logger.info(msg)
-
-        # Also output reply to file 'ROOToutput.log' in the run path
-        # This file is a keep file pattern in the TopLevelJobGroup.xml file.
-        h = open(os.path.join(self.runPath,'ROOToutput.log'),'a+')
-        h.write(replyString)
-        h.close()
-
-        return 0
-
-    def ROOTenvSetupAndLaunch(self):        
-        info = {}
-        sendToDevNull = ' >& /dev/null'
-
-        if self.runType=='kit':
-            aDir = str(self.distBase)
-            pathToLocalSetupScript =  os.path.join(str(self.distArea),
-                                                   'cmtsite/setup.sh')             
-        else:
-            aDir = os.path.join(self.distBase,
-                                self.nightlyPathFragment[self.branch])
-                        
-            pathToLocalSetupScript = os.path.join(self.workDirs['workBase/'],
-                                                  'cmtHome/setup.sh')
-                                    
-
-        localScriptTags = ' -tag=opt,AtlasOffline,' + str(self.release)
-        atlasOfflineRunTimeDir = os.path.join(aDir,'AtlasOffline',
-                                              str(self.release),
-                                              'AtlasOfflineRunTime/')        
-        pathToAtlasOfflineRunTimeSetupScript = os.path.join(
-            atlasOfflineRunTimeDir,'cmt/setup.sh')
-        
-        info['localSetupScriptCmd']   = 'source ' + pathToLocalSetupScript + localScriptTags + sendToDevNull
-        info['releaseSetupScriptCmd'] = 'source ' + pathToAtlasOfflineRunTimeSetupScript + sendToDevNull
-        info['launchRootCmd']         = 'root -b -q '
-        return info
-            
-    def makeRootCommand(self):
-        # returns a list of shell commands, untermintated by ';' to be added
-        # to a list of commands to be sent to shellCmd
-
-        cmds = self.ROOTenvSetupAndLaunch()
-
-        cmdLines = []
-        if self.macro:            
-            cmdLines.append(cmds['localSetupScriptCmd'])
-            cmdLines.append(cmds['releaseSetupScriptCmd'])
-            line = cmds['launchRootCmd'] + self.macro
-            if self.macroArg:
-                line += self.compileFlag+'('+self.macroArg+')'
-            line = self.escapeTheCommand(line)
-            cmdLines.append(line)
-
-        
-        return cmdLines
-    
-    def escapeTheCommand(self, cmd):
-
-        self.toBeEscaped = ['(',')','"']
-        
-        for char in self.toBeEscaped:
-            pieces = cmd.split(char)
-            
-            cmd = pieces[0]
-            for piece in pieces[1:]:
-                cmd += '\\'+char+piece
-
-        return cmd
-
-
-# ============EchoFile===================
-
-class EchoFile:
-    def __init__(self, argDict):
-
-        self.inFile   = argDict.get('datafile')
-        # self.logger = params['logger']
-        self.logger   = Logger()
-
-    def run(self):
-        inF  = open(self.inFile, 'r').readlines()
-        outF = open('echoedFile.txt', 'w')
-
-        for line in inF:
-            outF.write(line)
-
-        outF.close()
-        
-
-class FileToDict:
-    def __init__(self, argDict):
-
-        self.inFile   = argDict.get('datafile')
-        self.logger   = Logger()
-        # self.logger = params['logger']
-
-    def run(self):
-        self.DataDict      = self.makeDict(self.inFile)
-
-        return self.DataDict
-        
-
-    def retrieveFile(self, key, db):
-        pass
-    
-    
-    def linesCleanUp(self, lines):
-        ls =[]
-        for line in lines:
-            line =string.strip(line)
-            if len(line)>0:
-                ls.append(line)
-
-        return ls
-
-
-    def makeDict(self,file):
-        lines         = open(file, 'r').readlines()
-        lines         = self.linesCleanUp(lines)
-        
-        self.dict={}
-        for item in lines:
-            keyValuePairs     = string.split(item, ":")
-            for m in  keyValuePairs:
-                keyValuePairs = self.linesCleanUp(keyValuePairs)
-            self.dict[keyValuePairs[0]] = int( keyValuePairs[1])
-            
-        return self.dict
-
-
-
-
-# -------------------------------------------
-
-class RTTExplosion:
-    """An RTT test to run its own TestSuite under RTT control
-    This test checks that classes can be instantitiated. It also
-    creates the databass needed to run RTTRegression"""
-    
-
-    def __init__(self, argDict):
-
-
-        self.success =  0
-        self.error   = -1
-
-        # self.logger = self.argDict['logger']
-        self.logger = Logger()
-        msg = 'Instantiating RTTExplosion, args: %s' %str(argDict)
-        self.logger.debug(msg)
-        jDescriptor = argDict['JobDescriptor']
-        # fixtureDir is set in JobsXMLReader when reading in the config file.
-        self.fixtureDir = argDict['fixtureDir']
-        # the current directory
-        self.runPath   =  jDescriptor.runPath
-        self.logDir    =  jDescriptor.runPath
-        # directory of the source code under test
-        self.rttSrcDir =  os.path.join(self.runPath, 'Tools/RunTimeTester/src')
-        fixtureDir    =  os.path.basename(self.fixtureDir)
-
-    def run(self):
-
-        # prepend the source directory undertest (must be prepended,
-        # otherwise the RTT source used to run the job will be used
-        # for importing!
-        list = [self.rttSrcDir]
-        list.extend(sys.path)
-        sys.path = list
-        self.logger.debug('sys.path = %s' % str(sys.path))
-        os.chdir(self.rttSrcDir)
-        # sys.path.append(self.rttSrcDir)
-
-        try:
-            from TestSuiteRun2 import TestSuiteRun
-            self.logger.debug('RTTExplosion: Import of TestSuiteRun succeeded')
-        except Exception, e:
-            self.logger.error('RTTExplosion: Import of TestSuiteRun failed')
-            list    = sys.exc_info()
-            errmsg  = str(list[0])
-            errmsg += str(list[1])
-            errmsg += string.join(traceback.format_tb(list[2]))
-            self.logger.error(errmsg)
-            return self.error
-
-        try:
-            tsr = TestSuiteRun(self.rttSrcDir, self.fixtureDir, self.logDir)
-            self.logger.debug('RTTExplosion: TestSuiteRun execution complete')
-        except Exception, e:
-            self.logger.error('RTTExplosion: TestSuiteRun execution failed')
-            return self.error
-
-
-        if tsr.passed:
-            self.logger.debug('RTTExplosion Test passed')
-            return self.success
-        self.logger.debug('RTTExplosion Test failed')
-        return self.error
-        
-        
-# -------------------------------------------
-class RTTRegression:
-    """An RTT test to run its own TestSuite under RTT control
-    This test checks that classes can be instantitiated. It also
-    creates the databass needed to run RTTRegression"""
-    
-
-    def __init__(self, argDict):
-
-
-        self.success =  0
-        self.error   = -1
-
-        # self.logger = argDict['logger']
-        self.logger  = Logger()
-        msg = 'Instantiating RTTRegression, args: %s' %str(argDict)
-        self.logger.debug(msg)
-        # fixtureDir is set in JobsXMLReader when reading in the config file.
-        self.fixtureDir = argDict['fixtureDir']
-        # the current directory
-        jDescriptor    = argDict['JobDescriptor']
-        self.runPath   =  jDescriptor.runPath
-        self.logDir    =  jDescriptor.runPath
-        # directory of the source code under test
-        self.rttSrcDir =  os.path.join(self.runPath, 'Tools/RunTimeTester/src')
-        self.runPath   =  jDescriptor.runPath
-        fixture        =  os.path.basename(self.fixtureDir)
-        self.dbName    =  os.path.join(self.runPath, fixture+'.db')
-        self.refdbName =  os.path.join(self.runPath, 'refFile_'+fixture+'.db')
-        # do not open the dir does not exist yet
-        self.ofName    =  os.path.join(self.runPath, fixture+'_regression.log')
-
-        
-    def run(self):
-
-        outFile = open(self.ofName, 'w')
-
-        if not os.path.exists(self.dbName):
-            msg = 'None existant path: %s' %  self.dbName
-            self.logger.error(msg)
-            outFile.write(msg+'\n')
-            outFile.close()
-            
-            return self.error
-        if not os.path.exists(self.refdbName):
-            msg = 'None existant path: %s' %  self.refdbName
-            self.logger.error(msg)
-            outFile.write(msg+'\n')
-            outFile.close()
-            
-            return self.error
-
-        newDB = anydbm.open(self.dbName,    'r')
-        oldDB = anydbm.open(self.refdbName, 'r')
-        result = self.success
-            
-    
-        onlyInNew = [k for k in newDB.keys() if k not in oldDB.keys()]
-
-        text = 'Number of keys in reference db %d\n' % len(oldDB.keys())
-        text = 'Number of keys in new       db %d\n' % len(newDB.keys())
-        
-        
-        if onlyInNew:
-
-            result = self.error
-    
-            text += '\n'
-            text +='Reference - %s: date: %s\n' % (oldDB['fixtureDir'],
-                                                   oldDB['date'])
-            
-            text +=  'New       - %s: date: %s\n' % (newDB['fixtureDir'],
-                                                     newDB['date'])
-            
-            text += '\n'
-            text +=  ' keys in new database, but not in old\n'
-            text +=  str(onlyInNew)+'\n'
-            text += '\n'
-    
-        onlyInOld = [k for k in oldDB.keys() if k not in newDB.keys()]
-        if onlyInOld:
-
-            result = self.error
-        
-            text +=  '\n'
-            text +=  ' keys in old database, but not in new\n'
-            text +=  str(onlyInOld)+'\n'
-            text += '\n'
-    
-        keys = [k for k in oldDB.keys() if k in newDB.keys()]
-        toRemove = ['fixtureDir', 'date']
-        [keys.remove(k) for k in toRemove if k in keys]
-        if keys:
-            text +=  'differences:\n'
-            text += '\n'
-
-            for k in keys:
-                
-        
-                if oldDB[k] != newDB[k]:
-
-                    result = self.error
-                    
-                    text +=  'Key: %s\n' % k
-                    text += '\n'
-                    text +=  '   old:\n'
-                    text +=  '   ' +str(oldDB[k])+'\n'
-                    text += '\n'
-                    text +=  '   new:\n'
-                    text +=  '   '+str(newDB[k])+'\n'
-
-
-        text += '\n'
-
-        totTests = 0
-        text +=  'Number of points examined:\n'
-        for k in keys:
-            line = ''
-            line += k.ljust(30)
-            ntestOld = len(oldDB[k].split(','))
-            ntestNew = len(newDB[k].split(','))
-            # assert(ntestOld == ntestNew)
-            num = '%d' % ntestOld
-            line += num.ljust(5)
-            # print line
-            totTests += ntestOld
-
-        text +=  'No of test classes which pass: %d\n' % len(keys)
-        text +=  'Total number of tests passed:  %d\n ' %totTests
-
-        outFile.write(text)
-        outFile.close()
-        return result
-        
-        
-
-# ============JunkTest===================
-
-class JunkTest:
-    def __init__(self, argDict):
-        self.int = 1
-        self.logger   = Logger()
-
-    def run(self):
-        # self.logger.error('JunkTest RTT test run method')
-        print 'JunkTest RTT test run method'
-        return 0
-        
-
-# -------------------------------------------
-        
-if __name__ == '__main__':
-    open('junk.dat','w').write('aaaa\naaabb\naabbcccc\naabbddd\nzz\nyyzzzz\n')
-
-    
-    dict = {
-        'inputFile': 'junk*',
-        'outputFile': 'FileGrepper.log',
-        # 'searchList': 'aa,bb',
-        'vetoList': 'c,d',
-        'searchStrings': ['zz', 'yyzzzz'],
-        'vetoStrings':    ['y'],
-        'logger': logger,
-        'JobDescriptor': []
-        }
-
-    fg = FileGrepper(dict)
-    print fg.run()
-    
-    
diff --git a/Tools/RunTimeTester/share/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/share/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 689b3a502cca47d3ebb70a91c82f7ea5ab90639c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,153 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dt">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <rttContactPerson>RTT team (rtt AT hep (dot) ucl.ac dot uk)</rttContactPerson>
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-
-          <rttselftest/>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         
-       <jobGroup name="RTTSelfTest" parent="Top">
-         <keepFilePattern>*.txt</keepFilePattern>
-         <keepFilePattern>Fixtures*</keepFilePattern>
-         <binRefFile>testHist.root</binRefFile>
-
-         <auxFilePattern>testHist.cxx</auxFilePattern>
-
-         <test>
-            <modulename>RttLibraryTools</modulename>
-            <testname>ROOTMacroRunner</testname>
-            <arg>
-              <argname>macro</argname>
-              <argvalue>testHist.cxx</argvalue>
-            </arg>
-         </test>
-
-         <test>
-            <modulename>RttLibraryTools</modulename>
-            <testname>RTTExplosion0</testname>
-            <arg>
-              <argname>fixtureDir</argname>
-              <argvalue>TestSuite0/Fixtures0</argvalue>
-            </arg>
-         </test>
-         <test>
-            <modulename>RttLibraryTools</modulename>
-            <testname>RTTExplosion0</testname>
-            <arg>
-              <argname>fixtureDir</argname>
-              <argvalue>TestSuite0/Fixtures1</argvalue>
-            </arg>
-         </test>
-         <test>
-            <modulename>RttLibraryTools</modulename>
-            <testname>RTTExplosion0</testname>
-            <arg>
-              <argname>fixtureDir</argname>
-              <argvalue>TestSuite0/Fixtures2</argvalue>
-            </arg>
-         </test>
-         <test>
-            <modulename>RttLibraryTools</modulename>
-            <testname>RTTExplosion0</testname>
-            <arg>
-              <argname>fixtureDir</argname>
-              <argvalue>TestSuite0/Fixtures3</argvalue>
-            </arg>
-         </test>
-         <logChecker class="SimpleLogChecker">
-            <searchString>Creating cleanup scripts</searchString>
-         </logChecker>
-      </jobGroup>
-
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <test>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </test>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <test>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>RuleCheckerWatcherSummary</testname>
-            </test>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/share/UserScriptRunnerTemplate.py b/Tools/RunTimeTester/share/UserScriptRunnerTemplate.py
deleted file mode 100644
index 784d9efcbb87e61c82d79f5da68df8df5d8d7cd8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/UserScriptRunnerTemplate.py
+++ /dev/null
@@ -1,18 +0,0 @@
-import os
-import os.path
-import sys
-import cPickle
-
-class UserScriptRunner:
-    def run(self, picklefile):
-        sys.path.append("%USER_MODULE_LOCATION%")
-        sys.path.append("%RTT_SHARE_DIR%")
-        sys.path.append("%RTT_SRC_DIR%")
-
-        # sys.path = ["/afs/cern.ch/atlas/project/RTT/Results/dozer/api"] + sys.path
-        import USER_MODULE_NAME
-        inst = USER_MODULE_NAME.USER_MODULE_NAME(cPickle.load(file(picklefile)))
-        inst.run()
-
-if __name__ == '__main__':
-    UserScriptRunner().run(sys.argv[1].strip())
diff --git a/Tools/RunTimeTester/share/WatcherFilesWithStringsFinder.py b/Tools/RunTimeTester/share/WatcherFilesWithStringsFinder.py
deleted file mode 100755
index e25706820c24a05598e3f3b784f515da92faa0b3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/WatcherFilesWithStringsFinder.py
+++ /dev/null
@@ -1,29 +0,0 @@
-######################################################
-#                                                    #
-# Do NOT bind any objects (self.xxx) which contain   #
-# file objects (such as self.logger in this class    #
-# otherwise cannit shelve the objects                #
-# Instead, unload the necessary variables in         #
-# __init__                                           #
-#                                                    #
-######################################################
-
-
-from FilesWithStringsFinder import FilesWithStringsFinder
-
-class WatcherFilesWithStringsFinder(FilesWithStringsFinder):
-    def __init__(self, dict):
-        FilesWithStringsFinder.__init__(self, dict)
-        jDescriptor        = dict['JobDescriptor']
-        self.confederation = jDescriptor.confederation
-        
-    def filesInDirs(self, jDescriptor):
-        """return a list of files found using the minders of the Watcher's
-        confederation of Worker jobs"""
-
-        files = []
-        for minder in self.confederation:
-            files.extend(FilesWithStringsFinder.filesInDirs())
-
-        return files
-
diff --git a/Tools/RunTimeTester/share/aclicTweak.C b/Tools/RunTimeTester/share/aclicTweak.C
deleted file mode 100755
index 7f244b7752281c8acd0afdefef3bad73934facce..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/aclicTweak.C
+++ /dev/null
@@ -1,90 +0,0 @@
-/**
- * @file aclicTweak.C 
- * @author Krzysztof Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
- * @brief Changing flags for aclic compiler inside root to fit 32bit 
- *        binaries running in 64bit OS.
- *
- * Usage: 
- *  1. from command line:
- *    $bash> root aclicTweak.C myMacro.C++ 
- *
- *  2. inside root
- *   a)  root [1] .L aclicTweak.C
- *       root [2] aclicTweak()
- *       root [3] myMacro.C++ 
- *
- *   b)  root [1] .L aclicTweak.C
- *       root [2] aclicRun("myMacro.C")
- */
-
-#include <string>
-
-
-/**
- * aclic tweaker
- */
-void aclicTweak() {
-
-	
-	std::string::size_type pos = 
-		std::string(gSystem->GetMakeExe()).find("-m32", 0);
-
-	// for 32bit binary sizeof(long int) == 4
-	// for 64bit binary sizeof(long int) == 8
-
-	if ( sizeof(long int) == 4 && 
-			 pos == std::string::npos ) {
-		
-		std::string makeExe(gSystem->GetMakeExe());
-		std::string makeSharedLib(gSystem->GetMakeSharedLib());
-		
-		
-		// only for testing 
-		//std::cout << "***** beforeTweak *****" << std::endl;
-		//std::cout << "EXE   = " << makeExe << std::endl;
-		//std::cout << "SHLIB = " << makeSharedLib << std::endl;
-		
-		std::string::size_type gpp = makeExe.find("g++", 0);
-		if ( gpp != std::string::npos ) {
-			makeExe.insert(gpp+4, " -m32 ");
-		}
-
-		gpp = makeExe.find("g++", gpp+1);
-		if ( gpp != std::string::npos ) {
-			makeExe.insert(gpp+4, " -m32 ");
-		}
-
-		 
-		gpp = makeSharedLib.find("g++", 0);
-		if ( gpp != std::string::npos ) {
-			makeSharedLib.insert(gpp+4, " -m32 ");
-		}
-	 
-		gpp = makeSharedLib.find("g++", gpp+1);
-		if ( gpp != std::string::npos ) {
-			makeSharedLib.insert(gpp+4, " -m32 ");
-		}
-
-		// only for testing 
-		//std::cout << "***** afterTweak *****" << std::endl;
-		//std::cout << "EXE   = " << makeExe << std::endl;
-		//std::cout << "SHLIB = " << makeSharedLib << std::endl;
-		
-		gSystem->SetMakeExe( makeExe.c_str() );
-		gSystem->SetMakeSharedLib( makeSharedLib.c_str() );
-		
-	}
-}
-
-
-/**
- * @fn aclicRun
- * @param script name of root script to compile and run
- */
-void aclicRun(const char* script) {
-
-	aclicTweak();
-	const std:string command( ".x " + std::string(script) + "++");
-	gROOT->ProcessLine(command.c_str());
-
-} 
diff --git a/Tools/RunTimeTester/share/histAdder.C b/Tools/RunTimeTester/share/histAdder.C
deleted file mode 100755
index 9eeac386d71b4ee0d3a2776bf6e6aba6f58d8f80..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/histAdder.C
+++ /dev/null
@@ -1,198 +0,0 @@
-//********************************
-// Root macro to read in and sum histograms from a number of files,
-// and to result the sum back out
-//
-// Needs more work to ensure character buffers do not overflow.
-// PS 10/04
-//********************************
-
-
-TObjArray* getHists(char* fname){
-  //printf("getHists: start\n");
-  TObjArray* hists = new TObjArray(0);
-  int size = hists->GetSize();
-  //printf("getHists: size on creation: %d\n",size);
-  
-  TFile* f = new TFile(fname);
-
-  TIter next(f->GetListOfKeys());
-  TKey *key;
-  while ((key=(TKey*)next())) {
-    char* name = key->GetName();
-    //printf("getHists: obj name %s\n", name);
-    TObject* obj = f->Get(name);
-    hists->Add(obj);
-  }  
-  
-  int size = calcSize(hists);
-  printf("getHists: from %s found %d hists\n", fname, size);
-  return hists;
-}
-
-int calcSize(TObjArray* col){
-  //printf("calcSize: start:");
-  int count = 0;
-  TIter iter = col->MakeIterator();
-  while(iter.Next()){++count;}
-  return count;
-}
-bool addHists(char* fname, TObjArray* hsums, char* rfname, TObjArray* rsums){
-  TObjArray* hists  = getHists(fname);
-  TObjArray* rhists = getHists(rfname);
-
-  int hsize = calcSize(hists);
-  int ssize = calcSize(hsums);
-  int rsize = calcSize(rhists);
-
-  if(hsize != ssize || hsize != rsize){
-    printf("unequal numbers of hists: new h, new r, sum: %d %d %d",
-	   hsize, rsize, ssize);
-    printf("skipping %s\n", fname);
-    return 0;
-  }
-
-
-  accumulateHists(hists, hsums);
-  accumulateHists(rhists, rsums);
-
-  return 1;
-}
-
-void accumulateHists(TObjArray* hists, TObjArray* hsums){
-  TIter h_iter(hists);
-  TIter s_iter(hsums);
-  TH1* h;
-  while( h = (TH1*)h_iter.Next()){
-    s = (TH1*)s_iter.Next();
-    s->Add(h);
-  }
-}
-
-void writeHists(char* sumfile, TObjArray* hists){
-  cout<<"writing to..."<<sumfile<<endl;
-  //printf("number of hists to write: %d\n",calcSize(hists));
-  ofile = new TFile(sumfile, "recreate");
-  //printf("number of hists to write: %d\n",calcSize(hists));
-  hists->Write();
-  ofile->Close();
-}
-
-int nameWithoutCC(char* name, int n, FILE* fp){
-  //printf("nameWithoutCC: start \n");
-  char temp[201];    //buffer size now doubled 10-03-05.
-  char* s;
-
-  if (s = fgets(&temp, n, fp)){
-    //printf("before shortening %s", &temp);
-    //////////////////////////////////
-    // Mac only!!!!
-    int nchar = strlen(s)-1;
-    s = strncpy(name, &temp, nchar);
-    //////////////////////////////////
-    //////////////////////////////////
-    // Linux
-    // s = strcpy(name, &temp);
-    //////////////////////////////////
-    //printf("%d\n", nchar);
-    //printf("\n");
-    // printf("nameWithoutCC: end1: %s\n", s);
-    return 1;
-  }
-  // printf("nameWithoutCC: end2\n");
-
-  return 0;
-}
-bool fillInFileName(char* lp, char* fn){
-  int csize = strlen(lp)+strlen(fn)+strlen("/");
-  if (csize>200){
-    printf("lp character buffer too short  need %d",csize);
-    return 0;
-  }
-  strcat(lp,"/");
-  strcat(lp,fn);
-  return 1;
-}
-
-bool fillInRefFileName(char* lp, char* fn){
-  int csize = strlen(lp)+strlen(fn)+strlen("/refFile_");
-  if (csize>200){
-    printf("lp character buffer too short  need %d",csize);
-    return 0;
-  }
-  strcat(lp,"/refFile_");
-  strcat(lp,fn);
-  return 1;
-}
-
-
-int histAdder(char* fname){
-
-  //printf("histAddr: start\n");
-  //printf("histAddr: %s\n", fname);
-    
-  FILE *fp = fopen("histLocation.txt","r");     
-
-
-  TObjArray* hists  = new TObjArray(0);
-  TObjArray* rhists = new TObjArray(0);
-  int fpass = 1;
-
-  char location[201];
-  char* lp = &location;
-  char rlocation[201];
-  char* rlp = &rlocation;
-
-  while (nameWithoutCC(lp, 200, fp)) {
-
-    if (strlen(lp) >200){printf("Beware: string buffers are full!");}
-
-    strcpy(rlp, lp);
-
-    if (!(fillInFileName(lp, fname) && fillInRefFileName(rlp, fname))){
-      printf("problem with name lengths, skipping %s\n", lp);
-      continue;
-    }
-    printf("histAdder hist name %s\n",lp);
-    printf("histAdder ref hist name %s\n",rlp);
-    
-    if(fpass){
-      fpass =0;
-
-      hists  = getHists(lp);
-      rhists = getHists(rlp);
-
-      if(calcSize(hists) != calcSize(rhists)){
-	printf("no of hists != in ref & new hist files, skipping %s\n",lp);
-	fpass = 1;
-      }
-
-      if(calcSize(hists) == 0 ){
-	printf("no of hists in files = 0, skipping %s\n",lp);
-	fpass = 1;
-      }
-      
-      //printf("histAdder: fpass %d\n",fpass);
-    }else{
-      if(!addHists(lp, hists, rlp, rhists)){
-	printf("Problem accumulating hists %s\n",lp);
-      }
-    }
-    for(int i=0; i<201; ++i){location[i]  = '\0';}
-    for(int i=0; i<201; ++i){rlocation[i] = '\0';}
-  }
-  fclose(fp);
-
-  writeHists(fname, hists);
-
-  char refName[100]="refFile_";
-  char* rfname = &refName;
-  strcat(rfname, fname);
-
-  writeHists(rfname, rhists);
-
-  //int size = calcSize(hists);
-  //printf("histAdder size: %d\n", size);
-  hists->Delete();
-  delete hists;
-  return 0;
-}
diff --git a/Tools/RunTimeTester/share/histComparison.C b/Tools/RunTimeTester/share/histComparison.C
deleted file mode 100755
index 320253bb4094f206efba2ffae397374c2482b358..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/histComparison.C
+++ /dev/null
@@ -1,176 +0,0 @@
-//********************************
-// Root macro to read in and compare histograms from a hist and ref hist file.
-// Result is written to standard out which is caught by the Python popen object
-// used to launch the ROOT child process that ran this macro.
-//
-// BS 12/04
-//********************************
-
-//=========================================================================
-
-TObjArray* getHists(TFile * f){
-  //printf("getHists: start\n");
-  TObjArray* hists = new TObjArray(0);
-  int size = hists->GetSize();
-  //printf("getHists: size on creation: %d\n",size);  
-
-  TIter next(f->GetListOfKeys());
-  TKey *key;
-  while ((key=(TKey*)next())) {
-    char* name = key->GetName();
-    //printf("getHists: obj name %s\n", name);
-    TObject* obj = f->Get(name);
-    hists->Add(obj);
-  }  
-  
-  int size = calcSize(hists);
-  //  printf("getHists: from %s found %d hists\n", fname, size);
-  return hists;
-}
-
-//=========================================================================
-
-int calcSize(TObjArray* col){
-  //printf("calcSize: start:");
-  int count = 0;
-  TIter iter = col->MakeIterator();
-  while(iter.Next()){++count;}
-  return count;
-}
-
-//=========================================================================
-
-bool binByBinComparison(hist1,hist2){
-  Int_t nbins1 = hist1->GetNbinsX();
-  Int_t nbins2 = hist2->GetNbinsX();
-  if(nbins1 != nbins2){
-    return false;
-  }
-  else{
-    for(Int_t binNo=0;binNo < nbins1;binNo++){
-      if(hist1->GetBinContent(binNo) != hist2->GetBinContent(binNo)){
-	return false;
-      }
-    }
-  }
-
-  return true;
-}
-
-//=========================================================================
-
-bool statisticalComparison(hist1,hist2){
-  return true;
-  //  return (hist1->KolmogorovTest(hist2) < probValue);
-}
-
-//=========================================================================
-
-bool isSameHistogram(TH1* hist1, TH1* hist2, char * comparisonType){
-  if (comparisonType=='bin-by-bin'){
-    return binByBinComparison(hist1,hist2);
-  }
-  else{ //Kolmogorov test
-    return statisticalComparison(hist1,hist2);
-  }
-
-  return true;
-}
-
-//=========================================================================
-
-int cleanup(TObjArray * array1, TObjArray * array2, TFile * file1, TFile * file2){
-
-  array1->Delete();
-  delete array1;
-
-  array2->Delete();
-  delete array2;
-  
-  delete file1;
-  delete file2;
-
-  return 0;
-}
-
-//=========================================================================
-
-char * toString(TObjArray * histTitles){
-  TIter iter(histTitles);
-
-  TString total;
-  while( currStringObject = (TObjString*)iter.Next()){
-    TString tString = currStringObject->GetString();
-    total += tString->Data() + ", ";
-  }
-
-  return total->Data();
-}
-
-//=========================================================================
-
-int histComparison(char* newHistFile, char* refHistsFile, char * compType){
-  int successCode = 0;
-  int errorCode = 1;
-
-  //------------------------------------------------------------------
-  // Open the 2 histogram files and extract the histograms into 2 lists
-  //------------------------------------------------------------------
-  TFile* hists = new TFile(newHistFile); 
-  TFile* refHists = new TFile(refHistsFile); 
-    
-  listHists  = getHists(hists);
-  listRhists = getHists(refHists);
-  
-  //------------------------------------------------------------------
-  // Different number of histograms in the two histogram files -> EXIT
-  //------------------------------------------------------------------
-  if(calcSize(listHists) != calcSize(listRhists)){
-    cleanup(listHists,listRhists,hists,refHists);
-    cout << "histComparison.C return value: false info: the histogram files " << newHistFile << " and " << refHistsFile << " did not contain the same number of histograms" << endl;
-    return errorCode;
-  }
-
-  //------------------------------------------------------------------
-  // No histograms in the newly created histogram file -> EXIT
-  //------------------------------------------------------------------
-  if(calcSize(listHists) == 0 ){
-    cleanup(listHists,listRhists,hists,refHists);
-    cout << "histComparison.C return value: false info: the newly created histogram file" << newHistFile << " contains no histograms!" << endl;
-    return errorCode;
-  }
-  
-  //------------------------------------------------------------------
-  // Preliminary checks over, now compare each histogram in the two files
-  //------------------------------------------------------------------
-  TIter new_iter(listHists);
-  TIter ref_iter(listRhists);
-
-  TObjArray* failed = new TObjArray(0);
-  int sizeOnInit = failed->GetSize();
-
-  while( newHist = (TH1*)new_iter.Next()){
-    refHist = (TH1*)ref_iter.Next();
-    if (!isSameHistogram(newHist,refHist,compType)){
-      failed->Add(new TObjString(newHist->GetTitle()));
-    }
-  }
-
-  //------------------------------------------------------------------
-  // Some histograms were different -> EXIT
-  //------------------------------------------------------------------
-  if (failed->GetSize() > sizeOnInit){
-    cout << "failed size: " << failed->GetSize() << endl;
-    cout << "histComparison.C return value: false info: the following histograms were not identical in the two files " << newHistFile << " and " << refHistsFile << ": " << toString(failed) << endl;
-    cleanup(listHists,listRhists,hists,refHists);
-    failed->Delete();
-    delete failed;
-    return errorCode;
-  }
-  
-  //------------------------------------------------------------------
-  // All OK
-  //------------------------------------------------------------------
-  cout << "histComparison.C return value: true" << endl;
-  return successCode;
-}
diff --git a/Tools/RunTimeTester/share/importName.py b/Tools/RunTimeTester/share/importName.py
deleted file mode 100755
index d7d9c724baf4da6637e7ed034bd8fe4f54e2a78c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/importName.py
+++ /dev/null
@@ -1,10 +0,0 @@
-def importName(moduleName, name):
-    """From Python Cookbook
-       Import a named object from a module
-    """
-    try:
-        module = __import__(moduleName, globals(), locals(), [name])
-    except ImportError:
-        logger.error('ImportName failed to import %s %s' % (moduleName, name))
-        return None
-    return vars(module)[name]
diff --git a/Tools/RunTimeTester/share/old/RecExCommon_jobOptions.txt b/Tools/RunTimeTester/share/old/RecExCommon_jobOptions.txt
deleted file mode 100755
index c598395a1903c162d3561230be51f71645581757..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/RecExCommon_jobOptions.txt
+++ /dev/null
@@ -1,578 +0,0 @@
-//##############################################################
-//
-// Main jobOptions.file seting up general parameters 
-// and sourcing one auxilliary jobOption per major algorithm
-// To use it : athena RecExCommon_jobOptions.txt  
-//==============================================================
-//
-// uncomment to switch off dumping of jobOption file content
-//#pragma print off
- 
-// to read Zebra TDR data
-
-#include "AthenaCommon/Atlas_ZebraTDR.UnixStandardJob.txt"
-
-
-
-
-
-
-//particle property service
-#include "PartPropSvc/PartPropSvc.txt"
-
-
-// specify input file
-//
-ApplicationMgr.EvtSel = "FILE ZEBRA.P";
-//dummy (SHOULD NOT BE NECESSARY IN THE FUTURE ANY MORE)
-EventSelector.dataSetNames = {"Y01606.1"} ;
-
-// Select the events to read by their event number:
-//  Examples:   24,  35: 37, 45:   selects 24, 35 to 37 and 45 and above
-//              : 35               selects up to and including 35
-//              53:               selects from 53
-// be careful to respects the spaces after " and :
-//EventSelector.EventNumbers = " 24,  35: 37, 45: " ;
-
-//
-// Get only relevant digits from Event
-//
-EventSelector.calos = false;                        
-EventSelector.muons = false;                        
-EventSelector.trt   = false;
-EventSelector.sct   = false;
-EventSelector.pixel = false;
-
-
-// Use auditors
-ApplicationMgr.DLLs += { "GaudiAud" };
-//
-// write out a summary of the time spent
-//
-AuditorSvc.Auditors  += { "ChronoAuditor"};
-//
-// write out a short message upon entering or leaving each algorithm
-//
-//AuditorSvc.Auditors  += { "NameAuditor" };
-//
-// write out s summary of the memory usage
-//
-//AuditorSvc.Auditors  += { "MemStatAuditor" };
-//MemStatAuditor.OutputLevel = 4 ;
-//
-// write out a list of all Storegate collection with their keys and
-// lock/unlock state. Very useful for debugging purpose
-//StoreGateSvc.Dump = true ;
-//DetectorStore.Dump = true ;
-//ConditionStore.Dump = true ;
-
-// Set output level threshold 
-//(0=ALL 1=VERBOSE, 2=DEBUG, 3=INFO, 4=WARNING, 5=ERROR, 6=FATAL )
-MessageSvc.OutputLevel               = 3;
-
-// to get couloured messages (not in emacs though)
-//MessageSvc.useColors = true ;
-
-
-// Number of events to be processed 
-ApplicationMgr.EvtMax = 25;
-
-
-
-// Root Ntuple output 
-ApplicationMgr.DLLs += { "RootHistCnv" };
-ApplicationMgr.HistogramPersistency = "ROOT";
-NTupleSvc.Output = { "FILE1 DATAFILE='ntuple.root' OPT='NEW'" };
-//for root file, it is easier to start ntuple ID with a letter
-// to be removed for hbook ntuple
-CBNT_Athena.NtupleLocID="/FILE1/CBNT/t3333";
-
-
-//  Hbook Ntuple output  (uncomment these lines, comment the four previous lines)
-// Also please make sure that all ntuple ID are plain number (in
-//    particular CBNT_Athena)
-//NTupleSvc.Output =  { "FILE1 DATAFILE='ntuple.hbook' OPT='NEW'" };
-//ApplicationMgr.HistogramPersistency="HBOOK";
-//HbookHistSvc.NPAWC = 1500000 ;
-
-
-// Histogram output, if any. 
-HistogramPersistencySvc.OutputFile  = "histo.hbook";
-
-// Nova converters are necessary
-#include "NovaCnvSvc/NovaCnvSvc_jobOptions.txt" 
-#include "NovaConverters/NovaConverters_jobOptions.txt" 
-
-NovaCnvSvc.Host		= "atlasdev1.cern.ch";
-NovaCnvSvc.Database	= "parameters";
-
-// Get Geomodel and its tools
-ApplicationMgr.DLLs	+= { "GeoModelSvc", "SCT_GeoModel", "PixelGeoModel"};
-ApplicationMgr.ExtSvc	+= { "GeoModelSvc"};
-GeoModelSvc.Detectors	+= { "PixelDetectorTool", "SCT_DetectorTool" };
-//GeoModelSvc.PixelDetectorTool.Align = false;
-
-
-
-
-
-//
-// Detector description initialisation
-//
-// do not use only compact identificator 
-//DetDescrCnvSvc.CompactIDsOnly = false ;
-//
-
-#include "CaloDetMgrDetDescrCnv/CaloDetMgrDetDescrCnv_joboptions.txt"
-
-// AMDB database is needed by muon reco
-//
-#include "AmdbAth/AmdbAth_jobOptions.txt"
-//
-// muon mdet descr manager
-#include "MuonDetMgrDetDescrCnv/MuonDetMgrDetDescrCnv_joboptions.txt"
-
-
-
-// MagneticField Service 
-#include "BFieldAth/BFieldAth_jobOptions.txt"
-// if running on data with uniform magnetic field in Inner Detector
-// (which is the case for most data until dc0) this switch 
-// should be uncommented
-//MagFieldAthenaSvc.UniformFieldInID = 1 ;
-
-
-
-//--------------------------------------------------------------
-// Now specify the list of algorithms to be run
-// The order of the jobOption specify the order of the algorithms
-// (don't modify it)
-//--------------------------------------------------------------
-
-
-// CBNT_Athena.Members specify the ntuple block corresponding to a given ntuple
-// Comment CBNT_Athena.Members line to remove a ntuple block
-// It is also possible to disable a ntuple block by
-//       disabling the corresponding
-//       CBNT_Athena.Members with the line: <Member>.Enable = false ;
-// Comment #include line to remove an algorithm : it is 
-//   mandatory to also comment the corresponding
-//       CBNT_Athena line otherwise crash
-// The ntuple specification is in file CBNT_jobOptions.txt
-// and can be modified by adding properties below
-//
-// ----- CBNT_Athena algorithm
-//
-#include "RecExCommon/CBNT_jobOptions.txt"
-#include "RecExCommon/CBNT_Truth_jobOptions.txt"
-
-// run/event number
-#include "CBNT_Athena/CBNT_EventInfo_jobOptions.txt"
-
-
-
-// 
-//
-// functionality : Genz/kine->HepMc converter for true particle, Generators and Kine
-//
-ApplicationMgr.DLLs += { "GenzModule"};
-// run the  GenzKine->HepMC coverter
-ApplicationMgr.TopAlg += { "GenzModule"};
-
-// this algorithm dump the content of the MC event: big output
-//ApplicationMgr.DLLs += { "DumpMC" };
-//ApplicationMgr.TopAlg += { "DumpMC" };
-
-
-// truth information
-ApplicationMgr.DLLs += { "CBNT_Truth" };
-CBNT_Athena.Members += { "CBNT_Truth" };
-
-//
-
-
-
-//
-// functionality : Inner detector prereconstruction 
-//
-// use the new detector description (chose only one of these three files)
-//
-// for TDR data
-ApplicationMgr.DLLs += { "InDetRawUtils" };
-//#include "InDetDC1ReadPackage/SiTrackerDetDescrTDR_CnvOptions.txt"
-// use the DC1 Initial Layout (one less pixel layer, TRT wheel
-// missing) xml detector description
-//#include "InDetDC1ReadPackage/SiTrackerDetDescrInitialLayout_CnvOptions.txt"
-// use the DC1 xml detector description
-#include "InDetDC1ReadPackage/SiTrackerDetDescrDC0_CnvOptions.txt"
-
-
-#include "InDetDC1ReadPackage/PixelDetDescrCnvOptions.txt"
-#include "InDetDC1ReadPackage/SCT_DetDescrCnvOptions.txt"
-
-
-// Generate RDOs
-#include "InDetDC1ReadPackage/PixelRDOCnvOptions.txt"
-#include "InDetDC1ReadPackage/SCT_RDOCnvOptions.txt"
-#include "InDetDC1ReadPackage/TRT_RDOCnvOptions.txt"
-
-// Simulated data selection (noise/efficiency/threshold settings)
-#include "InDetSimDataSelector/PixelSimDataSelectorOptions.txt"
-#include "InDetSimDataSelector/SCT_SimDataSelectorOptions.txt"
-
-
-// and SiTrackerClusters
-#include "SiClusterization/MergedPixelsOptions.txt"
-#include "SiClusterization/SCT_ClustersOptions.txt"
-
-// and SiTrackerSpacePoints
-#include "SiSpacePointMaker/SiTrackerSpacePointFinderOptions.txt"
-
-
-//
-// functionality :  xKalman track finding
-//
-#include "xKalmanppAthena/xKalman_jobOptions.txt"
-CBNT_Track.TracksNames += { "Tracks_xKalman/1" };
-//
-// optionally switch back to RD's event 
-// 
-//XKaEventCollection.InputForTRTclusters     = 0 ; // 0->RD ,1->RDO (default),2->RIO
-//XKaEventCollection.InputForSILclusters     = 0 ; // 0->RD ,1->RI0 (default)
-
-
-
-//
-// functionality : iPatRec TrackFinder/Follower/Fitter
-//
-
-#include "iPatRecAlgs/iPatRec_jobOptions.txt"
-//temporary: reduce printlevel as the default is too verbose
-iPatRec.printLevel		= 0;
-
-// iPatRec TrackTruth
-#include "iPatRecAlgs/iPatTrackTruthAssociator_jobOptions.txt"
-
-
-// iPatRec statistics
-#include "iPatRecAlgs/iPatStatistics_jobOptions.txt"
-
-// iPatRec SimpleTrackCollection output
-#include "iPatRecAlgs/iPatSimpleTracks_jobOptions.txt"
-CBNT_Track.TracksNames += { "iPatSimpleTracks/2" };
-
-
-//
-// functionality : put tracks from any tracking algorithm in combined ntuple
-//
-ApplicationMgr.DLLs += { "SimpleTrackAlgs" };
-CBNT_Athena.Members += { "CBNT_Track"};
-
-
-//
-// functionality : LAr Calorimeter cell reconstruction 
-//
-//(only one of these  
-// two # include lines should be uncommented ) 
-//
-// - complete RDO and cell production (needed for the TRIGGER)
-//
-//#include "LArCellRec/LArCellFromLArRaw_MC_jobOptions.txt"
-//   uncomment this line if you want to switch on the noise in LArg calo
-//   (be aware that noise will add 60s/event)
-//digitmaker1.NoiseOnOff = true ; // switch noise on
-//
-// - direct cell production (only if TRIGGER is not run) 
-//
-#include "LArCellRec/LArCell_jobOptions.txt"             
-//   uncomment these lines if you want to switch on the noise in LArg cal
-//   (be aware that noise will add 16s/event)
-//cellbuilderEM.WithNoise=true; // switch noise on in e.m
-//cellbuilderHEC.WithNoise=true; // switch noise on in HEC
-//cellbuilderFCal.WithNoise=true; // switch noise on in FCAL
-
-// to use the BNL mysql noise database instead of the CERN one
-// uncomment this line (identical results)
-//LArConditionsDBSvc.Hostname = "atlassw1.phy.bnl.gov";
-// for info the cern server (default) is:
-//LArConditionsDBSvc.Hostname = "atlasdev1.cern.ch";
-
-
-CBNT_Athena.Members  += { "CBNT_CaloCell/CBNT_LArCell" }; 
-
-
-
-//
-// functionality : Tile Calorimeter cell reconstruction
-//
-#include "TileRec/TileRec_jobOptions.txt"
-// comment this line to switch off the noise in TIle calo
-TileInfoLoader.TileNoise		= false;
-// switch off zero suppresion in tile for time being
-TileInfoLoader.TileZeroSuppress         = false;
-
-CBNT_Athena.Members  += { "CBNT_CaloCell/CBNT_TileCell" }; 
-//
-
-//
-// functionality : Calorimeter combined clustering
-//
-
-#include "CaloRec/CaloCluster_jobOptions.txt"
-
-// only for very detailed calo info
-CBNT_Athena.Members += { "CBNT_CaloInfo"};
-
-//
-// functionality : LAr Calorimeter clustering
-//
-#include "LArClusterRec/LArCluster_jobOptions.txt"
-// introduce multisize possibility
-#include "LArClusterRec/LArCluster_MultiSize_jobOptions.txt"
-CBNT_Athena.Members += { "CBNT_CaloCluster"};
-// different clustersize
-CBNT_Athena.Members += { "CBNT_CaloCluster/CBNT_CaloCluster35" };
-CBNT_Athena.Members += { "CBNT_CaloCluster/CBNT_CaloCluster37" };
-
-//
-// functionality : Topological clustering
-//
-#include "LArClusterRec/LArTopoCluster_jobOptions.txt"
-CBNT_Athena.Members += { "CBNT_CaloCluster/CBNT_CaloClusterTopo" };
-
-//
-// functionality: Muonbox muon Reconstruction
-//
-//muonbox
-//#include "MuonboxAth/Muonbox_jobOptions.txt"
-// old DC1 data
-//MuonboxCosignalMaker.MaximumDriftTime = 500. ; 
-// redigitised data
-MuonboxCosignalMaker.MaximumDriftTime = 700. ; 
-MuonboxConfigMaker.PrintLevel            = 0    ;
-
-//new f90 version of muonbox (at cern/bnl only so far)
-// cannot be used together with MuonboxAthena
-#include "MuonboyAth/Muonboy_jobOptions.txt"
-// old DC1 data
-//MuonboyCosignalMaker.MaximumDriftTime = 500. ; 
-// redigitised data
-MuonboyCosignalMaker.MaximumDriftTime = 700. ; 
-MuonboyConfigMaker.PrintLevel            = 0    ;
-
-CBNT_Athena.Members  += { "CBNT_Muon" }; 
-
-
-//
-// functionality : Moore muon reconstruction
-//
-#include "MooAlgs/Moore_RecEx_jobOptions.txt"
-CBNT_Athena.Members  += { "MooMakeNtuple" }; 
-//old DC1 data 
-//MooMdtDigitToHitSvc.max_drift_time = 500; 
-//redigitised  data
-MooMdtDigitToHitSvc.max_drift_time = 700; 
-
-
-//------------------------------------------------
-// now combined reconstruction algorithms
-//------------------------------------------------
-//
-// functionality : primary vertex and vertexing interface
-//
-#include "VertexPrimary/Vertex_jobOptions.txt"
-CBNT_Athena.Members  += { "CBNT_Vertex" }; 
-
-//
-// functionality : photon Conversion
-//
-#include "ConversionAlgs/Conversion_jobOptions.txt"
-CBNT_Athena.Members  += { "CBNT_Conversion" }; 
-
-
-//
-// functionality : CaloTower protojets + preclustering + KT algorithm + CombinedJetAlg
-//
-#include "JetRec/KtJetTower_jobOptions.txt"
-//
-// to switch to cone reco jet uncomment this line and comment the next
-// BE CAREFUL to do the same for the truth jets 
-//
-//#include "JetRec/ConeJetTower_jobOptions.txt"
-//
-// CBNT without jet cell sums for H1 weights
-CBNT_Athena.Members += { "CBNT_CombinedJet" };
-CBNT_CombinedJet.inputLoc = "CombinedJets" ;
-//
-// CBNT including jet cell sums for H1 weights requires setting DoCellSums=1
-// combJet.DoCellSums = 1
-//CBNT_Athena.Members += { "CBNT_JetSums" };
-//
-
-//
-// functionality: MC Truth protojets + preclustering + KT algorithm + CombinedJetAlg
-//
-//
-CBNT_Athena.Members += { "CBNT_Jet/CBNT_JetTruth" };
-#include "JetRec/KtJetTruth_jobOptions.txt"
-CBNT_JetTruth.inputLoc = "TruthKtJets";
-//
-// to switch to truth cone jet uncomment this line and comment the next
-// BE CAREFUL to do the same for the reco jets 
-//
-//#include "JetRec/ConeJetTruth_jobOptions.txt"
-//CBNT_JetTruth.inputLoc = "SplitMergeTruthJets";
-
-//
-// functionality : tau identification
-//
-#include "tauRec/tauRec_jobOptions.txt"
-CBNT_Athena.Members  += { "CBNT_tau" }; 
-
-//
-// functionality : electron photon identification
-//
-//
-
-#include "egammaRec/egammaRec_jobOptions.txt"
-CBNT_Athena.Members += { "CBNT_egamma" };
-CBNT_Athena.Members += { "CBNT_egTruth" };
-
-//
-// functionality : muonIdentification 
-//
-
-#include "MuonIdentification/MuonIdentification_jobOptions.txt"
-CBNT_Athena.Members += { "MuidNtupleKINE"};     
-MuidNtupleKINE.maxNumberofKINE = 10 ;
-CBNT_Athena.Members += { "MuidIDNtuple"};
-MuidIDNtuple.maxNumberofRECO = 10 ;
-CBNT_Athena.Members += { "MuidInitNtuple"};
-MuidInitNtuple.maxNumberofRECO = 10 ;
-CBNT_Athena.Members += { "MuidNtuple"};
-MuidNtuple.maxNumberofRECO = 10 ;
-CBNT_Athena.Members += { "MuidCombNtuple"};
-MuidCombNtuple.maxNumberofRECO = 10 ;
-
-
-
-//
-// functionality : Missing Et
-//
-
-#include "MissingET/MissingET_jobOptions.txt"
-CBNT_Athena.Members += { "CBNT_MissingET" };
-CBNT_Athena.Members += { "CBNT_MissingETtruth" };
-
-//
-// functionality : energy flow 
-//
-#include "eflowRec/CaloClusterForEflow_jobOptions.txt"
-#include "eflowRec/eflowRec_jobOptions.txt"
-CBNT_Athena.Members += { "CBNT_eflow" };
-
-//
-// functionality : Jet-finding on eflowObjects 
-// can be performed by uncommenting the next
-// three lines (using cone or Kt jobOptions and CBNT parameters)
-
-//#include "JetRec/ConeJetEflow_jobOptions.txt"
-//CBNT_Athena.Members += { "CBNT_Jet/CBNT_JetEflow" };
-//CBNT_JetEflow.inputLoc = "EflowConeJets";
-
-
-
-//
-// functionality : All TRIGGER code bytestream, LVL1 and LVL2
-//
-
-//#include "TriggerRelease/jobOfragment_forRecExCommon.txt"
-// trigger combined ntuple
-//#include "TriggerRelease/jobOfragment_TriggerCBNT.txt"
-
-
-//
-// functionality : atlfast 
-// disabled by default.
-//uncomment these two lines
-//#include "AtlfastAlgs/Atlfast_CBNT.txt"
-//CBNT_Athena.Members  += { "Atlfast::CBNT_Atlfast/CBNT_Atlfast" }; 
-
-
-//
-// user code and property modifiers should come below
-//
-
-
-
-
-//
-// functionality : monitor memory and cpu time
-//
-#include "CBNT_Athena/CBNT_Audit_jobOptions.txt"
-
-//
-// functionality : build combined ntuple, 
-// gathering info from all the reco algorithms
-//
-#include "CBNT_Athena/CBNT_Athena_jobOptions.txt"
-
-
-// uncomment this block to write output
-// WARNING nova converters need be switched off
-/*
-//--------------------------------------------------------------
-// now write out Transient Event Store content in POOL
-//--------------------------------------------------------------
-//
-
-#include "AthenaPoolCnvSvc/WriteAthenaPool_jobOptions.txt"
-
-// check dictionary
-#include "AthenaSealSvc/AthenaSealSvc_joboptions.txt"
-AthenaSealSvc.CheckDictionary = true;
-
-// Define the output Db parameters (the default value are shown)
-PoolSvc.Output    = "SimplePoolFile.root";
-// PoolSvc.DbServer  =  "db1.usatlas.bnl.gov";
-// PoolSvc.DbAccount =  "athena";
-// PoolSvc.DbPassword = "";
-PoolSvc.DbType    = "root";
-// PoolSvc.ConnectionType = "MySQLCollection";
-// PoolSvc.FullConnection = "mysql://athena:insider@db1.usatlas.bnl.gov/pool_collection";
-PoolSvc.Collection = "NewPoolTry";
-
-
-
-// Converters:
-#include "GeneratorObjectsAthenaPool/GeneratorObjectsAthenaPool_joboptions.txt"
-#include "EventAthenaPool/EventAthenaPool_joboptions.txt"
-#include "RecAthenaPool/RecAthenaPool_joboptions.txt"
-
-
-// list of output objects 
-
-//HepMC record
-Stream1.ItemList+={"McEventCollection#*"};
-
-// MissingET
-Stream1.ItemList+={"MissingET#*"};
-
-// EventInfo
-Stream1.ItemList+={"EventInfo#*"};
-
-// SimpleTrackCollection
-Stream1.ItemList+={"SimpleTrackCollection#*"};
-
-// JetCollection<CombinedJet>
-Stream1.ItemList+={"JetCollection<CombinedJet>#*"};
-
-
-//--------------------------------------------------------------------
-// switch off the writing
-//ApplicationMgr.OutStream = { };
-*/
-
-
-
-
diff --git a/Tools/RunTimeTester/share/old/TestAtlfastDoNotTouch.txt b/Tools/RunTimeTester/share/old/TestAtlfastDoNotTouch.txt
deleted file mode 100755
index e292392851b710e0dd016ceb1d98a5b0b0b41672..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/TestAtlfastDoNotTouch.txt
+++ /dev/null
@@ -1,234 +0,0 @@
-//##############################################################
-// AtlfastDoNotTouch
-// To be changed by experts only
-//##############################################################
-Atlfast.members = {
-//    Global Event Data:
-        "Atlfast::GlobalEventDataMaker/GlobalEventDataMaker",
-// 
-//    Make calorimeter cell map for the event:
-        "Atlfast::CellMaker/CellMaker", 
-//
-        "Atlfast::TrackMaker/TrackMaker",
-
-//    Make clusters based on  cells
-        "Atlfast::ClusterMaker/ClusterMaker",
-//
-//    Make electrons 
-        "Atlfast::DefaultReconstructedParticleMaker/ElectronMaker",
-//
-//    Make photons
-        "Atlfast::DefaultReconstructedParticleMaker/PhotonMaker",
-//
-//    Make muons 
-	"Atlfast::DefaultReconstructedParticleMaker/MuonMaker",
-//
-//    Find Associations between eor gamma and clusters 
-//	"Atlfast::Associator/Associator",
-//
-//    Sort the electrons into isolated and non-isolated
-	"Atlfast::Isolator/ElectronIsolator",
-//
-//    Sort the photons into isolated and non-isolated
-        "Atlfast::Isolator/PhotonIsolator",
-//
-//    Sort the muons into isolated and non-isolated
-        "Atlfast::Isolator/MuonIsolator",
-//
-//    Make Jets
-        "Atlfast::JetMaker/JetMaker",
-//
-//    Correct Jets
-        "Atlfast::AtlfastB/AtlfastB",
-//
-//    Make event header
-        "Atlfast::EventHeaderMaker/EventHeaderMaker"
-};
-//---------------------------
-// Locations not to be touched
-//---------------------------
-// Cell Maker 
-CellMaker.InputLocation		= "/Event/McEventCollection";
-CellMaker.OutputLocation        = "/Event/AtlfastCells"; 
-//-------------
-//Cluster Maker
-//-------------
-//Cluster Maker
-ClusterMaker.ProcessCells       = true;
-ClusterMaker.ProcessTracks      = false;
-ClusterMaker.InputCellLocation  = "/Event/AtlfastCells";
-ClusterMaker.InputTrackLocation = "/Event/AtlfastTracks";
-ClusterMaker.OutputLocation     = "/Event/AtlfastClusters";
-ClusterMaker.UnusedCellLocation	= "/Event/AtlfastUnusedCells";
-//------------
-// Track Maker
-//------------
-TrackMaker.MC_eventLocation  = "/Event/McEventCollection";
-TrackMaker.OutputLocation    = "/Event/AtlfastTracks"; 
-//------------
-// Track Dumper
-//------------
-TrackDumper.InputLocation     = "/Event/AtlfastTracks"; 
-//------------------
-// Particles Makers
-//------------------
-// electron 
-ElectronMaker.OutputLocation               = "/Event/AtlfastElectrons";
-ElectronMaker.MC_eventLocation		   = "/Event/McEventCollection";
-//muon
-MuonMaker.OutputLocation                   = "/Event/AtlfastMuons";
-MuonMaker.MC_eventLocation		   = "/Event/McEventCollection";
-//photon
-PhotonMaker.OutputLocation                 = "/Event/AtlfastPhotons";
-PhotonMaker.MC_eventLocation		   = "/Event/McEventCollection";
-//----------
-// Jet Maker
-//----------
-JetMaker.InputLocation                     = "/Event/AtlfastClusters";
-JetMaker.OutputLocation                    = "/Event/AtlfastJets";
-JetMaker.MuonLocation                      = "/Event/AtlfastNonIsolatedMuons";
-JetMaker.MissingMomentumLocation	   = "/Event/AtlfastMissingMomentum";
-JetMaker.McTruthLocation		   = "/Event/McEventCollection";
-JetMaker.UnusedCellLocation		   = "/Event/AtlfastUnusedCells";
-//----------
-// AtlfastB
-//----------
-AtlfastB.InputLocation                     = "/Event/AtlfastJets";
-AtlfastB.OutputLocation                    = "/Event/AtlfastBJets";
-//-------------------
-// Event Header Maker 
-//------------------- 
-EventHeaderMaker.MissingMomentumLocation  = "/Event/AtlfastMissingMomentum";
-EventHeaderMaker.ElectronLocation	  = "/Event/AtlfastIsolatedElectrons";
-EventHeaderMaker.PhotonLocation		  = "/Event/AtlfastIsolatedPhotons";
-EventHeaderMaker.IsolatedMuonLocation     = "/Event/AtlfastIsolatedMuons";
-EventHeaderMaker.NonIsolatedMuonLocation  = "/Event/AtlfastNonIsolatedMuons";
-EventHeaderMaker.JetLocation              = "/Event/AtlfastJets";
-EventHeaderMaker.McTruthLocation          = "/Event/McEventCollection";
-EventHeaderMaker.OutputLocation           = "/Event/AtlfastEventHeader";
-//----------
-// Isolators
-//----------
-// electron
-ElectronIsolator.InputLocation             = "/Event/AtlfastElectrons";
-ElectronIsolator.IsolatedOutputLocation    = "/Event/AtlfastIsolatedElectrons";
-ElectronIsolator.NonIsolatedOutputLocation = "/Event/AtlfastNonIsolatedElectrons";
-ElectronIsolator.CellLocation              = "/Event/AtlfastCells";
-ElectronIsolator.Clusterlocation           = "/Event/AtlfastClusters";
-//photon
-PhotonIsolator.InputLocation               = "/Event/AtlfastPhotons";
-PhotonIsolator.IsolatedOutputLocation      = "/Event/AtlfastIsolatedPhotons";
-PhotonIsolator.NonIsolatedOutputLocation   = "/Event/AtlfastNonIsolatedPhotons";
-PhotonIsolator.CellLocation                = "/Event/AtlfastCells";
-PhotonIsolator.ClusterLocation             = "/Event/AtlfastClusters";
-//muon
-MuonIsolator.InputLocation                 = "/Event/AtlfastMuons";
-MuonIsolator.IsolatedOutputLocation        = "/Event/AtlfastIsolatedMuons";
-MuonIsolator.NonIsolatedOutputLocation     = "/Event/AtlfastNonIsolatedMuons";
-MuonIsolator.CellLocation                  = "/Event/AtlfastCells";
-MuonIsolator.ClusterLocation               = "/Event/AtlfastClusters";
-MuonIsolator.RClusterMatch                 = 0.0;
-//---------------------
-// StandardNtuple Maker 
-//--------------------- 
-StandardNtupleMaker.JetLocation		    = "/Event/AtlfastJets";
-StandardNtupleMaker.ElectronLocation	    = "/Event/AtlfastIsolatedElectrons";
-StandardNtupleMaker.IsolatedMuonLocation    = "/Event/AtlfastIsolatedMuons";
-StandardNtupleMaker.NonIsolatedMuonLocation = "/Event/AtlfastNonIsolatedMuons";
-StandardNtupleMaker.PhotonLocation	    = "/Event/AtlfastIsolatedPhotons";
-StandardNtupleMaker.McTruthLocation         = "/Event/McEventCollection";
-StandardNtupleMaker.EventHeaderLocation     = "/Event/AtlfastEventHeader";
-//-------------------------
-// FinalStateParticleDumper 
-//-------------------------
-FinalStateParticleDumper.InputLocation	= "/Event/McEventCollection";
-//-------
-//Dumpers
-//-------
-ElectronDumper.InputLocation               = "/Event/AtlfastElectrons";
-//
-MuonDumper.InputLocation                   = "/Event/AtlfastMuons";
-//
-IsolatedElectronDumper.InputLocation       = "/Event/AtlfastIsolatedElectrons";
-//
-PhotonDumper.InputLocation                 = "/Event/AtlfastPhotons";
-//
-IsolatedPhotonDumper.InputLocation         = "/Event/AtlfastIsolatedPhotons";
-//-----------------------
-//StandardHistogram Maker
-//-----------------------
-StandardHistogramMaker.ElectronLocation    = "/Event/AtlfastElectrons";
-StandardHistogramMaker.PhotonLocation      = "/Event/AtlfastPhotons";
-StandardHistogramMaker.IsolatedElectronLocation = "/Event/AtlfastIsolatedElectrons";
-StandardHistogramMaker.IsolatedPhotonLocation   = "/Event/AtlfastIsolatedPhotons";
-StandardHistogramMaker.CellLocation        = "/Event/AtlfastCells";
-StandardHistogramMaker.ClusterLocation     = "/Event/AtlfastClusters";
-//
-ElectronHistogramMaker.InputLocation       = "/Event/AtlfastElectrons";
-ElectronHistogramMaker.McTruthLocation     = "/Event/McEventCollection";
-ElectronHistogramMaker.ParticleType        = 11;
-//
-PhotonHistogramMaker.InputLocation         = "/Event/AtlfastPhotons";
-PhotonHistogramMaker.McTruthLocation       = "/Event/McEventCollection";
-PhotonHistogramMaker.ParticleType          = 22;
-//
-MuonHistogramMaker.InputLocation           = "/Event/AtlfastMuons";
-MuonHistogramMaker.McTruthLocation         = "/Event/McEventCollection";
-MuonHistogramMaker.ParticleType            = 13;
-//----------
-//SimpleTest
-//----------
-SimpleTest1.AtlfastEventLocation           = "/Event/Atlfast" ;
-SimpleTest1.JetLocation                    = "/Event/AtlfastJets";
-SimpleTest1.ElectronLocation               = "/Event/AtlfastIsolatedElectrons";
-SimpleTest1.IsolatedMuonLocation           = "/Event/AtlfastIsolatedMuons";
-SimpleTest1.NonIsolatedMuonLocation        = "/Event/AtlfastNonIsolatedMuons";
-SimpleTest1.PhotonLocation                 = "/Event/AtlfastIsolatedPhotons";
-//SimpleTest1.TrackLocation                  = "/Event/AtlfastTracks";
-SimpleTest1.McTruthLocation                = "/Event/McEventCollection";
-SimpleTest1.TriggerLocation                = "/Event/AtlfastTrigger";
-SimpleTest1.EventHeaderLocation            = "/Event/AtlfastEventHeader";
-SimpleTest1.outputLocation                 = "/Event/SimpleTest1";
-//
-SimpleTest2.AtlfastEventLocation           = "/Event/Atlfast" ;
-SimpleTest2.JetLocation                    = "/Event/AtlfastJets";
-SimpleTest2.ElectronLocation               = "/Event/AtlfastIsolatedElectrons";
-SimpleTest2.IsolatedMuonLocation           = "/Event/AtlfastIsolatedMuons";
-SimpleTest2.NonIsolatedMuonLocation        = "/Event/AtlfastNonIsolatedMuons";
-SimpleTest2.PhotonLocation                 = "/Event/AtlfastIsolatedPhotons";
-//SimpleTest2.TrackLocation                  = "/Event/AtlfastTracks";
-SimpleTest2.McTruthLocation                = "/Event/McEventCollection";
-SimpleTest2.TriggerLocation                = "/Event/AtlfastTrigger";
-SimpleTest2.EventHeaderLocation            = "/Event/AtlfastEventHeader";
-SimpleTest2.outputLocation                 = "/Event/SimpleTest2";
-//----------
-//CellDumper
-//----------
-CellDumper.inputLocation                   = "/Event/AtlfastCells";
-//------------------------
-//FinalStateParticleDumper
-//------------------------
-FinalStateParticleDumper.inputLocation     = "/Event/McEventCollection";
-//----------
-//SimpleDiff
-//----------
-SimpleDiff.inputLocation1                  = "/Event/SimpleTest1";
-SimpleDiff.inputLocation2                  = "/Event/SimpleTest2";
-//----------
-//Associator
-//----------
-Associator.ElectronLocation                = "/Event/AtlfastElectrons";
-Associator.PhotonLocation                  = "/Event/AtlfastPhotons"; 
-Associator.MuonLocation                    = "/Event/AtlfastMuons";   
-Associator.ClusterLocation                 = "/Event/AtlfastClusters";
-Associator.JetLocation                     = "/Event/AtlfastJets";
-Associator.AssocElectronLocation           = "/Event/AtlfastAssocElectrons";
-Associator.AssocPhotonLocation             = "/Event/AtlfastAssocPhotons";
-Associator.AssocMuonLocation               = "/Event/AtlfastAssocMuons";
-//--------------------------
-// End of DoNotTouch options
-//-------------------------- 
-
-
-
-
diff --git a/Tools/RunTimeTester/share/old/TestAtlfastOptions.txt b/Tools/RunTimeTester/share/old/TestAtlfastOptions.txt
deleted file mode 100755
index de598dda60a83c43570b707f22374ddaa8e62979..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/TestAtlfastOptions.txt
+++ /dev/null
@@ -1,123 +0,0 @@
-#include "RunTimeTester/TestAtlfastDoNotTouch.txt"
-//--------------------------------------------------------------
-// Algorithm Parameters set to Default Values 
-//--------------------------------------------------------------
-// Default Reconstructed Particle Maker
-//-------------------------------------
-// options for electron maker
-ElectronMaker.ParticleType               = 11;
-ElectronMaker.mcMinimumPt                = 0.0;
-ElectronMaker.mcMaximumEta               = 100.0;
-ElectronMaker.MinimumPt                  = 5000.0;
-ElectronMaker.MaximumEta                 = 2.5;
-ElectronMaker.DoSmearing		 = true;
-//
-// options for muon maker
-MuonMaker.ParticleType                   = 13;
-MuonMaker.mcMinimumPt                    = 500.0;
-MuonMaker.mcMaximumEta                   = 2.5;
-MuonMaker.MinimumPt                      = 6000.0;
-MuonMaker.MaximumEta                     = 100.0;
-MuonMaker.DoSmearing                     = true;
-MuonMaker.MuonSmearKey			 = 3;
-//
-//options for photon maker
-PhotonMaker.ParticleType                 = 22;
-PhotonMaker.mcMinimumPt                  = 0.0;
-PhotonMaker.mcMaximumEta                 = 100.0;
-PhotonMaker.MinimumPt                    = 5000.0;
-PhotonMaker.MaximumEta                   = 2.5;
-PhotonMaker.DoSmearing			 = true;
-//--------------------------------------------------------------
-// Cell Maker
-//-----------------------------------
-CellMaker.EtaCoverage        = 5.0;
-CellMaker.MinETCell          = 0.0;
-CellMaker.GranBarrelEta      = 0.1;
-CellMaker.GranBarrelPhi      = 0.1;
-CellMaker.GranForwardEta     = 0.2;
-CellMaker.GranForwardPhi     = 0.2;
-CellMaker.FastShower         = false;
-//--------------------------------------------------------------
-// Cluster Maker
-//-------------------------------------  
-ClusterMaker.ProcessCells       = true;
-ClusterMaker.ProcessTracks      = false;
-ClusterMaker.RConeBarrel        = 0.401;
-ClusterMaker.RConeForward       = 0.401;
-ClusterMaker.minInitiatorET     = 1500.;
-ClusterMaker.minClusterET       = 5000.0;
-ClusterMaker.Strategy           = "Cone";
-
-//--------------------------------------------------------------
-// Jet Maker
-//-------------------------------------  
-JetMaker.MinimumPT        = 10000.0;
-JetMaker.MaximumEta       = 5.0;
-JetMaker.DoSmearing       = true;
-JetMaker.RconeB           = 0.400;
-JetMaker.RconeF		  = 0.400;
-JetMaker.bPtMin		  = 5000.0;  
-JetMaker.bMaxDeltaR       = 0.2;
-JetMaker.cPtMin		  = 5000.0;
-JetMaker.cMaxDeltaR       = 0.2;
-JetMaker.tauPtMin         = 10000.0;
-JetMaker.tauMaxDeltaR     = 0.3;
-JetMaker.etaTagMax        = 2.5;
-JetMaker.tauJetPtRatio    = 0.9;
-//--------------------------------------------------------------
-// AtlfastB (jet corrector)
-//-------------------------------------
- AtlfastB.AtlfBjetSwitch   =true;
-AtlfastB.AtlfCalSwitch     =true;
-AtlfastB.AtlfTauSwitch     =true;
-AtlfastB.AtlfTauVetoSwitch =false;
-AtlfastB.AtlfTrigMuoSwitch =false;
-AtlfastB.JetCorrFile       ="./AtlfastBjet.dat";
-AtlfastB.CJetCorrFile      ="./AtlfastBcjet.dat";
-AtlfastB.AtlfBNSet         = 1; 
-AtlfastB.TauEff            = 0.05;
-AtlfastB.TauVetoOption     = 1;
-//--------------------------------------------------------------
-// Track Maker
-//------------------------------------- 
-TrackMaker.McPtMinimum	     = 500.;
-TrackMaker.McEtaMaximum      = 2.5;
-TrackMaker.vtMaximum         = 3.0;
-TrackMaker.vlMaximum         = 40.0;
-TrackMaker.DoSmearing	      = true;
-TrackMaker.BField	      = 2.0;
-//--------------------------------------------------------------
-// Event Header Maker 
-//------------------------------------- 
-EventHeaderMaker.BeamEnergy               = 7000000.0;
-EventHeaderMaker.TestMode		  = 0;
-//--------------------------------------------------------------
-// FinalStateParticleDumper 
-//------------------------------------
-// choose one of the following values for SelectorName
-//FinalStateParticleDumper.SelectorName	= "Z0selector";
-//FinalStateParticleDumper.SelectorName	= " ";
-//FinalStateParticleDumper.SelectorName	= "qselector";
-//FinalStateParticleDumper.SelectorName	= "bSelector";
-FinalStateParticleDumper.SelectorName	= "All";
-//-------------------------------------- 
-// GlobalEventData Maker 
-//------------------------------------- 
-//Luminosity option: 1=low, 2-high
-GlobalEventDataMaker.Luminosity		= 1;
-GlobalEventDataMaker.Invisibles		= {12,14,16}; 
-GlobalEventDataMaker.RandomSeed		= 12345;
-GlobalEventDataMaker.BFieldOn		= true;
-GlobalEventDataMaker.BarrelForwardEta	= 3.2;
-
-
-
-
-
-
-
-
-
-
-
diff --git a/Tools/RunTimeTester/share/old/artemis.C b/Tools/RunTimeTester/share/old/artemis.C
deleted file mode 100755
index 961b7d6dca71baec7440984d3bb694e18b29988c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/artemis.C
+++ /dev/null
@@ -1,89 +0,0 @@
-
-class ArtHists{
-public:
-  ArtHists(int base, char* fn):{
-    cout<<"ArtHists base "<<base<<endl;
-    TFile* f1 =new TFile(fn);
-    if(base == 100){   
-    m_h1 = (TH1F*)f1.Get("100");
-    m_h2 = (TH1F*)f1.Get("101");
-    m_h3 = (TH1F*)f1.Get("102");
-    m_h4 = (TH1F*)f1.Get("200");
-    m_h5 = (TH1F*)f1.Get("201");
-    m_h6 = (TH1F*)f1.Get("202");
-    }elseif(base==110){
-    m_h1 = (TH1F*)f1.Get("110");
-    m_h2 = (TH1F*)f1.Get("111");
-    m_h3 = (TH1F*)f1.Get("112");
-    m_h4 = (TH1F*)f1.Get("210");
-    m_h5 = (TH1F*)f1.Get("211");
-    m_h6 = (TH1F*)f1.Get("212");
-    }elseif(base==120){
-    m_h1 = (TH1F*)f1.Get("120");
-    m_h2 = (TH1F*)f1.Get("121");
-    m_h3 = (TH1F*)f1.Get("122");
-    m_h4 = (TH1F*)f1.Get("220");
-    m_h5 = (TH1F*)f1.Get("221");
-    m_h6 = (TH1F*)f1.Get("222");
-    }elseif(base==130){
-    m_h1 = (TH1F*)f1.Get("130");
-    m_h2 = (TH1F*)f1.Get("131");
-    m_h3 = (TH1F*)f1.Get("132");
-    m_h4 = (TH1F*)f1.Get("230");
-    m_h5 = (TH1F*)f1.Get("231");
-    m_h6 = (TH1F*)f1.Get("232");
-    }
-    
-  }
-  void Draw(TCanvas* can){
-    can->Clear();
-    can->Divide(3,2);
-    can->cd(1);
-    m_h1->Draw();
-    can->cd(2);
-    m_h2->Draw();
-    can->cd(3);
-    m_h3->Draw();
-    can->cd(4);
-    m_h4->Draw();
-    can->cd(5);
-    m_h5->Draw();
-    can->cd(6);
-    m_h6->Draw();
-  }
-  void Print(TCanvas* c){
-    c->Print(fn);
-  }
- 
-private:
-  TH1F* m_h1;
-  TH1F* m_h2;
-  TH1F* m_h3;
-  TH1F* m_h4;
-  TH1F* m_h5;
-  TH1F* m_h6;
-  
-};
-
-int artemis(){
-   
-  TCanvas* can = new TCanvas;
-  
-  ArtHists a1(100, "histo.hbook");
-  a1.Draw(can);
-  can->Print("plot_0", "gif");
-  
-  ArtHists a2(110, "histo.hbook");
-  a2.Draw(can);
-  can->Print("plot_1.ps");
-  
-  ArtHists a3(120, "histo.hbook");
-  a3.Draw(can);
-  can->Print("plot_2.ps");
-  
-  ArtHists a4(130, "histo.hbook");
-  a4.Draw(can);
-  can->Print("plot_3.ps");
-
-  return 0;
-}
diff --git a/Tools/RunTimeTester/share/old/atlfast.C b/Tools/RunTimeTester/share/old/atlfast.C
deleted file mode 100755
index a88fe705ebd8eed6a05e38eaee1e71475800144d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/atlfast.C
+++ /dev/null
@@ -1,426 +0,0 @@
-class AtlHists{
-public:
-  AtlHists(char* fn){
-    m_file =new TFile(fn);
-
-    this->nonIsolatedElectronHists();    
-    this->isolatedElectronHists();
-    this->nonIsolatedPhotonHists();
-    this->isolatedPhotonHists();
-    this->nonIsolatedMuonHists();
-    this->isolatedMuonHists();
-    this->nonIsolatedMuonHists();
-    this->isolatedMuonHists();
-    this->cellHists();
-    this->clusterHists();
-    this->jetHists();
-    this->jetBHists();
-    this->eventHeaderHists1();
-    this->eventHeaderHists2();
-  }
-
-  void nonIsolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/1");
-    m_h2 = (TH1F*)m_file->Get("simple/2");
-    m_h3 = (TH1F*)m_file->Get("simple/3");
-    m_h4 = (TH1F*)m_file->Get("simple/4");
-    m_h5 = (TH1F*)m_file->Get("simple/5");
-    m_h6 = (TH1F*)m_file->Get("simple/6");
-    
-    TPostScript ps("nonIsolatedElectron.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-
-  void isolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/7");
-    m_h2 = (TH1F*)m_file->Get("simple/8");
-    m_h3 = (TH1F*)m_file->Get("simple/9");
-    m_h4 = (TH1F*)m_file->Get("simple/10");
-    m_h5 = (TH1F*)m_file->Get("simple/11");
-    m_h6 = (TH1F*)m_file->Get("simple/12");
-
-    TPostScript ps("isolatedElectron.ps", 112);
-    this->Draw();
-    ps.Close();
-
-  }
-    
-
-  void nonIsolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/13");
-    m_h2 = (TH1F*)m_file->Get("simple/14");
-    m_h3 = (TH1F*)m_file->Get("simple/15");
-    m_h4 = (TH1F*)m_file->Get("simple/16");
-    m_h5 = (TH1F*)m_file->Get("simple/17");
-    m_h6 = (TH1F*)m_file->Get("simple/18");
-
-    TPostScript ps("nonIsolatedPhoton.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-    return;
-  }
-
-  void isolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/19");
-    m_h2 = (TH1F*)m_file->Get("simple/20");
-    m_h3 = (TH1F*)m_file->Get("simple/21");
-    m_h4 = (TH1F*)m_file->Get("simple/22");
-    m_h5 = (TH1F*)m_file->Get("simple/23");
-    m_h6 = (TH1F*)m_file->Get("simple/24");
-
-    TPostScript ps("isolatedPhoton.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-    
-
-  void nonIsolatedMuonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/25");
-    m_h2 = (TH1F*)m_file->Get("simple/26");
-    m_h3 = (TH1F*)m_file->Get("simple/27");
-    m_h4 = (TH1F*)m_file->Get("simple/28");
-    m_h5 = (TH1F*)m_file->Get("simple/29");
-    m_h6 = (TH1F*)m_file->Get("simple/30");
-
-
-    TPostScript ps("nonIsolatedMuon.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-    return;
-  }
-
-  void isolatedMuonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/31");
-    m_h2 = (TH1F*)m_file->Get("simple/32");
-    m_h3 = (TH1F*)m_file->Get("simple/33");
-    m_h4 = (TH1F*)m_file->Get("simple/34");
-    m_h5 = (TH1F*)m_file->Get("simple/35");
-    m_h6 = (TH1F*)m_file->Get("simple/36");
-
-
-    TPostScript ps("isolatedMuon.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-    
-
-  void cellHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/37");
-    m_h2 = (TH1F*)m_file->Get("simple/38");
-    m_h3 = (TH1F*)m_file->Get("simple/39");
-    m_h4 = (TH1F*)m_file->Get("simple/40");
-    m_h5 = (TH1F*)m_file->Get("simple/41");
-    m_h6 = (TH1F*)m_file->Get("simple/42");
-
-    TPostScript ps("cells.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-
-  void clusterHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/43");
-    m_h2 = (TH1F*)m_file->Get("simple/44");
-    m_h3 = (TH1F*)m_file->Get("simple/45");
-    m_h4 = (TH1F*)m_file->Get("simple/46");
-    m_h5 = (TH1F*)m_file->Get("simple/47");
-    m_h6 = (TH1F*)m_file->Get("simple/48");
-
-    TPostScript ps("clusters.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void jetHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/49");
-    m_h2 = (TH1F*)m_file->Get("simple/50");
-    m_h3 = (TH1F*)m_file->Get("simple/51");
-    m_h4 = (TH1F*)m_file->Get("simple/52");
-    m_h5 = (TH1F*)m_file->Get("simple/53");
-    m_h6 = (TH1F*)m_file->Get("simple/54");
-
-
-    TPostScript ps("jets.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-  void jetBHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/55");
-    m_h2 = (TH1F*)m_file->Get("simple/56");
-    m_h3 = (TH1F*)m_file->Get("simple/57");
-    m_h4 = (TH1F*)m_file->Get("simple/58");
-    m_h5 = (TH1F*)m_file->Get("simple/59");
-    m_h6 = (TH1F*)m_file->Get("simple/60");
-
-    TPostScript ps("bjets.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-    
-
-  void eventHeaderHists1(){
-    m_h1 = (TH1F*)m_file->Get("simple/55");
-    m_h2 = (TH1F*)m_file->Get("simple/56");
-    m_h3 = (TH1F*)m_file->Get("simple/57");
-    m_h4 = (TH1F*)m_file->Get("simple/58");
-    m_h5 = (TH1F*)m_file->Get("simple/59");
-    m_h6 = (TH1F*)m_file->Get("simple/60");
-
-
-    TPostScript ps("eventHeader1.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void eventHeaderHists2(){
-    m_h1 = (TH1F*)m_file->Get("simple/61");
-    m_h2 = (TH1F*)m_file->Get("simple/62");
-    m_h3 = (TH1F*)m_file->Get("simple/63");
-    m_h4 = new TH1F;
-    m_h5 = new TH1F;
-    m_h6 = new TH1F;
-
-    TPostScript ps("eventHeader2.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-
-  }
-
-  void Draw(){
-    TCanvas* can = new TCanvas;
-    can->Clear();
-    can->Divide(3,2);
-    can->cd(1);
-    m_h1->Draw();
-    can->cd(2);
-    m_h2->Draw();
-    can->cd(3);
-    m_h3->Draw();
-    can->cd(4);
-    m_h4->Draw();
-    can->cd(5);
-    m_h5->Draw();
-    can->cd(6);
-    m_h6->Draw();
-  }
-  void Print(TCanvas* c){
-    c->Print(fn);
-  }
- 
-private:
-  TH1F* m_h1;
-  TH1F* m_h2;
-  TH1F* m_h3;
-  TH1F* m_h4;
-  TH1F* m_h5;
-  TH1F* m_h6;
-  TFile* m_file;
-  
-};
-
-class HistoRegression{
-public:
- HistoRegression(char* file, char* refFile){
-   bool initOK = true;
-   m_file =new TFile(file);
-   if (m_file->IsZombie()) {
-     initOK = false;
-   }
-
-   m_refFile = new TFile(refFile);
-   if (m_refFile->IsZombie()) {
-     initOK = false;
-   }
-
-   cout <<"initOK "<<initOK<<endl;
-   if(initOK){
-     cout<<"performing ...."<<endl;
-     this->nonIsolatedElectronHists(); 
-     this->isolatedElectronHists();
-     this->nonIsolatedPhotonHists();
-     this->isolatedPhotonHists();
-   }
- }
-
-  void nonIsolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/1");
-    m_h2 = (TH1F*)m_file->Get("simple/2");
-    m_h3 = (TH1F*)m_file->Get("simple/3");
-    m_h4 = (TH1F*)m_file->Get("simple/4");
-    m_h5 = (TH1F*)m_file->Get("simple/5");
-    m_h6 = (TH1F*)m_file->Get("simple/6");
-
-    m_refh1 = (TH1F*)m_refFile->Get("simple/1");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/2");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/3");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/4");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/5");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/6");
-    
-    TPostScript ps("nonIsolatedElectron_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-  
-  void isolatedElectronHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/7");
-    m_h2 = (TH1F*)m_file->Get("simple/8");
-    m_h3 = (TH1F*)m_file->Get("simple/9");
-    m_h4 = (TH1F*)m_file->Get("simple/10");
-    m_h5 = (TH1F*)m_file->Get("simple/11");
-    m_h6 = (TH1F*)m_file->Get("simple/12");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/7");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/8");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/9");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/10");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/11");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/12");
-    TPostScript ps("isolatedElectron_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-  }
-    
-
-  void nonIsolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/13");
-    m_h2 = (TH1F*)m_file->Get("simple/14");
-    m_h3 = (TH1F*)m_file->Get("simple/15");
-    m_h4 = (TH1F*)m_file->Get("simple/16");
-    m_h5 = (TH1F*)m_file->Get("simple/17");
-    m_h6 = (TH1F*)m_file->Get("simple/18");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/13");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/14");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/15");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/16");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/17");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/18");
-    TPostScript ps("nonIsolatedPhoton_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-    return;
-  }
-
-  void isolatedPhotonHists(){
-    m_h1 = (TH1F*)m_file->Get("simple/19");
-    m_h2 = (TH1F*)m_file->Get("simple/20");
-    m_h3 = (TH1F*)m_file->Get("simple/21");
-    m_h4 = (TH1F*)m_file->Get("simple/22");
-    m_h5 = (TH1F*)m_file->Get("simple/23");
-    m_h6 = (TH1F*)m_file->Get("simple/24");
-    m_refh1 = (TH1F*)m_refFile->Get("simple/19");
-    m_refh2 = (TH1F*)m_refFile->Get("simple/20");
-    m_refh3 = (TH1F*)m_refFile->Get("simple/21");
-    m_refh4 = (TH1F*)m_refFile->Get("simple/22");
-    m_refh5 = (TH1F*)m_refFile->Get("simple/23");
-    m_refh6 = (TH1F*)m_refFile->Get("simple/24");
-    TPostScript ps("isolatedPhoton_regression.ps", 112);
-    this->Draw();
-    ps.Close();
-
-
-  }
-
-  void setLineColor(int i){
-    gStyle->SetHistLineColor(i)
-  }
-  
-  void Draw(){
-    TCanvas* can = new TCanvas;
-    can->Clear();
-    can->Divide(3,2);
-    can->cd(1);
-    setLineColor(1);
-    m_h1->Draw();
-    setLineColor(2);
-    m_refh1->Draw("same");
-    can->cd(2);
-    setLineColor(1);
-    m_h2->Draw();
-    setLineColor(2);
-    m_refh2->Draw("same");
-    can->cd(3);
-    setLineColor(1);
-    m_h3->Draw();
-    setLineColor(2);
-    m_refh3->Draw("same");
-    can->cd(4);
-    setLineColor(1);
-    m_h4->Draw();
-    setLineColor(2);
-    m_refh4->Draw("same");
-    can->cd(5);
-    setLineColor(1);
-    m_h5->Draw();
-    setLineColor(2);
-    m_refh5->Draw("same");
-    can->cd(6);
-    setLineColor(1);
-    m_h6->Draw();
-    setLineColor(2);
-    m_refh6->Draw("same");
-  }
-
-  void Print(TCanvas* c){
-    c->Print(fn);
-  }
- 
-private:
-  TH1F* m_h1;
-  TH1F* m_h2;
-  TH1F* m_h3;
-  TH1F* m_h4;
-  TH1F* m_h5;
-  TH1F* m_h6;
-  TH1F* m_refh1;
-  TH1F* m_refh2;
-  TH1F* m_refh3;
-  TH1F* m_refh4;
-  TH1F* m_refh5;
-  TH1F* m_refh6;
-  TFile* m_file;
-  TFile* m_refFile;
-
-};
-
-int atlfast(){
-   
-  //  cout<<"in atlfast.C"<<endl;
-  AtlHists a1("histo.hbook");
-
-  //  cout<<"doing regression histograms..."<<endl;
-  HistoRegression hr("histo.hbook","refHisto.hbook");
-
-  return 0;
-}
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_herwig_default.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_herwig_default.txt
deleted file mode 100755
index 2d1d1ca328ec6876b6de508a6c04fd7c9136cd56..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_herwig_default.txt
+++ /dev/null
@@ -1,6 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Herwig_i"};
-Generator.Members = {"Herwig"};
-AtRndmGenSvc.Seeds = {"HERWIG 390020611 821000366", "HERWIG_INIT 820021 2347532"};
-Herwig.HerwigCommand = {"iproc 1499","modpdf 1", "autpdf CTEQ  "};
-
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_isajet_default.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_isajet_default.txt
deleted file mode 100755
index 530caba8c2d6481a08b048d52dffb5135a22d94b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_isajet_default.txt
+++ /dev/null
@@ -1,5 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Isajet_i"};
-Generator.Members = {"Isajet"};
-AtRndmGenSvc.Seeds = {"ISAJET 1857585203 424674947", "ISAJET_INIT 820021 2347532"};
-
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_default.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_default.txt
deleted file mode 100755
index 2f416453f6a6a50ea84c4b7650768c9c13f979ea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_default.txt
+++ /dev/null
@@ -1,4 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ele.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ele.txt
deleted file mode 100755
index caa1ad735f5d5fc0517720edfb50101371e73b7b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ele.txt
+++ /dev/null
@@ -1,46 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = {"pysubs msel 13",
-                              "pypars mstp 43 2",
-			      "pydat3 mdcy 23 1 1",
-			      "pydat3 mdcy 23 3 14",	      
-			      "pydat3 mdme 174 1 0",
-			      "pydat3 mdme 175 1 0",
-			      "pydat3 mdme 176 1 0",
-			      "pydat3 mdme 177 1 0",
-			      "pydat3 mdme 178 1 0",
-			      "pydat3 mdme 179 1 0",
-			      "pydat3 mdme 180 1 0",
-			      "pydat3 mdme 181 1 0",
-			      "pydat3 mdme 182 1 1",
-			      "pydat3 mdme 183 1 0",
-			      "pydat3 mdme 184 1 0",
-			      "pydat3 mdme 185 1 0",
-			      "pydat3 mdme 186 1 0",
-			      "pydat3 mdme 187 1 0",
-			      "pydat3 brat 174 0",
-			      "pydat3 brat 175 0",
-			      "pydat3 brat 176 0",
-			      "pydat3 brat 177 0", 
-			      "pydat3 brat 178 0",
-			      "pydat3 brat 179 0", 
-			      "pydat3 brat 180 0", 
-			      "pydat3 brat 181 0", 
-			      "pydat3 brat 182 1", 
-			      "pydat3 brat 183 0", 
-			      "pydat3 brat 184 0", 
-			      "pydat3 brat 185 0", 
-			      "pydat3 brat 186 0", 
-			      "pydat3 brat 187 0" 
-			                };
-
-
-
-
-
-
-
-
-
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_mu.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_mu.txt
deleted file mode 100755
index cf89e3eda432b6a2612d201f6c375a58f16a6149..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_mu.txt
+++ /dev/null
@@ -1,37 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = {"pysubs msel 13",
-                              "pypars mstp 43 2",
-"pydat3 mdcy 23 1 1",
-"pydat3 mdcy 23 3 14",
-"pydat3 mdme 174 1 0",
-"pydat3 mdme 175 1 0",
-"pydat3 mdme 176 1 0",
-"pydat3 mdme 177 1 0",
-"pydat3 mdme 178 1 0",
-"pydat3 mdme 179 1 0",
-"pydat3 mdme 180 1 0",
-"pydat3 mdme 181 1 0",
-"pydat3 mdme 182 1 0",
-"pydat3 mdme 183 1 0",
-"pydat3 mdme 184 1 1",
-"pydat3 mdme 185 1 0",
-"pydat3 mdme 186 1 0",
-"pydat3 mdme 187 1 0",
-"pydat3 brat 174 0",
-"pydat3 brat 175 0",
-"pydat3 brat 176 0",
-"pydat3 brat 177 0", 
-"pydat3 brat 178 0",
-"pydat3 brat 179 0", 
-"pydat3 brat 180 0", 
-"pydat3 brat 181 0", 
-"pydat3 brat 182 0", 
-"pydat3 brat 183 0", 
-"pydat3 brat 184 1", 
-"pydat3 brat 185 0",
-"pydat3 brat 186 0",
-"pydat3 brat 187 0" 
-              };
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_qcd.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_qcd.txt
deleted file mode 100755
index 2642e8b83a68ed3d3f40df33cd3e42d39a86d7ac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_qcd.txt
+++ /dev/null
@@ -1,16 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = { "pydat1 mstu 11 10"
-                              ,"pypars mstp 51 7"
-                              ,"pysubs msel 0"
-                              ,"pysubs msub 11 1"
-                              ,"pysubs msub 12 1"
-                              ,"pysubs msub 13 1"
-                              ,"pysubs msub 28 1"
-                              ,"pysubs msub 53 1"
-                              ,"pysubs msub 68 1"
-                              ,"pysubs ckin 3 17.0"
-              };
-
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_tau.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_tau.txt
deleted file mode 100755
index d66806f2d8c09e4e16de1c64e2060d0bee41d744..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_tau.txt
+++ /dev/null
@@ -1,37 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = {"pysubs msel 13",
-                              "pypars mstp 43 2",
-"pydat3 mdcy 23 1 1",
-"pydat3 mdcy 23 3 14",
-"pydat3 mdme 174 1 0",
-"pydat3 mdme 175 1 0",
-"pydat3 mdme 176 1 0",
-"pydat3 mdme 177 1 0",
-"pydat3 mdme 178 1 0",
-"pydat3 mdme 179 1 0",
-"pydat3 mdme 180 1 0",
-"pydat3 mdme 181 1 0",
-"pydat3 mdme 182 1 0",
-"pydat3 mdme 183 1 0",
-"pydat3 mdme 184 1 0",
-"pydat3 mdme 185 1 0",
-"pydat3 mdme 186 1 1",
-"pydat3 mdme 187 1 0",
-"pydat3 brat 174 0",
-"pydat3 brat 175 0",
-"pydat3 brat 176 0",
-"pydat3 brat 177 0", 
-"pydat3 brat 178 0",
-"pydat3 brat 179 0", 
-"pydat3 brat 180 0", 
-"pydat3 brat 181 0", 
-"pydat3 brat 182 0", 
-"pydat3 brat 183 0", 
-"pydat3 brat 184 0", 
-"pydat3 brat 185 0", 
-"pydat3 brat 186 1",
-"pydat3 brat 187 0"
-              };
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ttbar.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ttbar.txt
deleted file mode 100755
index af3526518d74f1cef76fb71ba67763165c484d6c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_ttbar.txt
+++ /dev/null
@@ -1,5 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = {"pysubs kfin 2 6 1"};
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_whbb.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_whbb.txt
deleted file mode 100755
index d7b9b6e7a72dc7ea3fd15142a8f4205af9ee5225..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_pythia_whbb.txt
+++ /dev/null
@@ -1,112 +0,0 @@
-#include "AthenaCommon/Atlas_Gen.UnixStandardJob.txt"
-ApplicationMgr.DLLs += { "Pythia_i"};
-Generator.Members = {"Pythia"};
-AtRndmGenSvc.Seeds = {"PYTHIA 4789899 989240512", "PYTHIA_INIT 820021 2347532"};
-Pythia.PythiaCommand = { 
-                               "pydat1 mstu 11 10"
-                              ,"pypars mstp 51 7"
-                              ,"pysubs msel 0"
-                              ,"pysubs msub 26 1"
-                              ,"pydat2 pmas 25 1 100."
-                              ,"pydat3 mdcy 25 1 1"
-                              ,"pydat3 mdme 210 1 0"
-                              ,"pydat3 mdme 211 1 0"
-                              ,"pydat3 mdme 212 1 0"
-                              ,"pydat3 mdme 213 1 0"
-                              ,"pydat3 mdme 214 1 1"
-                              ,"pydat3 mdme 215 1 0"
-                              ,"pydat3 mdme 216 1 0"
-                              ,"pydat3 mdme 217 1 0"
-                              ,"pydat3 mdme 218 1 0"
-                              ,"pydat3 mdme 219 1 0"
-                              ,"pydat3 mdme 220 1 0"
-                              ,"pydat3 mdme 221 1 0"
-                              ,"pydat3 mdme 222 1 0"
-                              ,"pydat3 mdme 223 1 0"
-                              ,"pydat3 mdme 224 1 0"
-                              ,"pydat3 mdme 225 1 0"
-                              ,"pydat3 mdme 226 1 0"
-                              ,"pydat3 mdme 227 1 0"
-                              ,"pydat3 mdme 228 1 0"
-                              ,"pydat3 mdme 229 1 0"
-                              ,"pydat3 mdme 230 1 0"
-                              ,"pydat3 mdme 231 1 0"
-                              ,"pydat3 mdme 232 1 0"
-                              ,"pydat3 mdme 233 1 0"
-                              ,"pydat3 mdme 234 1 0"
-                              ,"pydat3 mdme 235 1 0"
-                              ,"pydat3 mdme 236 1 0"
-                              ,"pydat3 mdme 237 1 0"
-                              ,"pydat3 mdme 238 1 0"
-                              ,"pydat3 mdme 239 1 0"
-                              ,"pydat3 mdme 240 1 0"
-                              ,"pydat3 mdme 241 1 0"
-                              ,"pydat3 mdme 242 1 0"
-                              ,"pydat3 mdme 243 1 0"
-                              ,"pydat3 mdme 244 1 0"
-                              ,"pydat3 mdme 245 1 0"
-                              ,"pydat3 mdme 246 1 0"
-                              ,"pydat3 mdme 247 1 0"
-                              ,"pydat3 mdme 248 1 0"
-                              ,"pydat3 mdme 249 1 0"
-                              ,"pydat3 mdme 250 1 0"
-                              ,"pydat3 mdme 251 1 0"
-                              ,"pydat3 mdme 252 1 0"
-                              ,"pydat3 mdme 253 1 0"
-                              ,"pydat3 mdme 254 1 0"
-                              ,"pydat3 mdme 255 1 0"
-                              ,"pydat3 mdme 256 1 0"
-                              ,"pydat3 mdme 257 1 0"
-                              ,"pydat3 mdme 258 1 0"
-                              ,"pydat3 mdme 259 1 0"
-                              ,"pydat3 mdme 260 1 0"
-                              ,"pydat3 mdme 261 1 0"
-                              ,"pydat3 mdme 262 1 0"
-                              ,"pydat3 mdme 263 1 0"
-                              ,"pydat3 mdme 264 1 0"
-                              ,"pydat3 mdme 265 1 0"
-                              ,"pydat3 mdme 266 1 0"
-                              ,"pydat3 mdme 267 1 0"
-                              ,"pydat3 mdme 268 1 0"
-                              ,"pydat3 mdme 269 1 0"
-                              ,"pydat3 mdme 270 1 0"
-                              ,"pydat3 mdme 271 1 0"
-                              ,"pydat3 mdme 272 1 0"
-                              ,"pydat3 mdme 273 1 0"
-                              ,"pydat3 mdme 274 1 0"
-                              ,"pydat3 mdme 275 1 0"
-                              ,"pydat3 mdme 276 1 0"
-                              ,"pydat3 mdme 277 1 0"
-                              ,"pydat3 mdme 278 1 0"
-                              ,"pydat3 mdme 279 1 0"
-                              ,"pydat3 mdme 280 1 0"
-                              ,"pydat3 mdme 281 1 0"
-                              ,"pydat3 mdme 282 1 0"
-                              ,"pydat3 mdme 283 1 0"
-                              ,"pydat3 mdme 284 1 0"
-                              ,"pydat3 mdme 285 1 0"
-                              ,"pydat3 mdme 286 1 0"
-                              ,"pydat3 mdme 287 1 0"
-                              ,"pydat3 mdme 288 1 0"
-                              ,"pydat3 mdcy 24 1 1"
-                              ,"pydat3 mdme 190 1 0"
-                              ,"pydat3 mdme 191 1 0"
-                              ,"pydat3 mdme 192 1 0"
-                              ,"pydat3 mdme 193 1 0"
-                              ,"pydat3 mdme 194 1 1"
-                              ,"pydat3 mdme 195 1 0"
-                              ,"pydat3 mdme 196 1 0"
-                              ,"pydat3 mdme 197 1 0"
-                              ,"pydat3 mdme 198 1 1"
-                              ,"pydat3 mdme 199 1 0"
-                              ,"pydat3 mdme 200 1 0"
-                              ,"pydat3 mdme 201 1 0"
-                              ,"pydat3 mdme 202 1 0"
-                              ,"pydat3 mdme 203 1 0"
-                              ,"pydat3 mdme 204 1 0"
-                              ,"pydat3 mdme 205 1 0"
-                              ,"pydat3 mdme 206 1 1"
-                              ,"pydat3 mdme 207 1 1"
-                              ,"pydat3 mdme 208 1 0"
-                              ,"pydat3 mdme 209 1 0"
-                        };
diff --git a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_zebra_default.txt b/Tools/RunTimeTester/share/old/hold_GeneratorOptions_zebra_default.txt
deleted file mode 100755
index 220c957de2052acdd39976085a25841ba4a27945..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/hold_GeneratorOptions_zebra_default.txt
+++ /dev/null
@@ -1,11 +0,0 @@
-#include "AthenaCommon/Atlas_ZebraTDR.UnixStandardJob.txt"
-//ApplicationMgr.EvtSel = "FILE ZEBRA.P";
-ApplicationMgr.EvtSel = "EventSelector";
-EventSelector.dataSetNames = {"Y01606.1"} ;
-ApplicationMgr.DLLs += { "GenzModule"};
-Generator.Members += { "GenzModule"};
-EventSelector.calos = false;                        
-EventSelector.muons = false;                        
-EventSelector.trt   = false;
-EventSelector.sct   = false;
-EventSelector.pixel = false;
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/old/jobOptionsAF.tpl b/Tools/RunTimeTester/share/old/jobOptionsAF.tpl
deleted file mode 100755
index 22f5480a2192e3470a5ca59c9f9d2b2ec5c8a52b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/jobOptionsAF.tpl
+++ /dev/null
@@ -1,76 +0,0 @@
-//##############################################################
-//
-// Job options file for running simpleTest
-//
-//==============================================================
-//
-
-#include "RunTimeTester/GeneratorOptions_!$PROCESS$!.txt"
-#include "PartPropSvc/PartPropSvc.txt"
-ApplicationMgr.DLLs += { "GaudiAlg" };	
-ApplicationMgr.DLLs += { "GaudiAud" };
-AuditorSvc.Auditors  = { "ChronoAuditor" };
-
-
-ApplicationMgr.ExtSvc += {"AtRndmGenSvc"};
-
-ApplicationMgr.DLLs += { "AtlfastAlgs","TestAtlfast"};
-//--------------------------------------------------------------
-// Event related parameters
-//--------------------------------------------------------------
-// Number of events to be processed (default is 10)
-ApplicationMgr.EvtMax = 1000;
-
-//--------------------------------------------------------------
-// Private Application Configuration options
-//--------------------------------------------------------------
-
-
-// Execution of algorithms given in sequencer
-//
-ApplicationMgr.TopAlg = {"Sequencer/TopSequence"};
-TopSequence.Members = {"Sequencer/Generator","Sequencer/Atlfast"};
-
-#include "RunTimeTester/TestAtlfastOptions.txt"
-
-//set up individual jobOptions
-ElectronMaker.DoSmearing		 = !$SMEARING$!;
-MuonMaker.DoSmearing                     = !$SMEARING$!;
-PhotonMaker.DoSmearing			 = !$SMEARING$!;
-JetMaker.DoSmearing       	         = !$SMEARING$!;
-TrackMaker.DoSmearing       	         = !$SMEARING$!;
-
-GlobalEventDataMaker.Luminosity		=  !$LUMINOSITY$!;	
-
-
-ClusterMaker.ProcessCells       = !$CLUSTERCELLS$!;
-ClusterMaker.ProcessTracks      = !$CLUSTERTRACKS$!;
-
-ClusterMaker.Strategy           = "!$CLUSTERSTRATEGY$!";
-
-//SimpleTest1
-Atlfast.members += {"TestAtlfast::SimpleTest/SimpleTest1"};
-SimpleTest1.writeRef                     = true;
-SimpleTest1.outFile                      = "./outFile_simpleTest.txt";
-SimpleTest1.refFile                      = "./refFile_outFile_simpleTest.txt";
-SimpleTest1.cmpFile                      = "./cmpFile_simpleTest.txt";
-
-
-Isajet.IsajetCommand = {"isalis !$FLAGS$!.lis"};
-MessageSvc.OutputLevel               = 3;
-
-//==============================================================
-//
-// End of job options file
-//
-//##############################################################
-
-
-
-
-
-
-
-
-
-
diff --git a/Tools/RunTimeTester/share/old/jobOptionsG.tpl b/Tools/RunTimeTester/share/old/jobOptionsG.tpl
deleted file mode 100755
index c9ff4a5e85c62b9f4f5899e68fb36fa9d5c98a43..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/old/jobOptionsG.tpl
+++ /dev/null
@@ -1,54 +0,0 @@
-//##############################################################
-//
-// Job options file for running simpleTest
-//
-//==============================================================
-//
-#include "RunTimeTester/GeneratorOptions_!$PROCESS$!.txt"
-#include "PartPropSvc/PartPropSvc.txt"
-ApplicationMgr.DLLs += { "GaudiAlg" };	
-ApplicationMgr.DLLs += { "GaudiAud" };
-AuditorSvc.Auditors  = { "ChronoAuditor" };
-
-
-ApplicationMgr.ExtSvc += {"AtRndmGenSvc"};
-
-ApplicationMgr.DLLs += { "TestAtlfast"};
-//--------------------------------------------------------------
-// Event related parameters
-//--------------------------------------------------------------
-// Number of events to be processed (default is 10)
-ApplicationMgr.EvtMax = 1000;
-
-// Execution of algorithms given in sequencer
-//
-ApplicationMgr.TopAlg = {"Sequencer/TopSequence"};
-TopSequence.Members = {"Sequencer/Generator","Sequencer/Atlfast"};
-
-//GeneratorsTest
-Atlfast.members += {"TestAtlfast::GeneratorsTest/SimpleTest1"};
-SimpleTest1.writeRef                     = true;
-SimpleTest1.outFile                      = "./outFile_Generators_!$FLAGS$!.txt";
-SimpleTest1.refFile                      = "./refFile_Generators_!$FLAGS$!.txt";
-SimpleTest1.cmpFile                      = "./cmpFile_Generators_!$FLAGS$!.txt";
-
-Isajet.IsajetCommand = {"isalis !$FLAGS$!.lis"};
-
-
-MessageSvc.OutputLevel               = 4;
-
-//==============================================================
-//
-// End of job options file
-//
-//##############################################################
-
-
-
-
-
-
-
-
-
-
diff --git a/Tools/RunTimeTester/share/rttSummaryMacro.cxx b/Tools/RunTimeTester/share/rttSummaryMacro.cxx
deleted file mode 100755
index 2397cbc9269cc117169cd01c3f0ef0bc86b28178..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/rttSummaryMacro.cxx
+++ /dev/null
@@ -1,383 +0,0 @@
-#include <fstream>
-#include <vector>
-#include <string>
-#include "TH1.h"
-#include "TCanvas.h"
-#include "TPostScript.h"
-#include "TPaveText.h"
-#include <cstdlib>
-#include <iostream>
-#include <stdlib.h>
-
-
-void printInt(int i){std::cout<<i<<" ";}
-
-class Group{
-public:
-  Group(std::ifstream& f, int nRelease,
-	const std::vector<std::string>& releases,
-	const std::vector<std::string>& times) {
-
-    // constructor reads in the number of jobs, succesful jobs
-    // tests, succesful tests, for a job group as a function of
-    // release, then make superimposed (all, succcess) hists.
-
-    int iRelease;
-    int thing;
-    f>>m_group;
-
-    for(iRelease=0; iRelease<nRelease; ++iRelease){
-      f>>thing;
-      std::cout<<"debug "<<thing<<std::endl;
-      m_jobs.push_back(thing);
-    }
-    
-    for(iRelease=0; iRelease<nRelease; ++iRelease){
-      f>>thing;
-      m_jSucc.push_back(thing);
-    }
-
-    for(iRelease=0; iRelease<nRelease; ++iRelease){
-      f>>thing;
-      m_tests.push_back(thing);
-    }
-    
-    for(iRelease=0; iRelease<nRelease; ++iRelease){
-      f>>thing;
-      m_tSucc.push_back(thing);
-    }
-
-
-    m_jobAllHist  = 0;
-    m_jobSucHist  = 0;
-    m_testAllHist = 0;
-    m_testSucHist = 0;
-    makeHists(releases, times);
-  }
-
-  void dump(int gNum){
-    
-    std::cout<<std::endl;
-    std::cout<<std::endl;
-    std::cout<<gNum<<std::endl;
-    std::cout<<m_group<<std::endl;
-    
-    std::cout<<std::endl<<"jobs"<<std::endl;
-    std::for_each(m_jobs.begin(),   m_jobs.end(), printInt);
-
-    std::cout<<std::endl<<"jSucc"<<std::endl;
-    std::for_each(m_jSucc.begin(),  m_jSucc.end(), printInt);
-
-    std::cout<<std::endl<<"tests"<<std::endl;
-    std::for_each(m_tests.begin(),  m_tests.end(), printInt);
-
-    std::cout<<std::endl<<"tSucc"<<std::endl;
-    std::for_each(m_tSucc.begin(), m_tSucc.end(), printInt);
-    
-    std::cout<<std::endl;
-    
-    }
-  TH1F* jobAllHist() {return m_jobAllHist;}
-  TH1F* jobSucHist() {return m_jobSucHist;}
-  TH1F* testAllHist(){return m_testAllHist;}
-  TH1F* testSucHist(){return m_testSucHist;}
-
-  /*
-    Dont know why this destructor causes a crash
-  ~Group(){
-    std::cout<<"Group destructor: pointers to hists"<<std::endl;
-    std::cout<<m_jobAllHist<<std::endl;
-    std::cout<<m_jobSucHist<<std::endl;
-    std::cout<<m_testAllHist<<std::endl;
-    std::cout<<m_testSucHist<<std::endl;
-    delete m_jobAllHist;
-    std::cout<<"del0"<<std::endl;
-    delete m_jobSucHist;
-    std::cout<<"del1"<<std::endl;
-    delete m_testAllHist;
-    std::cout<<"del2"<<std::endl;
-    delete m_testSucHist;
-    std::cout<<"del3"<<std::endl;
-  }
-  */
-  TPaveText* textBox(){return m_textBox;}
-private:
-  std::string      m_group;
-  std::vector<int> m_jobs;
-  std::vector<int> m_jSucc;
-  std::vector<int> m_tests;
-  std::vector<int> m_tSucc;
-  TH1F* m_jobAllHist;
-  TH1F* m_jobSucHist;
-  TH1F* m_testAllHist;
-  TH1F* m_testSucHist;
-  
-  TPaveText* m_textBox;
-
-  void makeHists(std::vector<std::string> releases,
-		 std::vector<std::string> times){
-    std::cout<<"m_group1 "<<m_group<<std::endl;
-    std::string label;
-
-
-
-    label = "All Jobs "+m_group;
-    m_jobAllHist = new TH1F(label.c_str(), m_group.c_str(), 7, 0., 7.);
-    
-    label = "Successful Jobs "+m_group;
-    m_jobSucHist = new TH1F(label.c_str(), m_group.c_str(), 7, 0., 7.);
-    
-    label = "All Tests "+m_group;
-    m_testAllHist = new TH1F(label.c_str(), m_group.c_str(), 7, 0., 7.);
-    
-    label = "Successful Tests "+m_group;
-    m_testSucHist = new TH1F(label.c_str(), m_group.c_str(), 7, 0., 7.);
-    
-
-    
-    const char* relLabels[7] = {" ", " ", " ", " ", " ", " ", " "};
-
-    m_textBox = new TPaveText(0, 0, 1, 1);
-			      
-    std::vector<std::string>::const_iterator releaseIter = releases.begin();
-    std::vector<std::string>::const_iterator timeIter    = times.begin();
-    std::string text;
-    for(; releaseIter != releases.end(); ++releaseIter, ++timeIter){
-      text = "";
-      text += *releaseIter;
-      text += ": ";
-      text += *timeIter;
-      m_textBox -> AddText(text.c_str());
-    }
-    m_textBox ->SetTextSize(0.07);
-
-    releaseIter = releases.begin();
-    int ind = 0;
-    // the x-axis labels are formed from the release name and the time 
-    // the RTT was run
-    for(; releaseIter != releases.end(); ++releaseIter){
-      relLabels[ind++]=(*releaseIter).c_str();
-      if (ind == 7) break;
-    }
-    
-    m_jobAllHist->SetBit(TH1::kCanRebin);
-    m_jobSucHist->SetBit(TH1::kCanRebin);
-    m_testAllHist->SetBit(TH1::kCanRebin);
-    m_testSucHist->SetBit(TH1::kCanRebin);
-
-    int bin = 0;
-    std::vector<int>::const_iterator iter;
-    for(iter = m_jobs.begin(); iter != m_jobs.end(); ++iter){
-      m_jobAllHist->Fill(relLabels[bin++], double(*iter));
-    }
-    
-    bin = 0;
-    for(iter = m_jSucc.begin(); iter != m_jSucc.end(); ++iter){
-      m_jobSucHist->Fill(relLabels[bin++], double(*iter));
-    }
-
-    bin = 0;
-    for(iter = m_tests.begin(); iter != m_tests.end(); ++iter){
-      m_testAllHist->Fill(relLabels[bin++], double(*iter));
-     }
-
-    bin = 0;
-    for(iter = m_tSucc.begin(); iter != m_tSucc.end(); ++iter){
-      m_testSucHist->Fill(relLabels[bin++], double(*iter));
-     }
-  }
-};
-
-/*  
-class PrintGroup{
-public:
-  PrintGroup(){}
-  void operator()(Group g){}
-};
-*/    
-
-class PrintGroup{
-public:
-  PrintGroup(){
-    m_gNum = 0;
-  }
-  void operator()(Group g){
-    g.dump(m_gNum++);
-  }
-private:
-  int m_gNum;
-};
-
-class Plot{
-public:
-  Plot(){} // root seems to want this.....
-
-  Plot(std::string fn){
-
-    int maxPadSide   = 2;
-
-    m_title      = fn.c_str();
-    m_pad        = 0;
-    m_lot        = 0;
-    m_fn         = fn;
-    m_maxPad     = maxPadSide*maxPadSide;
-    m_maxPadSide = maxPadSide;
-    m_can        = new TCanvas(m_title, m_title, 1);
-    m_can        -> Divide(maxPadSide, maxPadSide);
-    m_can->Print((m_fn+"[").c_str());
-
-  }
-
-  void plot(TH1* all, TH1* succ, TPaveText* text){
-
-
-    if(++m_pad == m_maxPad){
-
-      std::cout<<"wrting canvas with current pad = "<<m_pad<<std::endl;
-      m_can->Print( (m_fn+"(").c_str() );
-      m_can->Clear();
-      m_can-> Divide(m_maxPadSide, m_maxPadSide);
-      m_pad = 1;
-    }
-    
-    m_can->cd(m_pad);
-    all->SetFillColor(1);
-    all->SetStats(kFALSE);
-    all->Draw();
-    succ->SetFillColor(3);
-    succ->SetStats(kFALSE);
-    succ->Draw("same");
-
-    // draw the text box giving the date and time for each release
-    // to th last pad
-    if (++m_pad == m_maxPad){
-      m_can->cd(m_pad);
-      text->Draw();
-    }
-    --m_pad;
-
-  }
-  void close(){
-    if (m_pad != m_maxPad){
-      std::cout<<"writing last canvas with current pad = "<<m_pad<<std::endl;
-      m_can->Print( (m_fn+"(").c_str() );
-    }
-    m_can->Print( (m_fn+"]").c_str());
-  }
-
-  void operator()(Group& g){}
-
-private:
-  TCanvas* m_can;
-  int m_maxPad;
-  int m_maxPadSide;
-  int m_pad;
-  int m_lot;
-  const char* m_title;
-  std::string m_fn;
-};
-
-class PlotJobs: public Plot{
-  // function object to plot all jobs superimposed by succesful jobs
-public:
-  PlotJobs(std::string fn): Plot(fn+ ":Jobs.ps"){}
-  
-  void operator()(Group& g){
-    this->plot(g.jobAllHist(), g.jobSucHist(), g.textBox());
-  }
-};
-
-class PlotTests: public Plot{
-  // function object to plot all tests superimposed by succesful tests
-public:
-  PlotTests(std::string fn): Plot(fn+":Tests.ps"){}
-  void operator()(Group& g){
-    this->plot(g.testAllHist(), g.testSucHist(), g.textBox());
-  }
-};
-
-
-int doit(std::string fileName){
-
-  const char* fn = fileName.c_str();
-  std::cout<<"Summary File = "<<fn<<std::endl;
-  std::ifstream from(fn);
-  if (!from) {
-    std::cout<<"cannot open "<<fn<<std::endl;
-    return 1;
-  }
-
-  int nRelease;
-  from >> nRelease;
-
-
-  std::vector<std::string> releases;
-  for (int release =0; release<nRelease; ++release){
-    std::string rel;
-    from >> rel;
-    std::cout<<"release in = "<<rel<<std::endl;
-    releases.push_back(rel);
-  }
-
-  std::vector<std::string> times;
-  for (int release =0; release<nRelease; ++release){
-    std::string time;
-    from >> time;
-    std::cout<<"release time = "<<time<<std::endl;
-    times.push_back(time);
-  }
-
-  int nGroups;
-  from>>nGroups;
-  std::vector<Group> groups;
-  for(int iGroup = 0; iGroup<nGroups; ++iGroup)
-    {groups.push_back(Group(from, nRelease, releases, times));}
-  
-  from.close();
-  
-  std::for_each(groups.begin(), groups.end(), PrintGroup());  
-  
-  PlotJobs plotJobs(fileName);
-  plotJobs = std::for_each(groups.begin(), groups.end(), plotJobs);
-  plotJobs.close();
-
-  PlotTests plotTests(fileName);
-  plotTests = std::for_each(groups.begin(), groups.end(), plotTests);
-  plotTests.close();
-  
-  
-  
-  return 0;
-}
-int rttSummaryMacro(){
-  // read in a list of files to process.
-  // each file contains information for a number of releases.
-  // for the information is only available for nightlies.
-  // The information is written out by the RTT for each run, then
-  // collected and massages by RTTSumFromXML.py
-
-  std::cout<<"hello1"<<std::endl;
-  const char* fn = "./rttSummaryList.txt";
-  std::ifstream listIn(fn);
-  std::vector<std::string> summaryFiles;
-  std::string summaryFile;
-  while (listIn>>summaryFile) summaryFiles.push_back(summaryFile);
- 
-
-  if (summaryFiles.size() == 0){
-    std::cout<<"No summary files found"<<std::endl;
-    return 0;
-  }
-  std::for_each(summaryFiles.begin(), summaryFiles.end(), doit);
-  std::cout<<"hello2"<<std::endl;
-  
-  return 0;
-}
-/*
-int rttSummaryMacro(){
-  std::cout<<"hello"<<std::endl;
-  return 0;
-}
-*/
-
-
diff --git a/Tools/RunTimeTester/share/testHist.cxx b/Tools/RunTimeTester/share/testHist.cxx
deleted file mode 100755
index 6b4a8b49385edde63197a9c64449c1378366bb32..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/testHist.cxx
+++ /dev/null
@@ -1,35 +0,0 @@
-#include "TPostScript.h"
-#include "TH1F.h"
-#include "TFile.h"
-
-int testHist(){
-  TH1F* h1 = new TH1F("h1", "TestHist", 7, 0.., 7.);
-  h1->Fill(0, 0.);
-  h1->Fill(1, 1.);
-  h1->Fill(2, 2.);
-  h1->Fill(3, 3.);
-  h1->Fill(4, 2.);
-  h1->Fill(5, 1.);
-  h1->Fill(6, 0.);
-
-  // h1->Draw();
-  
-
-  TFile* f1 = new TFile("testHist.root", "RECREATE");
-  h1->Write();
-  f1->Close();
-
-
-  //TFile* ff1 = new TFile("refFile_testHist.root");
-  //TH1F* g1 = (TH1F*)ff1->Get("h1");
-
-  //g1->SetLineColor(kRed);
-
-  TPostScript ps("RTTtest.ps",112);
-  h1->Draw();
-  //g1->Scale(0.9);
-  //g1->Draw("SAME");
-  ps.Close();
-  
-
-}
diff --git a/Tools/RunTimeTester/share/web/RTT.dtd b/Tools/RunTimeTester/share/web/RTT.dtd
deleted file mode 100755
index 1c5841b1e13e3e033ce8228358b25e653024e758..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/RTT.dtd
+++ /dev/null
@@ -1,85 +0,0 @@
-<!ELEMENT rttconfig (mode,                release,             refRelease,  
-                     cmtConfig,           branch,              runType,
-                     topProject,          otherProject?,       distArea,
-                     workBasePath,        resultsBasePath,     resultsBaseURL,
-                     site,                dataSetCatalog,      jobGroupConfig,    
-                     pathToVetoFile?,     dCubeCfgFile?,    
-                     refDB?,              moniDB?,             confFileSrc, 
-                     packageAccept*,      packageVeto*,        webpageBranchVeto*,
-                     testMode?,           maxTime?,            versionDir?,  
-                     libToolsLoc?,        confFileSuffix?,
-                     rttKitTest?,         releaseType?,        rttPilotJobs?,
-                     atnPilotJobs?,       concurrent?,         disableMailTo?,    devRun?, 
-                     deleteNonKeepFiles?, bigFiles?)>
-    
-   <!ELEMENT mode                   (#PCDATA)>
-   <!ELEMENT release                (#PCDATA)>
-   <!ELEMENT refRelease             (#PCDATA)>
-   <!ELEMENT cmtConfig              (#PCDATA)>
-   <!ELEMENT branch                 (#PCDATA)>
-   <!ELEMENT runType                (releaseRun | kitRun)>
-   <!ELEMENT topProject             (#PCDATA)>  
-   <!ELEMENT otherProject           (#PCDATA)>  
-   <!ELEMENT distArea               (#PCDATA)>
-   <!ELEMENT workBasePath           (#PCDATA)>
-   <!ELEMENT resultsBasePath        (#PCDATA)>
-   <!ELEMENT resultsBaseURL         (#PCDATA)>
-   <!ELEMENT site                   (#PCDATA)>
-   <!ELEMENT dataSetCatalog         (#PCDATA)>
-   <!ELEMENT jobGroupConfig         (#PCDATA)>
-   <!ELEMENT pathToVetoFile         (#PCDATA)>
-   <!ELEMENT dCubeCfgFile           (#PCDATA)>
-   <!ELEMENT confFileSrc            (  (rttProvides,  
-                                         localConfFiles?, 
-                                         installedConfFiles?) |
-                                       (rttProvides?, 
-                                         localConfFiles,  
-                                         installedConfFiles?) |
-                                       (rttProvides?, 
-                                         localConfFiles?, 
-                                         installedConfFiles) )>
-
-   <!ELEMENT releaseRun             EMPTY>
-   <!ELEMENT kitRun                 (installArea)>
-   <!ELEMENT installArea            (#PCDATA)>
-   <!ELEMENT rttProvides            (packageConfFile+)>
-      <!ELEMENT packageConfFile     (package, pathToFile, cmtpath, version,
-                                     installArea, shareArea)>
-         <!ELEMENT package          (#PCDATA)>
-         <!ELEMENT pathToFile       (#PCDATA)>
-         <!ELEMENT cmtpath          (#PCDATA)>
-         <!ELEMENT version          (#PCDATA)>
-         <!ELEMENT shareArea        (#PCDATA)>
-   <!ELEMENT localConfFiles         (userHomeCmtDir, testProjectCmtDir)>
-   <!ELEMENT installedConfFiles     EMPTY>
-   <!ELEMENT refDB                  (file, enable)>
-      <!ELEMENT file                (#PCDATA)>
-      <!ELEMENT enable              (#PCDATA)>
-   <!ELEMENT moniDB                 (file, enable)>
-   <!ELEMENT packageAccept          (#PCDATA)>
-   <!ELEMENT packageVeto            (#PCDATA)>
-   <!ELEMENT webpageBranchVeto      (#PCDATA)>
-   <!ELEMENT testMode               (#PCDATA)>
-   <!ELEMENT maxTime                ((hours, minutes?) | minutes)>
-      <!ELEMENT hours               (#PCDATA)>
-      <!ELEMENT minutes             (#PCDATA)>
-   <!ELEMENT versionDir             (#PCDATA)>
-   <!ELEMENT libToolsLoc            (#PCDATA)>
-   <!ELEMENT confFileSuffix         (#PCDATA)>
-   <!ELEMENT rttKitTest             EMPTY>   
-   <!ELEMENT releaseType            (#PCDATA)> 
-   <!ELEMENT rttPilotJobs               EMPTY>
-   <!ELEMENT atnPilotJobs               EMPTY>
-   <!ELEMENT userHomeCmtDir         (#PCDATA)> 
-   <!ELEMENT testProjectCmtDir      (#PCDATA)> 
-  <!ELEMENT concurrent             (wpcServer)>
-     <!ELEMENT wpcServer            (#PCDATA)>
-   <!ELEMENT disableMailTo              EMPTY>
-   <!ELEMENT devRun                     EMPTY>
-   <!ELEMENT deleteNonKeepFiles         EMPTY>
-   <!ELEMENT bigFiles               (bigFilesArea, bigFilesSize)>
-   <!ELEMENT bigFilesArea           (#PCDATA)>
-   <!ELEMENT bigFilesSize           (#PCDATA)>
-
-
-  
diff --git a/Tools/RunTimeTester/share/web/RTTpage1_LinksAtTopOfPage.js b/Tools/RunTimeTester/share/web/RTTpage1_LinksAtTopOfPage.js
deleted file mode 100644
index b44608f021207d68179a67c996a94ea06b8efe63..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/RTTpage1_LinksAtTopOfPage.js
+++ /dev/null
@@ -1,49 +0,0 @@
-
-/*********************/
-/***** Base URLs *****/
-/*********************/
-
-uclHEP    = 'http://www.hep.ucl.ac.uk/atlas/AtlasTesting/';
-cernRTT   = 'http://cern.ch/atlasrtt/';
-atlasComp = 'http://atlas-computing.web.cern.ch/atlas-computing/';
-
-/*********************/
-/***** Style CSS *****/
-/*********************/
-
-styleClass  = '<style>';
-styleClass += 'a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}';
-styleClass += 'a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}';
-styleClass += '</style>';
-
-/*********************/
-/***** The links *****/
-/*********************/
-
-link00 = '<a class="pageLink" href="page0.php">All RTT results</a>';
-link0 = '<a class="pageLink" href="weeklyTable/weekly.php">Summary of the week</a>';
-link1 = '<a class="pageLink" href="https://savannah.cern.ch/bugs/?group=atlas-infra">Report RTT bugs</a>';
-link2 = '<a class="pageLink" href="' + uclHEP + '">RTT documentation</a>';
-link3 = '<a class="pageLink" href="' + atlasComp + 'projects/releases/releases.php">Release Status</a>';
-link4 = '<a class="pageLink" href="' + atlasComp +'links/distDirectory/nightlies/global/">NICOS</a>';
-link5 = '<a class="pageLink" href="' + cernRTT + 'DatasetManager.report">Dataset Manager Report</a>';
-link6 = '<a class="pageLink" href="' + uclHEP + '../RTTmonolithic/RTTpage1.html">Old RTT monolithic results</a>';
-link7 = '<a class="pageLink" href="' + uclHEP + 'DTD/unifiedTestConfiguration.dtd">Package DTD</a>';
-link8 = '<a class="pageLink" href="' + uclHEP + 'DTD/RTT.dtd">Top level DTD</a>';
-
-
-document.write(
-   styleClass
- + '<div align="center">'
- + link00 + ' | '
- + link0 + ' | '
- + link1 + ' | '
- + link2 + ' | '
- + link3 + ' | '
- + link4 + ' | '
- + link5 + ' | '
- + link6 + ' | '
- + link7 + ' | '
- + link8
- + '</div><br />' 
-);
diff --git a/Tools/RunTimeTester/share/web/effects.js b/Tools/RunTimeTester/share/web/effects.js
deleted file mode 100644
index 06f59b47698e61d0f4c63bd8a5147147d0217b04..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/effects.js
+++ /dev/null
@@ -1,1090 +0,0 @@
-// script.aculo.us effects.js v1.7.0, Fri Jan 19 19:16:36 CET 2007
-
-// Copyright (c) 2005, 2006 Thomas Fuchs (http://script.aculo.us, http://mir.aculo.us)
-// Contributors:
-//  Justin Palmer (http://encytemedia.com/)
-//  Mark Pilgrim (http://diveintomark.org/)
-//  Martin Bialasinki
-// 
-// script.aculo.us is freely distributable under the terms of an MIT-style license.
-// For details, see the script.aculo.us web site: http://script.aculo.us/ 
-
-// converts rgb() and #xxx to #xxxxxx format,  
-// returns self (or first argument) if not convertable  
-String.prototype.parseColor = function() {  
-  var color = '#';
-  if(this.slice(0,4) == 'rgb(') {  
-    var cols = this.slice(4,this.length-1).split(',');  
-    var i=0; do { color += parseInt(cols[i]).toColorPart() } while (++i<3);  
-  } else {  
-    if(this.slice(0,1) == '#') {  
-      if(this.length==4) for(var i=1;i<4;i++) color += (this.charAt(i) + this.charAt(i)).toLowerCase();  
-      if(this.length==7) color = this.toLowerCase();  
-    }  
-  }  
-  return(color.length==7 ? color : (arguments[0] || this));  
-}
-
-/*--------------------------------------------------------------------------*/
-
-Element.collectTextNodes = function(element) {  
-  return $A($(element).childNodes).collect( function(node) {
-    return (node.nodeType==3 ? node.nodeValue : 
-      (node.hasChildNodes() ? Element.collectTextNodes(node) : ''));
-  }).flatten().join('');
-}
-
-Element.collectTextNodesIgnoreClass = function(element, className) {  
-  return $A($(element).childNodes).collect( function(node) {
-    return (node.nodeType==3 ? node.nodeValue : 
-      ((node.hasChildNodes() && !Element.hasClassName(node,className)) ? 
-        Element.collectTextNodesIgnoreClass(node, className) : ''));
-  }).flatten().join('');
-}
-
-Element.setContentZoom = function(element, percent) {
-  element = $(element);  
-  element.setStyle({fontSize: (percent/100) + 'em'});   
-  if(navigator.appVersion.indexOf('AppleWebKit')>0) window.scrollBy(0,0);
-  return element;
-}
-
-Element.getOpacity = function(element){
-  return $(element).getStyle('opacity');
-}
-
-Element.setOpacity = function(element, value){
-  return $(element).setStyle({opacity:value});
-}
-
-Element.getInlineOpacity = function(element){
-  return $(element).style.opacity || '';
-}
-
-Element.forceRerendering = function(element) {
-  try {
-    element = $(element);
-    var n = document.createTextNode(' ');
-    element.appendChild(n);
-    element.removeChild(n);
-  } catch(e) { }
-};
-
-/*--------------------------------------------------------------------------*/
-
-Array.prototype.call = function() {
-  var args = arguments;
-  this.each(function(f){ f.apply(this, args) });
-}
-
-/*--------------------------------------------------------------------------*/
-
-var Effect = {
-  _elementDoesNotExistError: {
-    name: 'ElementDoesNotExistError',
-    message: 'The specified DOM element does not exist, but is required for this effect to operate'
-  },
-  tagifyText: function(element) {
-    if(typeof Builder == 'undefined')
-      throw("Effect.tagifyText requires including script.aculo.us' builder.js library");
-      
-    var tagifyStyle = 'position:relative';
-    if(/MSIE/.test(navigator.userAgent) && !window.opera) tagifyStyle += ';zoom:1';
-    
-    element = $(element);
-    $A(element.childNodes).each( function(child) {
-      if(child.nodeType==3) {
-        child.nodeValue.toArray().each( function(character) {
-          element.insertBefore(
-            Builder.node('span',{style: tagifyStyle},
-              character == ' ' ? String.fromCharCode(160) : character), 
-              child);
-        });
-        Element.remove(child);
-      }
-    });
-  },
-  multiple: function(element, effect) {
-    var elements;
-    if(((typeof element == 'object') || 
-        (typeof element == 'function')) && 
-       (element.length))
-      elements = element;
-    else
-      elements = $(element).childNodes;
-      
-    var options = Object.extend({
-      speed: 0.1,
-      delay: 0.0
-    }, arguments[2] || {});
-    var masterDelay = options.delay;
-
-    $A(elements).each( function(element, index) {
-      new effect(element, Object.extend(options, { delay: index * options.speed + masterDelay }));
-    });
-  },
-  PAIRS: {
-    'slide':  ['SlideDown','SlideUp'],
-    'blind':  ['BlindDown','BlindUp'],
-    'appear': ['Appear','Fade']
-  },
-  toggle: function(element, effect) {
-    element = $(element);
-    effect = (effect || 'appear').toLowerCase();
-    var options = Object.extend({
-      queue: { position:'end', scope:(element.id || 'global'), limit: 1 }
-    }, arguments[2] || {});
-    Effect[element.visible() ? 
-      Effect.PAIRS[effect][1] : Effect.PAIRS[effect][0]](element, options);
-  }
-};
-
-var Effect2 = Effect; // deprecated
-
-/* ------------- transitions ------------- */
-
-Effect.Transitions = {
-  linear: Prototype.K,
-  sinoidal: function(pos) {
-    return (-Math.cos(pos*Math.PI)/2) + 0.5;
-  },
-  reverse: function(pos) {
-    return 1-pos;
-  },
-  flicker: function(pos) {
-    return ((-Math.cos(pos*Math.PI)/4) + 0.75) + Math.random()/4;
-  },
-  wobble: function(pos) {
-    return (-Math.cos(pos*Math.PI*(9*pos))/2) + 0.5;
-  },
-  pulse: function(pos, pulses) { 
-    pulses = pulses || 5; 
-    return (
-      Math.round((pos % (1/pulses)) * pulses) == 0 ? 
-            ((pos * pulses * 2) - Math.floor(pos * pulses * 2)) : 
-        1 - ((pos * pulses * 2) - Math.floor(pos * pulses * 2))
-      );
-  },
-  none: function(pos) {
-    return 0;
-  },
-  full: function(pos) {
-    return 1;
-  }
-};
-
-/* ------------- core effects ------------- */
-
-Effect.ScopedQueue = Class.create();
-Object.extend(Object.extend(Effect.ScopedQueue.prototype, Enumerable), {
-  initialize: function() {
-    this.effects  = [];
-    this.interval = null;
-  },
-  _each: function(iterator) {
-    this.effects._each(iterator);
-  },
-  add: function(effect) {
-    var timestamp = new Date().getTime();
-    
-    var position = (typeof effect.options.queue == 'string') ? 
-      effect.options.queue : effect.options.queue.position;
-    
-    switch(position) {
-      case 'front':
-        // move unstarted effects after this effect  
-        this.effects.findAll(function(e){ return e.state=='idle' }).each( function(e) {
-            e.startOn  += effect.finishOn;
-            e.finishOn += effect.finishOn;
-          });
-        break;
-      case 'with-last':
-        timestamp = this.effects.pluck('startOn').max() || timestamp;
-        break;
-      case 'end':
-        // start effect after last queued effect has finished
-        timestamp = this.effects.pluck('finishOn').max() || timestamp;
-        break;
-    }
-    
-    effect.startOn  += timestamp;
-    effect.finishOn += timestamp;
-
-    if(!effect.options.queue.limit || (this.effects.length < effect.options.queue.limit))
-      this.effects.push(effect);
-    
-    if(!this.interval) 
-      this.interval = setInterval(this.loop.bind(this), 15);
-  },
-  remove: function(effect) {
-    this.effects = this.effects.reject(function(e) { return e==effect });
-    if(this.effects.length == 0) {
-      clearInterval(this.interval);
-      this.interval = null;
-    }
-  },
-  loop: function() {
-    var timePos = new Date().getTime();
-    for(var i=0, len=this.effects.length;i<len;i++) 
-      if(this.effects[i]) this.effects[i].loop(timePos);
-  }
-});
-
-Effect.Queues = {
-  instances: $H(),
-  get: function(queueName) {
-    if(typeof queueName != 'string') return queueName;
-    
-    if(!this.instances[queueName])
-      this.instances[queueName] = new Effect.ScopedQueue();
-      
-    return this.instances[queueName];
-  }
-}
-Effect.Queue = Effect.Queues.get('global');
-
-Effect.DefaultOptions = {
-  transition: Effect.Transitions.sinoidal,
-  duration:   1.0,   // seconds
-  fps:        60.0,  // max. 60fps due to Effect.Queue implementation
-  sync:       false, // true for combining
-  from:       0.0,
-  to:         1.0,
-  delay:      0.0,
-  queue:      'parallel'
-}
-
-Effect.Base = function() {};
-Effect.Base.prototype = {
-  position: null,
-  start: function(options) {
-    this.options      = Object.extend(Object.extend({},Effect.DefaultOptions), options || {});
-    this.currentFrame = 0;
-    this.state        = 'idle';
-    this.startOn      = this.options.delay*1000;
-    this.finishOn     = this.startOn + (this.options.duration*1000);
-    this.event('beforeStart');
-    if(!this.options.sync)
-      Effect.Queues.get(typeof this.options.queue == 'string' ? 
-        'global' : this.options.queue.scope).add(this);
-  },
-  loop: function(timePos) {
-    if(timePos >= this.startOn) {
-      if(timePos >= this.finishOn) {
-        this.render(1.0);
-        this.cancel();
-        this.event('beforeFinish');
-        if(this.finish) this.finish(); 
-        this.event('afterFinish');
-        return;  
-      }
-      var pos   = (timePos - this.startOn) / (this.finishOn - this.startOn);
-      var frame = Math.round(pos * this.options.fps * this.options.duration);
-      if(frame > this.currentFrame) {
-        this.render(pos);
-        this.currentFrame = frame;
-      }
-    }
-  },
-  render: function(pos) {
-    if(this.state == 'idle') {
-      this.state = 'running';
-      this.event('beforeSetup');
-      if(this.setup) this.setup();
-      this.event('afterSetup');
-    }
-    if(this.state == 'running') {
-      if(this.options.transition) pos = this.options.transition(pos);
-      pos *= (this.options.to-this.options.from);
-      pos += this.options.from;
-      this.position = pos;
-      this.event('beforeUpdate');
-      if(this.update) this.update(pos);
-      this.event('afterUpdate');
-    }
-  },
-  cancel: function() {
-    if(!this.options.sync)
-      Effect.Queues.get(typeof this.options.queue == 'string' ? 
-        'global' : this.options.queue.scope).remove(this);
-    this.state = 'finished';
-  },
-  event: function(eventName) {
-    if(this.options[eventName + 'Internal']) this.options[eventName + 'Internal'](this);
-    if(this.options[eventName]) this.options[eventName](this);
-  },
-  inspect: function() {
-    var data = $H();
-    for(property in this)
-      if(typeof this[property] != 'function') data[property] = this[property];
-    return '#<Effect:' + data.inspect() + ',options:' + $H(this.options).inspect() + '>';
-  }
-}
-
-Effect.Parallel = Class.create();
-Object.extend(Object.extend(Effect.Parallel.prototype, Effect.Base.prototype), {
-  initialize: function(effects) {
-    this.effects = effects || [];
-    this.start(arguments[1]);
-  },
-  update: function(position) {
-    this.effects.invoke('render', position);
-  },
-  finish: function(position) {
-    this.effects.each( function(effect) {
-      effect.render(1.0);
-      effect.cancel();
-      effect.event('beforeFinish');
-      if(effect.finish) effect.finish(position);
-      effect.event('afterFinish');
-    });
-  }
-});
-
-Effect.Event = Class.create();
-Object.extend(Object.extend(Effect.Event.prototype, Effect.Base.prototype), {
-  initialize: function() {
-    var options = Object.extend({
-      duration: 0
-    }, arguments[0] || {});
-    this.start(options);
-  },
-  update: Prototype.emptyFunction
-});
-
-Effect.Opacity = Class.create();
-Object.extend(Object.extend(Effect.Opacity.prototype, Effect.Base.prototype), {
-  initialize: function(element) {
-    this.element = $(element);
-    if(!this.element) throw(Effect._elementDoesNotExistError);
-    // make this work on IE on elements without 'layout'
-    if(/MSIE/.test(navigator.userAgent) && !window.opera && (!this.element.currentStyle.hasLayout))
-      this.element.setStyle({zoom: 1});
-    var options = Object.extend({
-      from: this.element.getOpacity() || 0.0,
-      to:   1.0
-    }, arguments[1] || {});
-    this.start(options);
-  },
-  update: function(position) {
-    this.element.setOpacity(position);
-  }
-});
-
-Effect.Move = Class.create();
-Object.extend(Object.extend(Effect.Move.prototype, Effect.Base.prototype), {
-  initialize: function(element) {
-    this.element = $(element);
-    if(!this.element) throw(Effect._elementDoesNotExistError);
-    var options = Object.extend({
-      x:    0,
-      y:    0,
-      mode: 'relative'
-    }, arguments[1] || {});
-    this.start(options);
-  },
-  setup: function() {
-    // Bug in Opera: Opera returns the "real" position of a static element or
-    // relative element that does not have top/left explicitly set.
-    // ==> Always set top and left for position relative elements in your stylesheets 
-    // (to 0 if you do not need them) 
-    this.element.makePositioned();
-    this.originalLeft = parseFloat(this.element.getStyle('left') || '0');
-    this.originalTop  = parseFloat(this.element.getStyle('top')  || '0');
-    if(this.options.mode == 'absolute') {
-      // absolute movement, so we need to calc deltaX and deltaY
-      this.options.x = this.options.x - this.originalLeft;
-      this.options.y = this.options.y - this.originalTop;
-    }
-  },
-  update: function(position) {
-    this.element.setStyle({
-      left: Math.round(this.options.x  * position + this.originalLeft) + 'px',
-      top:  Math.round(this.options.y  * position + this.originalTop)  + 'px'
-    });
-  }
-});
-
-// for backwards compatibility
-Effect.MoveBy = function(element, toTop, toLeft) {
-  return new Effect.Move(element, 
-    Object.extend({ x: toLeft, y: toTop }, arguments[3] || {}));
-};
-
-Effect.Scale = Class.create();
-Object.extend(Object.extend(Effect.Scale.prototype, Effect.Base.prototype), {
-  initialize: function(element, percent) {
-    this.element = $(element);
-    if(!this.element) throw(Effect._elementDoesNotExistError);
-    var options = Object.extend({
-      scaleX: true,
-      scaleY: true,
-      scaleContent: true,
-      scaleFromCenter: false,
-      scaleMode: 'box',        // 'box' or 'contents' or {} with provided values
-      scaleFrom: 100.0,
-      scaleTo:   percent
-    }, arguments[2] || {});
-    this.start(options);
-  },
-  setup: function() {
-    this.restoreAfterFinish = this.options.restoreAfterFinish || false;
-    this.elementPositioning = this.element.getStyle('position');
-    
-    this.originalStyle = {};
-    ['top','left','width','height','fontSize'].each( function(k) {
-      this.originalStyle[k] = this.element.style[k];
-    }.bind(this));
-      
-    this.originalTop  = this.element.offsetTop;
-    this.originalLeft = this.element.offsetLeft;
-    
-    var fontSize = this.element.getStyle('font-size') || '100%';
-    ['em','px','%','pt'].each( function(fontSizeType) {
-      if(fontSize.indexOf(fontSizeType)>0) {
-        this.fontSize     = parseFloat(fontSize);
-        this.fontSizeType = fontSizeType;
-      }
-    }.bind(this));
-    
-    this.factor = (this.options.scaleTo - this.options.scaleFrom)/100;
-    
-    this.dims = null;
-    if(this.options.scaleMode=='box')
-      this.dims = [this.element.offsetHeight, this.element.offsetWidth];
-    if(/^content/.test(this.options.scaleMode))
-      this.dims = [this.element.scrollHeight, this.element.scrollWidth];
-    if(!this.dims)
-      this.dims = [this.options.scaleMode.originalHeight,
-                   this.options.scaleMode.originalWidth];
-  },
-  update: function(position) {
-    var currentScale = (this.options.scaleFrom/100.0) + (this.factor * position);
-    if(this.options.scaleContent && this.fontSize)
-      this.element.setStyle({fontSize: this.fontSize * currentScale + this.fontSizeType });
-    this.setDimensions(this.dims[0] * currentScale, this.dims[1] * currentScale);
-  },
-  finish: function(position) {
-    if(this.restoreAfterFinish) this.element.setStyle(this.originalStyle);
-  },
-  setDimensions: function(height, width) {
-    var d = {};
-    if(this.options.scaleX) d.width = Math.round(width) + 'px';
-    if(this.options.scaleY) d.height = Math.round(height) + 'px';
-    if(this.options.scaleFromCenter) {
-      var topd  = (height - this.dims[0])/2;
-      var leftd = (width  - this.dims[1])/2;
-      if(this.elementPositioning == 'absolute') {
-        if(this.options.scaleY) d.top = this.originalTop-topd + 'px';
-        if(this.options.scaleX) d.left = this.originalLeft-leftd + 'px';
-      } else {
-        if(this.options.scaleY) d.top = -topd + 'px';
-        if(this.options.scaleX) d.left = -leftd + 'px';
-      }
-    }
-    this.element.setStyle(d);
-  }
-});
-
-Effect.Highlight = Class.create();
-Object.extend(Object.extend(Effect.Highlight.prototype, Effect.Base.prototype), {
-  initialize: function(element) {
-    this.element = $(element);
-    if(!this.element) throw(Effect._elementDoesNotExistError);
-    var options = Object.extend({ startcolor: '#ffff99' }, arguments[1] || {});
-    this.start(options);
-  },
-  setup: function() {
-    // Prevent executing on elements not in the layout flow
-    if(this.element.getStyle('display')=='none') { this.cancel(); return; }
-    // Disable background image during the effect
-    this.oldStyle = {};
-    if (!this.options.keepBackgroundImage) {
-      this.oldStyle.backgroundImage = this.element.getStyle('background-image');
-      this.element.setStyle({backgroundImage: 'none'});
-    }
-    if(!this.options.endcolor)
-      this.options.endcolor = this.element.getStyle('background-color').parseColor('#ffffff');
-    if(!this.options.restorecolor)
-      this.options.restorecolor = this.element.getStyle('background-color');
-    // init color calculations
-    this._base  = $R(0,2).map(function(i){ return parseInt(this.options.startcolor.slice(i*2+1,i*2+3),16) }.bind(this));
-    this._delta = $R(0,2).map(function(i){ return parseInt(this.options.endcolor.slice(i*2+1,i*2+3),16)-this._base[i] }.bind(this));
-  },
-  update: function(position) {
-    this.element.setStyle({backgroundColor: $R(0,2).inject('#',function(m,v,i){
-      return m+(Math.round(this._base[i]+(this._delta[i]*position)).toColorPart()); }.bind(this)) });
-  },
-  finish: function() {
-    this.element.setStyle(Object.extend(this.oldStyle, {
-      backgroundColor: this.options.restorecolor
-    }));
-  }
-});
-
-Effect.ScrollTo = Class.create();
-Object.extend(Object.extend(Effect.ScrollTo.prototype, Effect.Base.prototype), {
-  initialize: function(element) {
-    this.element = $(element);
-    this.start(arguments[1] || {});
-  },
-  setup: function() {
-    Position.prepare();
-    var offsets = Position.cumulativeOffset(this.element);
-    if(this.options.offset) offsets[1] += this.options.offset;
-    var max = window.innerHeight ? 
-      window.height - window.innerHeight :
-      document.body.scrollHeight - 
-        (document.documentElement.clientHeight ? 
-          document.documentElement.clientHeight : document.body.clientHeight);
-    this.scrollStart = Position.deltaY;
-    this.delta = (offsets[1] > max ? max : offsets[1]) - this.scrollStart;
-  },
-  update: function(position) {
-    Position.prepare();
-    window.scrollTo(Position.deltaX, 
-      this.scrollStart + (position*this.delta));
-  }
-});
-
-/* ------------- combination effects ------------- */
-
-Effect.Fade = function(element) {
-  element = $(element);
-  var oldOpacity = element.getInlineOpacity();
-  var options = Object.extend({
-  from: element.getOpacity() || 1.0,
-  to:   0.0,
-  afterFinishInternal: function(effect) { 
-    if(effect.options.to!=0) return;
-    effect.element.hide().setStyle({opacity: oldOpacity}); 
-  }}, arguments[1] || {});
-  return new Effect.Opacity(element,options);
-}
-
-Effect.Appear = function(element) {
-  element = $(element);
-  var options = Object.extend({
-  from: (element.getStyle('display') == 'none' ? 0.0 : element.getOpacity() || 0.0),
-  to:   1.0,
-  // force Safari to render floated elements properly
-  afterFinishInternal: function(effect) {
-    effect.element.forceRerendering();
-  },
-  beforeSetup: function(effect) {
-    effect.element.setOpacity(effect.options.from).show(); 
-  }}, arguments[1] || {});
-  return new Effect.Opacity(element,options);
-}
-
-Effect.Puff = function(element) {
-  element = $(element);
-  var oldStyle = { 
-    opacity: element.getInlineOpacity(), 
-    position: element.getStyle('position'),
-    top:  element.style.top,
-    left: element.style.left,
-    width: element.style.width,
-    height: element.style.height
-  };
-  return new Effect.Parallel(
-   [ new Effect.Scale(element, 200, 
-      { sync: true, scaleFromCenter: true, scaleContent: true, restoreAfterFinish: true }), 
-     new Effect.Opacity(element, { sync: true, to: 0.0 } ) ], 
-     Object.extend({ duration: 1.0, 
-      beforeSetupInternal: function(effect) {
-        Position.absolutize(effect.effects[0].element)
-      },
-      afterFinishInternal: function(effect) {
-         effect.effects[0].element.hide().setStyle(oldStyle); }
-     }, arguments[1] || {})
-   );
-}
-
-Effect.BlindUp = function(element) {
-  element = $(element);
-  element.makeClipping();
-  return new Effect.Scale(element, 0,
-    Object.extend({ scaleContent: false, 
-      scaleX: false, 
-      restoreAfterFinish: true,
-      afterFinishInternal: function(effect) {
-        effect.element.hide().undoClipping();
-      } 
-    }, arguments[1] || {})
-  );
-}
-
-Effect.BlindDown = function(element) {
-  element = $(element);
-  var elementDimensions = element.getDimensions();
-  return new Effect.Scale(element, 100, Object.extend({ 
-    scaleContent: false, 
-    scaleX: false,
-    scaleFrom: 0,
-    scaleMode: {originalHeight: elementDimensions.height, originalWidth: elementDimensions.width},
-    restoreAfterFinish: true,
-    afterSetup: function(effect) {
-      effect.element.makeClipping().setStyle({height: '0px'}).show(); 
-    },  
-    afterFinishInternal: function(effect) {
-      effect.element.undoClipping();
-    }
-  }, arguments[1] || {}));
-}
-
-Effect.SwitchOff = function(element) {
-  element = $(element);
-  var oldOpacity = element.getInlineOpacity();
-  return new Effect.Appear(element, Object.extend({
-    duration: 0.4,
-    from: 0,
-    transition: Effect.Transitions.flicker,
-    afterFinishInternal: function(effect) {
-      new Effect.Scale(effect.element, 1, { 
-        duration: 0.3, scaleFromCenter: true,
-        scaleX: false, scaleContent: false, restoreAfterFinish: true,
-        beforeSetup: function(effect) { 
-          effect.element.makePositioned().makeClipping();
-        },
-        afterFinishInternal: function(effect) {
-          effect.element.hide().undoClipping().undoPositioned().setStyle({opacity: oldOpacity});
-        }
-      })
-    }
-  }, arguments[1] || {}));
-}
-
-Effect.DropOut = function(element) {
-  element = $(element);
-  var oldStyle = {
-    top: element.getStyle('top'),
-    left: element.getStyle('left'),
-    opacity: element.getInlineOpacity() };
-  return new Effect.Parallel(
-    [ new Effect.Move(element, {x: 0, y: 100, sync: true }), 
-      new Effect.Opacity(element, { sync: true, to: 0.0 }) ],
-    Object.extend(
-      { duration: 0.5,
-        beforeSetup: function(effect) {
-          effect.effects[0].element.makePositioned(); 
-        },
-        afterFinishInternal: function(effect) {
-          effect.effects[0].element.hide().undoPositioned().setStyle(oldStyle);
-        } 
-      }, arguments[1] || {}));
-}
-
-Effect.Shake = function(element) {
-  element = $(element);
-  var oldStyle = {
-    top: element.getStyle('top'),
-    left: element.getStyle('left') };
-    return new Effect.Move(element, 
-      { x:  20, y: 0, duration: 0.05, afterFinishInternal: function(effect) {
-    new Effect.Move(effect.element,
-      { x: -40, y: 0, duration: 0.1,  afterFinishInternal: function(effect) {
-    new Effect.Move(effect.element,
-      { x:  40, y: 0, duration: 0.1,  afterFinishInternal: function(effect) {
-    new Effect.Move(effect.element,
-      { x: -40, y: 0, duration: 0.1,  afterFinishInternal: function(effect) {
-    new Effect.Move(effect.element,
-      { x:  40, y: 0, duration: 0.1,  afterFinishInternal: function(effect) {
-    new Effect.Move(effect.element,
-      { x: -20, y: 0, duration: 0.05, afterFinishInternal: function(effect) {
-        effect.element.undoPositioned().setStyle(oldStyle);
-  }}) }}) }}) }}) }}) }});
-}
-
-Effect.SlideDown = function(element) {
-  element = $(element).cleanWhitespace();
-  // SlideDown need to have the content of the element wrapped in a container element with fixed height!
-  var oldInnerBottom = element.down().getStyle('bottom');
-  var elementDimensions = element.getDimensions();
-  return new Effect.Scale(element, 100, Object.extend({ 
-    scaleContent: false, 
-    scaleX: false, 
-    scaleFrom: window.opera ? 0 : 1,
-    scaleMode: {originalHeight: elementDimensions.height, originalWidth: elementDimensions.width},
-    restoreAfterFinish: true,
-    afterSetup: function(effect) {
-      effect.element.makePositioned();
-      effect.element.down().makePositioned();
-      if(window.opera) effect.element.setStyle({top: ''});
-      effect.element.makeClipping().setStyle({height: '0px'}).show(); 
-    },
-    afterUpdateInternal: function(effect) {
-      effect.element.down().setStyle({bottom:
-        (effect.dims[0] - effect.element.clientHeight) + 'px' }); 
-    },
-    afterFinishInternal: function(effect) {
-      effect.element.undoClipping().undoPositioned();
-      effect.element.down().undoPositioned().setStyle({bottom: oldInnerBottom}); }
-    }, arguments[1] || {})
-  );
-}
-
-Effect.SlideUp = function(element) {
-  element = $(element).cleanWhitespace();
-  var oldInnerBottom = element.down().getStyle('bottom');
-  return new Effect.Scale(element, window.opera ? 0 : 1,
-   Object.extend({ scaleContent: false, 
-    scaleX: false, 
-    scaleMode: 'box',
-    scaleFrom: 100,
-    restoreAfterFinish: true,
-    beforeStartInternal: function(effect) {
-      effect.element.makePositioned();
-      effect.element.down().makePositioned();
-      if(window.opera) effect.element.setStyle({top: ''});
-      effect.element.makeClipping().show();
-    },  
-    afterUpdateInternal: function(effect) {
-      effect.element.down().setStyle({bottom:
-        (effect.dims[0] - effect.element.clientHeight) + 'px' });
-    },
-    afterFinishInternal: function(effect) {
-      effect.element.hide().undoClipping().undoPositioned().setStyle({bottom: oldInnerBottom});
-      effect.element.down().undoPositioned();
-    }
-   }, arguments[1] || {})
-  );
-}
-
-// Bug in opera makes the TD containing this element expand for a instance after finish 
-Effect.Squish = function(element) {
-  return new Effect.Scale(element, window.opera ? 1 : 0, { 
-    restoreAfterFinish: true,
-    beforeSetup: function(effect) {
-      effect.element.makeClipping(); 
-    },  
-    afterFinishInternal: function(effect) {
-      effect.element.hide().undoClipping(); 
-    }
-  });
-}
-
-Effect.Grow = function(element) {
-  element = $(element);
-  var options = Object.extend({
-    direction: 'center',
-    moveTransition: Effect.Transitions.sinoidal,
-    scaleTransition: Effect.Transitions.sinoidal,
-    opacityTransition: Effect.Transitions.full
-  }, arguments[1] || {});
-  var oldStyle = {
-    top: element.style.top,
-    left: element.style.left,
-    height: element.style.height,
-    width: element.style.width,
-    opacity: element.getInlineOpacity() };
-
-  var dims = element.getDimensions();    
-  var initialMoveX, initialMoveY;
-  var moveX, moveY;
-  
-  switch (options.direction) {
-    case 'top-left':
-      initialMoveX = initialMoveY = moveX = moveY = 0; 
-      break;
-    case 'top-right':
-      initialMoveX = dims.width;
-      initialMoveY = moveY = 0;
-      moveX = -dims.width;
-      break;
-    case 'bottom-left':
-      initialMoveX = moveX = 0;
-      initialMoveY = dims.height;
-      moveY = -dims.height;
-      break;
-    case 'bottom-right':
-      initialMoveX = dims.width;
-      initialMoveY = dims.height;
-      moveX = -dims.width;
-      moveY = -dims.height;
-      break;
-    case 'center':
-      initialMoveX = dims.width / 2;
-      initialMoveY = dims.height / 2;
-      moveX = -dims.width / 2;
-      moveY = -dims.height / 2;
-      break;
-  }
-  
-  return new Effect.Move(element, {
-    x: initialMoveX,
-    y: initialMoveY,
-    duration: 0.01, 
-    beforeSetup: function(effect) {
-      effect.element.hide().makeClipping().makePositioned();
-    },
-    afterFinishInternal: function(effect) {
-      new Effect.Parallel(
-        [ new Effect.Opacity(effect.element, { sync: true, to: 1.0, from: 0.0, transition: options.opacityTransition }),
-          new Effect.Move(effect.element, { x: moveX, y: moveY, sync: true, transition: options.moveTransition }),
-          new Effect.Scale(effect.element, 100, {
-            scaleMode: { originalHeight: dims.height, originalWidth: dims.width }, 
-            sync: true, scaleFrom: window.opera ? 1 : 0, transition: options.scaleTransition, restoreAfterFinish: true})
-        ], Object.extend({
-             beforeSetup: function(effect) {
-               effect.effects[0].element.setStyle({height: '0px'}).show(); 
-             },
-             afterFinishInternal: function(effect) {
-               effect.effects[0].element.undoClipping().undoPositioned().setStyle(oldStyle); 
-             }
-           }, options)
-      )
-    }
-  });
-}
-
-Effect.Shrink = function(element) {
-  element = $(element);
-  var options = Object.extend({
-    direction: 'center',
-    moveTransition: Effect.Transitions.sinoidal,
-    scaleTransition: Effect.Transitions.sinoidal,
-    opacityTransition: Effect.Transitions.none
-  }, arguments[1] || {});
-  var oldStyle = {
-    top: element.style.top,
-    left: element.style.left,
-    height: element.style.height,
-    width: element.style.width,
-    opacity: element.getInlineOpacity() };
-
-  var dims = element.getDimensions();
-  var moveX, moveY;
-  
-  switch (options.direction) {
-    case 'top-left':
-      moveX = moveY = 0;
-      break;
-    case 'top-right':
-      moveX = dims.width;
-      moveY = 0;
-      break;
-    case 'bottom-left':
-      moveX = 0;
-      moveY = dims.height;
-      break;
-    case 'bottom-right':
-      moveX = dims.width;
-      moveY = dims.height;
-      break;
-    case 'center':  
-      moveX = dims.width / 2;
-      moveY = dims.height / 2;
-      break;
-  }
-  
-  return new Effect.Parallel(
-    [ new Effect.Opacity(element, { sync: true, to: 0.0, from: 1.0, transition: options.opacityTransition }),
-      new Effect.Scale(element, window.opera ? 1 : 0, { sync: true, transition: options.scaleTransition, restoreAfterFinish: true}),
-      new Effect.Move(element, { x: moveX, y: moveY, sync: true, transition: options.moveTransition })
-    ], Object.extend({            
-         beforeStartInternal: function(effect) {
-           effect.effects[0].element.makePositioned().makeClipping(); 
-         },
-         afterFinishInternal: function(effect) {
-           effect.effects[0].element.hide().undoClipping().undoPositioned().setStyle(oldStyle); }
-       }, options)
-  );
-}
-
-Effect.Pulsate = function(element) {
-  element = $(element);
-  var options    = arguments[1] || {};
-  var oldOpacity = element.getInlineOpacity();
-  var transition = options.transition || Effect.Transitions.sinoidal;
-  var reverser   = function(pos){ return transition(1-Effect.Transitions.pulse(pos, options.pulses)) };
-  reverser.bind(transition);
-  return new Effect.Opacity(element, 
-    Object.extend(Object.extend({  duration: 2.0, from: 0,
-      afterFinishInternal: function(effect) { effect.element.setStyle({opacity: oldOpacity}); }
-    }, options), {transition: reverser}));
-}
-
-Effect.Fold = function(element) {
-  element = $(element);
-  var oldStyle = {
-    top: element.style.top,
-    left: element.style.left,
-    width: element.style.width,
-    height: element.style.height };
-  element.makeClipping();
-  return new Effect.Scale(element, 5, Object.extend({   
-    scaleContent: false,
-    scaleX: false,
-    afterFinishInternal: function(effect) {
-    new Effect.Scale(element, 1, { 
-      scaleContent: false, 
-      scaleY: false,
-      afterFinishInternal: function(effect) {
-        effect.element.hide().undoClipping().setStyle(oldStyle);
-      } });
-  }}, arguments[1] || {}));
-};
-
-Effect.Morph = Class.create();
-Object.extend(Object.extend(Effect.Morph.prototype, Effect.Base.prototype), {
-  initialize: function(element) {
-    this.element = $(element);
-    if(!this.element) throw(Effect._elementDoesNotExistError);
-    var options = Object.extend({
-      style: {}
-    }, arguments[1] || {});
-    if (typeof options.style == 'string') {
-      if(options.style.indexOf(':') == -1) {
-        var cssText = '', selector = '.' + options.style;
-        $A(document.styleSheets).reverse().each(function(styleSheet) {
-          if (styleSheet.cssRules) cssRules = styleSheet.cssRules;
-          else if (styleSheet.rules) cssRules = styleSheet.rules;
-          $A(cssRules).reverse().each(function(rule) {
-            if (selector == rule.selectorText) {
-              cssText = rule.style.cssText;
-              throw $break;
-            }
-          });
-          if (cssText) throw $break;
-        });
-        this.style = cssText.parseStyle();
-        options.afterFinishInternal = function(effect){
-          effect.element.addClassName(effect.options.style);
-          effect.transforms.each(function(transform) {
-            if(transform.style != 'opacity')
-              effect.element.style[transform.style.camelize()] = '';
-          });
-        }
-      } else this.style = options.style.parseStyle();
-    } else this.style = $H(options.style)
-    this.start(options);
-  },
-  setup: function(){
-    function parseColor(color){
-      if(!color || ['rgba(0, 0, 0, 0)','transparent'].include(color)) color = '#ffffff';
-      color = color.parseColor();
-      return $R(0,2).map(function(i){
-        return parseInt( color.slice(i*2+1,i*2+3), 16 ) 
-      });
-    }
-    this.transforms = this.style.map(function(pair){
-      var property = pair[0].underscore().dasherize(), value = pair[1], unit = null;
-
-      if(value.parseColor('#zzzzzz') != '#zzzzzz') {
-        value = value.parseColor();
-        unit  = 'color';
-      } else if(property == 'opacity') {
-        value = parseFloat(value);
-        if(/MSIE/.test(navigator.userAgent) && !window.opera && (!this.element.currentStyle.hasLayout))
-          this.element.setStyle({zoom: 1});
-      } else if(Element.CSS_LENGTH.test(value)) 
-        var components = value.match(/^([\+\-]?[0-9\.]+)(.*)$/),
-          value = parseFloat(components[1]), unit = (components.length == 3) ? components[2] : null;
-
-      var originalValue = this.element.getStyle(property);
-      return $H({ 
-        style: property, 
-        originalValue: unit=='color' ? parseColor(originalValue) : parseFloat(originalValue || 0), 
-        targetValue: unit=='color' ? parseColor(value) : value,
-        unit: unit
-      });
-    }.bind(this)).reject(function(transform){
-      return (
-        (transform.originalValue == transform.targetValue) ||
-        (
-          transform.unit != 'color' &&
-          (isNaN(transform.originalValue) || isNaN(transform.targetValue))
-        )
-      )
-    });
-  },
-  update: function(position) {
-    var style = $H(), value = null;
-    this.transforms.each(function(transform){
-      value = transform.unit=='color' ?
-        $R(0,2).inject('#',function(m,v,i){
-          return m+(Math.round(transform.originalValue[i]+
-            (transform.targetValue[i] - transform.originalValue[i])*position)).toColorPart() }) : 
-        transform.originalValue + Math.round(
-          ((transform.targetValue - transform.originalValue) * position) * 1000)/1000 + transform.unit;
-      style[transform.style] = value;
-    });
-    this.element.setStyle(style);
-  }
-});
-
-Effect.Transform = Class.create();
-Object.extend(Effect.Transform.prototype, {
-  initialize: function(tracks){
-    this.tracks  = [];
-    this.options = arguments[1] || {};
-    this.addTracks(tracks);
-  },
-  addTracks: function(tracks){
-    tracks.each(function(track){
-      var data = $H(track).values().first();
-      this.tracks.push($H({
-        ids:     $H(track).keys().first(),
-        effect:  Effect.Morph,
-        options: { style: data }
-      }));
-    }.bind(this));
-    return this;
-  },
-  play: function(){
-    return new Effect.Parallel(
-      this.tracks.map(function(track){
-        var elements = [$(track.ids) || $$(track.ids)].flatten();
-        return elements.map(function(e){ return new track.effect(e, Object.extend({ sync:true }, track.options)) });
-      }).flatten(),
-      this.options
-    );
-  }
-});
-
-Element.CSS_PROPERTIES = $w(
-  'backgroundColor backgroundPosition borderBottomColor borderBottomStyle ' + 
-  'borderBottomWidth borderLeftColor borderLeftStyle borderLeftWidth ' +
-  'borderRightColor borderRightStyle borderRightWidth borderSpacing ' +
-  'borderTopColor borderTopStyle borderTopWidth bottom clip color ' +
-  'fontSize fontWeight height left letterSpacing lineHeight ' +
-  'marginBottom marginLeft marginRight marginTop markerOffset maxHeight '+
-  'maxWidth minHeight minWidth opacity outlineColor outlineOffset ' +
-  'outlineWidth paddingBottom paddingLeft paddingRight paddingTop ' +
-  'right textIndent top width wordSpacing zIndex');
-  
-Element.CSS_LENGTH = /^(([\+\-]?[0-9\.]+)(em|ex|px|in|cm|mm|pt|pc|\%))|0$/;
-
-String.prototype.parseStyle = function(){
-  var element = Element.extend(document.createElement('div'));
-  element.innerHTML = '<div style="' + this + '"></div>';
-  var style = element.down().style, styleRules = $H();
-  
-  Element.CSS_PROPERTIES.each(function(property){
-    if(style[property]) styleRules[property] = style[property]; 
-  });
-  if(/MSIE/.test(navigator.userAgent) && !window.opera && this.indexOf('opacity') > -1) {
-    styleRules.opacity = this.match(/opacity:\s*((?:0|1)?(?:\.\d*)?)/)[1];
-  }
-  return styleRules;
-};
-
-Element.morph = function(element, style) {
-  new Effect.Morph(element, Object.extend({ style: style }, arguments[2] || {}));
-  return element;
-};
-
-['setOpacity','getOpacity','getInlineOpacity','forceRerendering','setContentZoom',
- 'collectTextNodes','collectTextNodesIgnoreClass','morph'].each( 
-  function(f) { Element.Methods[f] = Element[f]; }
-);
-
-Element.Methods.visualEffect = function(element, effect, options) {
-  s = effect.gsub(/_/, '-').camelize();
-  effect_class = s.charAt(0).toUpperCase() + s.substring(1);
-  new Effect[effect_class](element, options);
-  return $(element);
-};
-
-Element.addMethods();
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/messageOfTheDay.js b/Tools/RunTimeTester/share/web/messageOfTheDay.js
deleted file mode 100644
index b765b3c9cb8de6d1fc814f9f6515e113634ff0b6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/messageOfTheDay.js
+++ /dev/null
@@ -1,223 +0,0 @@
-
-/************************************************/
-/*                                              */
-/*              EXPERT CORNER                   */
-/*                                              */
-/*                                              */
-/*  (If you don't want to change the way that)  */
-/*  (messages are displayed, rather you just )  */
-/*  (want to edit messages, then skip down to)  */
-/*  (the USER CORNER                         )  */
-/*                                              */
-/************************************************/
-
-var messages = new Array();
-
-/******************************************/
-/* A simple Javascript class for messages */
-/******************************************/
-
-function Message(date,text,isImportant){
-   this.theDate     = date;
-   this.theText     = text;
-   this.isImportant = isImportant;
-   messages.push(this);
-}
-
-/***************************************************/
-/* Functions to do with displaying on the web page */
-/***************************************************/
-
-/* Display style for the message text */
-function textStyle(isImportant){
-   style = 'font-family:arial;font-size:10pt;';
-   if (isImportant){
-       style = style + 'font-weight:bold;color:red';
-   }
-   else{
-       style = style + 'color:black';
-   }
-
-   return style;
-}
-
-/* Display style for the date*/
-function dateStyle(){
-   return 'font-family:arial;font-size:10pt;color:black;font-weight:bold';
-}
-
-function titleStyle(){
-   return 'font-family:sans-serif;font-weight:bold;font-size:10pt;color:#88f;background-color:#ccf';
-}
-
-/* HTML code for displaying a message*/
-function display(numbMessagesToDisplay,title){
-   htmlCode  = '<div align="center">';
-   htmlCode += '<table width="100%" border="0" cellspacing="0" cellpadding="0" bgcolor="#dddddd">';
-   htmlCode += '<tr><td colspan= "4" align="center" style="' + titleStyle() + '">' + title + '</td></tr>'
-
-   for (var i=messages.length-1;i >= (messages.length - numbMessagesToDisplay);i--){
-      message = messages[i];
-      htmlCode += '<tr style="' + dateStyle() + '">';
-      htmlCode += '<td>' + message.theDate + '</td>';
-      htmlCode += '</tr>';
-      htmlCode += '<tr style="' + textStyle(message.isImportant) + '">';
-      htmlCode += '<td>' + message.theText + '</td>';
-      htmlCode += '</tr>';
-   }
-
-   htmlCode += '</table>';
-   htmlCode += '</div>';
-
-   document.write(htmlCode);
-
-}
-
-/************************************************/
-/*                                              */
-/*                USER CORNER                   */
-/*                                              */
-/*                                              */
-/************************************************/
-
-new Message('27th February 2007','We currently run SLC3 and SLC4 (32 bit) AFS dev nightlies. Things to note:<ul><li>rel_2: SLC4_32 platform available for checking</li><li>rel_3: SLC3 platform available for checking</li><li>rel_4: RTT run, but low quality build.</li><li>rel_0/1/4/5/6: both platforms available for checking</li></ul>',false);
-
-new Message('1st March 2007','Today\'s rel_4 runs have completed successfully. Problems identified with yesterday\'s SLC4 rel_3 run:<ul><li>Run manually shutdown: 8 InDetRTT jobs affected. Cause: Batch system did not return these jobs.</li></ul>',false);
-
-new Message('2nd March 2007','The JOBOPTIONSSEARCHPATH env var has changed format in today\'s rel_5 compared to yesterday. The RTT runs have been delayed while we adapted our code. SLC4 has now completed. SLC3 ongoing.',false);
-
-new Message('3rd March 2007','Today\'s rel_6 SLC4 run complete. SLC3 run starts at 13:00. Yesterday\'s SLC3 run did not shut down correctly, but all packages completed.',false);
-
-new Message('5th March 2007','Problems with SLC3 runs since Friday. Investigation ongoing.',false);
-
-new Message('6th March 2007','Yesterday\'s rel_1 SLC3 build has been re-run and is complete. Will re-run rel_0 SLC3 shortly. 12.0.6 run seems to have been killed (probably by the network problems) before completion. Is re-running now.<br /><b>Note: RTT pilot jobs have been switched off for the past few days. Hence total number of jobs has dropped considerably.</b>',false);
-
-new Message('7th March 2007','rel_0 has been re-run. 12.0.6 is essentially done (there are a couple of jobs stuck on the batch system apparently). Today\'s rel_3 SLC4 run has almost completed.',false);
-
-new Message('8th March 2007','Today\'s rel_4 SLC4 had problems due to planned IT network intervention. Re-running. No SLC3 build was available yesterday. None will be available today (SLC3 build machines down).',false);
-
-new Message('9th March 2007','Today\'s rel_5 SLC4 is running. No SLC3 run available yet. None for past 2 days.',false);
-
-new Message('10th March 2007','Today\'s rel_6 SLC4 has run except for one long queue job. No SLC3 run available yet.',false);
-
-new Message('12th March 2007','Today\'s rel_1 SLC4 has run. No SLC3 run available yet.',false);
-
-new Message('13th March 2007','Today\'s rel_2 SLC4 has run. SLC3 build is running.',false);
-
-new Message('14th March 2007','Today\'s SLC4 run accidentally terminated. Re-running.',false);
-
-new Message('16th March 2007','Today\'s SLC4 rel_5 run ongoing.',false);
-
-new Message('18th March 2007','Problems with these web pages being addressed.',true);
-
-new Message('19th March 2007','Web pages accessible again. Previous tag back into production for now, rel_1 SLC4 ongoing.',false);
-
-new Message('20th March 2007','rel_2 SLC4 done. SLC3 ongoing.',false);
-
-new Message('21st March 2007','This morning\'s SLC4 run has just started. The current tag remains in production for now.',false);
-
-new Message('22nd March 2007','This morning\'s rel_4 SLC4 run is complete. No SLC3 build yet available.',false);
-
-new Message('25th March 2007','This morning\'s rel_0 SLC4 run is proceeding very slowly.',false);
-
-new Message('26th March 2007','This morning\'s rel_1 dev SLC4 run postponed to allow running of val SLC4 nightly.',false);
-
-new Message('27th March 2007','This morning\'s rel_2 val SLC4 run ongoing. SLC4 dev rel_2 starts 13:00.',false);
-
-new Message('2nd April 2007','This morning\'s rel_1 val SLC4 run ongoing. SLC4 dev rel_1 starts 13:00.',false);
-
-new Message('5th April 2007','This morning\'s rel_3 SLC4 bugfix run complete (44/243 jobs OK). rel_3 val starts 12:00; rel_3 SLC3 bugfix at 18:00.',false);
-
-new Message('6th April 2007','This morning\'s rel_5 SLC4 bugfix AFS build is not yet available.',false);
-
-new Message('9th April 2007','There was no rel_0 SLC3 bugfix build available yesterday. SLC4 was available, and was run (as was the val nightly)',false);
-
-new Message('9th April 2007','Batch queues have been inactivated (by an administrator, not us). Consequently no RTT runs are making progress.',true);
-
-new Message('10th April 2007','Batch queues appear normal again. This problem resulted in:<ul><li>Yesterday\'s rel_1 bugfix SLC3 build not run</li><li>Majority of jobs in rel_1 val/bugfix SLC4 build timed out.</li></ul>',true);
-
-new Message('11th April 2007','rel_3 SLC4 bugfix done. val build ongoing (not yet web visible).',false);
-
-new Message('12th April 2007','rel_4 SLC4 bugfix done. val build ongoing.',false);
-
-new Message('13th April 2007','rel_5 SLC4 bugfix and val runs ongoing.',false);
-
-new Message('14th April 2007','rel_6 SLC4 bugfix is complete, but should be ignored. The NICOS-produced flag that the RTT team was told could be used as a reliable indication that the release was ready appears, today, to have been set before all projects were truly available. The RTT thus processed an incomplete build.', true);
-
-new Message('17th April 2007','The machine on which RTT production runs are launched was switched off by IT a couple of days ago due to problems. It is still off, and so we have switched to the only other machine available to us. Unfortunately that machine appears sluggish so we have not yet been able to launch an RTT run.',false);
-
-new Message('17th April 2007','We are currently running yesterday\'s rel_1 SLC4 bugfix.',false);
-
-new Message('18th April 2007','Our main RTT production machine remains down whilst IT investigate. A second production machine to which we switched was rebooted last night whilst we were running yesterday\'s rel_2 SLC4 bugfix. The run was of course killed. We are currently re-running it but it will be stopped shortly to allow us to put in place RunTimeTester-00-01-85. When in place, we\'ll start up today\'s rel_3 val build.',false);
-
-new Message('18th April 2007','The ongoing run of yesterday\'s rel_2 SLC4 bugfix build has been halted. Today\'s rel_3 val build launches at 18:30 (using new tag RunTimeTester-00-01-85).',false);
-
-new Message('18th April 2007','Today\'s rel_3 val is almost done. SLC3 bugfix ongoing.',false);
-
-new Message('19th April 2007','Yesterday\'s ongoing rel_3 SLC3 bugfix and val runs were killed for reasons not understood. Today\'s rel_4 SLC4 bugfix build is done and available. rel_4 val is ongoing. If available, rel_4 SLC3 bugfix will launch tonight.',false);
-
-new Message('20th April 2007','<ul><li>Today\'s rel_5 SLC4 bugfix run ongoing.</li><li>Yesterday\'s rel_4 val run closed down with loss of 6 jobs (G4AtlasApps/JetRec packages). Why? Machine was under heavy stress and SLC3 bugfix run was starting up. 2 ongoing runs would have probably crashed the machine.</li></ul>',false);
-
-new Message('20th April 2007','Today\'s rel_5 SLC4 bugfix is done. val run launches at 14:20.',false);
-
-new Message('20th April 2007','Today\'s rel_5 SLC4 bugfix is done. val run finished, but is being launched again: batch system failed to return ~70 job logs.',false);
-
-new Message('20th April 2007','No rel_5 SLC3 bugfix build today.',false);
-
-new Message('23rd April 2007','Yesterday\'s bugfix and today\'s SLC4 bugfix runs should be ignored: AFS build was bad. Edit: This morning\'s SLC4 bugfix has been fixed, and so will be re-run.',false);
-
-new Message('23rd April 2007','Over the weekend RTT runs were ongoing, but failed to appear on the web pages. The pages will be manually updated shortly to make visible these results. <b>Edit: DONE. rel_1 SLC4 bugfix ongoing.</b>', false);
-
-new Message('24th April 2007','No rel_1 val build available yesterday. No rel_2 builds available so far today.',false);
-
-new Message('25th April 2007','No rel_2 bugfix SLC3 build available yesterday. Today\'s rel_3 bugfix almost done. val nightly to be re-run following fix to the build.',false);
-
-new Message('26th April 2007','Today\'s rel_4 val run ongoing. No bugfix builds available yet.', false);
-
-new Message('27th April 2007','No builds available yet.', false);
-
-new Message('30th April 2007','Since Friday the flag set by NICOS and used by the RTT to indicate that a given build is ready, has not been set. The RTT did not run any builds, except for yesterday\'s rel_0 val, for which the flag was set manually by Alex (thanks!).', true);
-
-new Message('30th April 2007','Today\'s rel_1 SLC4 bugfix almost done. val nightly ongoing.', false);
-
-new Message('2nd May 2007','Today\'s rel_3 val nightly is launching soon. No bugfix builds available yet.', false);
-
-new Message('4th May 2007','Today\'s rel_5 val ongoing; but build looks corrupt.', false);
-
-new Message('5th May 2007','Today\'s rel_6 val run launches at 11:00.', false);
-
-new Message('5th May 2007','Today\'s rel_6 val run ongoing.', false);
-
-new Message('7th May 2007','Disk quota exceeded over the weekend. Both rel_6 bugfix runs affected: 2-3 jobs (CaloRecEx/egammaRec) not completed.', false);
-
-new Message('8th May 2007','Today\'s rel_2 SLC4 bugfix and val runs are ongoing.', false);
-
-new Message('8th May 2007','All today\'s rel_2 runs complete (in all runs, some jobs from CaloRecEx/egammaRec got stuck in batch and were aborted manually).', false);
-
-new Message('13th May 2007','There was no rel_6 val build available yesterday.', false);
-
-new Message('16th May 2007','The empty "Project" column below (if you are on the start page with the list of available RTT runs) will gradually be filled.', false);
-
-new Message('7th June 2007','<h2>Here is the only working link to <a target="main_frame" href="http://cern.ch/atlasrtt/13.0.10/build/i686-slc4-gcc34-opt/offline/page2.xml">13.0.10 results</a>.</h2>', true);
-
-new Message('18th June 2007','<h2>There are problems with RTT cluster. Because of these, there were no runs yesterday 17th and none this morning.Please check RTT hypernews forum for updates.</h2>', true);
-
-new Message('20th June 2007','<h3>RTT running has resumed.</h3>', false);
-
-new Message('21st June 2007','<ul><li><h2>Here is the only working link to <a target="main_frame" href="http://cern.ch/atlasrtt/13.0.10/build/i686-slc4-gcc34-opt/offline/page2.xml">13.0.10 results</a>.</h2></li></ul>', true);
-
-new Message('6th June 2007','<ul><li><h2>Here are the only working links to <a href="http://cern.ch/atlasrtt/page1.php?xml=13.0.10/build/i686-slc4-gcc34-opt/offline/RTTSummary.xml&viewBy=package">13.0.10 results</a> and <a href="http://cern.ch/atlasrtt/page1.php?xml=13.0.20/build/i686-slc4-gcc34-opt/offline/RTTSummary.xml&viewBy=package">13.0.20 results</a>.</h2></li></ul>', true);
-
-/***********************************************************/
-/*        ADDING A NEW MESSAGE                             */
-/***********************************************************/
-/* Add your latest message above this comment box          */
-/* Skeleton code is:                                       */
-/*   new Message('theDate','theMessage',false);            */
-/*                                                         */
-/*  where false means the message gets displayed normal    */
-/*        true means message displayed in red and bold     */
-/***********************************************************/
-
-/* Show the last how many messages ? */
-display(numberOfMessagesToDisplay = 1,title=":: Message of the Day ::");
diff --git a/Tools/RunTimeTester/share/web/page0.php b/Tools/RunTimeTester/share/web/page0.php
deleted file mode 100644
index fe7a13dcc720d60d5b9f6e27e1d1761ee76ac702..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/page0.php
+++ /dev/null
@@ -1,163 +0,0 @@
-<html>
-<head>
-<style>
-body{font-family:Trebuchet,sans-serif;background-color:#eeeeff;text-align:center;margin:0}
-#admin{font-size:0.7em;font-weight:bold;letter-spacing:2px;background-color:#8888df;color:white;font-family:verdana,sans-serif;cursor:pointer}
-#resultsTable{font-size:1.2em;font-weight:bold;border:1px gray solid;padding:2px;background-color:#bbbbff;margin:auto;}
-#tableHeader td {background-color:#ccccff;padding:3px;text-align:center}
-.tableRow td {background-color:#ddddff;font-weight:normal;font-size:0.7em;padding:5px}
-.tableRow td a:link{text-decoration:underline; color:black;}   
-.tableRow td a:visited{text-decoration:underline; color:black;}   
-.tableRow td a:hover{text-decoration:none; color:orange;}
-.tableRow td a:active{text-decoration:underline; color:black;}   
-#tableHeader td a:link{text-decoration:underline; color:black;}   
-#tableHeader td a:visited{text-decoration:underline; color:black;}   
-#tableHeader td a:hover{text-decoration:none; color:orange;}
-#tableHeader td a:active{text-decoration:underline; color:black;}   
-#latestnews{position:fixed;bottom:0px;left:20%;right:20%;font-size:10pt;text-align:center;border:1px solid red;width:60%;font-family:sans-serif;background-color:#ddaaaa}
-</style>
-<title>ATLAS RunTimeTester Results</title>
-<script src="prototype.js" type="text/javascript"></script>
-<script src="effects.js" type="text/javascript"></script>
-<script type="text/javascript">
-<!--
-function toggleThis(srcID, tgtID){
-  var el = document.getElementById(srcID);
-  if(document.getElementById(tgtID).style.display == 'none'){
-    new Effect.Appear(tgtID);
-    el.innerHTML = "Click me to Hide";
-  }
-  else{
-    new Effect.DropOut(tgtID);
-    el.innerHTML = "Click me to Show";
-  }
-}
-
-function emphasize(){
-  setTimeout("toggleThis('admin','topFrame')", 500);
-}
-// -->
-</script>
-
-</head>
-<body>
-<div id='admin' onclick="toggleThis('admin','topFrame');">Click me to Show</div>
-
-<!------------------------------>
-<!-- TOP FRAME WITH THE LINKS -->
-<!------------------------------>
-<div id='topFrame' style="display:none">
-<script src="RTTpage1_LinksAtTopOfPage.js" type="text/JavaScript"></script>
-</div>
-<center>
-<div id="latestnews"><b>Latest News:</b><br />            
-8th Jan@09:30: Last night a new RTT version went into production, providing default data access to
-the atlasgroupdisk CASTOR area. If your job is failing due to dataset access problems this is probably why.
-</div>  
-</center>
-<br />
-
-<?
-$sortBy = $_GET['sort'];
-?>
-
-<?
-include("tools.php");
-
-
-$summaryFiles = openAndReadSummaryFilePaths('summaryFilePaths.txt');
-
-if(is_null($summaryFiles) || sizeOf($summaryFiles)==0){
-  echo "<h1>No data available</h1>";
-  exit;
- }
-
-$runs = array();
-
-foreach($summaryFiles as $pathToSummFile){
-  $overviewFile = dirname($pathToSummFile)."/overview.xml";
-  if(!file_exists($pathToSummFile) 
-     || !file_exists($overviewFile) 
-     || filesize($pathToSummFile)==0
-     || filesize($overviewFile)==0){continue;}
-
-  $oNode = getRootNodeFromXMLfile($overviewFile);
-
-  $run = array();
-
-  // now grab info from that overview node
-  $run['summFile']     = $pathToSummFile;
-  $run['release']      = getTagContent($oNode, 'release');
-  $run['branch']       = getTagContent($oNode, 'originalBranch');
-  $run['platform']     = getTagContent($oNode, 'targetCMTCONFIG');
-  $run['topProject']   = getTagContent($oNode, 'topProject');
-  $run['otherProject'] = getTagContent($oNode, 'otherProject');
-  $run['started']      = getTagContent($oNode, 'startTime1');
-  $run['completed']    = getTagContent($oNode, 'endTime1');
-  $run['nicos']        = getTagContent($oNode, 'nicosDate1');
-  $run['status']       = getTagContent($oNode, 'statusText');
-
-  $runs[] = $run;
-  $oNode->unlink();
-  unset($oNode);
-}
-
-$sortBy = $_GET['sort'];
-$sortMap = array('byRelease'   => 'sortSummaryFilesByRelease',
-                 'byBranch'    => 'sortSummaryFilesByBranch',
-                 'byProject'   => 'sortSummaryFilesByProject',
-                 'byPlatform'  => 'sortSummaryFilesByPlatform',
-                 'byStarted'   => 'sortSummaryFilesByStarted',
-                 'byCompleted' => 'sortSummaryFilesByCompleted');
- 
- 
-if(trim($sortBy) == ''){$sortBy='byRelease';}
-usort($runs, $sortMap[$sortBy]);
-?>
-
-<table id='resultsTable'>
-<tr id='tableHeader'>
-   <td>View By</td>
-   <td><a href="page0.php?sort=byRelease">Release</a></td>
-   <td><a href="page0.php?sort=byBranch">Branch</a></td>
-   <td><a href="page0.php?sort=byPlatform">Platform</a></td>
-   <td><a href="page0.php?sort=byProject">Project</a></td>
-   <td><a href="page0.php?sort=byStarted">Started</a></td>
-   <td><a href="page0.php?sort=byCompleted">Completed</a></td>
-   <td>NICOS</td>
-   <td>Status</td>
-</tr>
-
-<?
-foreach($runs as $run){
-  //Special cases
-  if ($run['branch']=='bugfix'){$displayBranch = '14.2.X';}
-  else if ($run['branch']=='val'){$displayBranch = '14.2.X-VAL';}
-  else{$displayBranch=$run['branch'];}
-  $byPackageLink = 'page1.php?xml='.$run['summFile'].'&viewBy=package';
-  $byCPCPackageLink = 'page1.php?xml='.$run['summFile'].'&viewBy=cpcByPackage';
-  $byCPCLink = 'page1.php?xml='.$run['summFile'].'&viewBy=cpc';
-  echo "<tr class='tableRow'>";
-  echo "<td><a href='".$byCPCLink."'>[CPC]</a>&nbsp;<a href='".$byCPCPackageLink."'>[CPC-by-package]</a>&nbsp;<a href='".$byPackageLink."'>[package]</a></td>";
-  echo "<td>".$run['release']."</td>";
-  echo "<td>".$displayBranch."</td>";
-  echo "<td>".$run['platform']."</td>";
-  if($run['otherProject']!=''){
-     echo "<td>".$run['topProject']."/".$run['otherProject']."</td>";
-  }
-  else{
-     echo "<td>".$run['topProject']."</td>";
-  }   
-  echo "<td>".$run['started']."</td>";
-  echo "<td>".$run['completed']."</td>";
-  echo "<td>".$run['nicos']."</td>";
-  echo "<td>".$run['status']."</td>";
-  echo "</tr>";
-}
-
-?>
-
-</table>
-
-</body>
-</html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/page1.php b/Tools/RunTimeTester/share/web/page1.php
deleted file mode 100644
index 7154bf0783f548960ef3c991b8565659054c3219..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/page1.php
+++ /dev/null
@@ -1,1065 +0,0 @@
-<?
-ob_start("ob_gzhandler"); 
-header("Content-type: text/html; charset=utf-8");
-?>
-
-<html>
-<head>
-<style>
-body{font-family:Trebuchet,sans-serif;background-color:#eeeeff;text-align:center;margin:0;color:black}
-.pageHeader{font-size:1.5em;font-weight:bold;}
-.sectionHeader{font-size:1em;font-weight:bold;}
-.statusOK{color:green}
-.statusBad{color:red}
-.statusNeutral{color:black}
-.statusRunning{color:brown}
-#viewTitle{font-size:1.2em;font-weight:bold;letter-spacing:3px;font-family:verdana, sans-serif}
-.statHeader{font-weight:bold;background-color:#bbbbff;text-align:left;font-size:0.9em}
-.statTitle{font-weight:bold;background-color:#ccccff;font-size:0.8em;text-align:right}
-.statEntry{font-weight:normal;font-size:0.8em;text-align:left}
-#admin{font-size:0.7em;font-weight:bold;letter-spacing:2px;background-color:#8888df;color:white;font-family:verdana, sans-serif}
-a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-a.switchView{font-family:arial;font-size:0.7em;color:black;text-decoration:underline}
-a.switchView:hover{font-family:arial;color:black;text-decoration:none}
-span.stats{font-family:arial;font-size:0.8em;color:black;text-decoration:none}
-#tableHeader td {background-color:#bbbbff;padding:3px;text-align:center;font-weight:bold}
-.packageTableHeader{background-color:#bbbbbb;padding:3px;text-align:left;font-weight:bold;font-size:1.3em;color:#666666}
-.frameworkHeader{background-color:#ddddff;padding:3px;text-align:center;font-size:1.0em;font-weight:bold}
-.cpcTableHeader{background-color:#ddddff;text-align:right;font-weight:normal;}
-.cpcTableColHeader{background-color:#aaaaff;text-align:center;font-weight:bold}
-.minderEntry{background-color:#eeeeff;padding:3px;text-align:left;font-size:0.8em;font-weight:normal}
-.stats{font-weight:normal;font-size:0.8em}
-.tableRow td {background-color:white;font-weight:normal;font-size:0.7em;padding:5px;text-align:center}
-.tableRow td a:link{text-decoration:underline; color:black;}   
-.tableRow td a:visited{text-decoration:underline; color:black;}   
-.tableRow td a:hover{text-decoration:none; color:orange;}
-.tableRow td a:active{text-decoration:underline; color:black;}
-.tableRow td span{background-color:white;font-weight:normal;font-size:0.8em;padding:5px;text-align:center}
-.tableRow2 td {background-color:#dddddd;font-weight:bold;font-size:0.7em;padding:5px;text-align:center}
-.packageRow td {background-color:#ddddff;font-weight:normal;font-size:0.7em;padding:5px;text-align:center}
-.packageRow td a:link{text-decoration:underline; color:black;}   
-.packageRow td a:visited{text-decoration:underline; color:black;}   
-.packageRow td a:hover{text-decoration:none; color:orange;}
-.packageRow td a:active{text-decoration:underline; color:black;}
-.showJobs{font-weight:normal;font-size:0.8em;text-decoration:underline;cursor:pointer;}
-#tableHeader td a:link{text-decoration:underline; color:black;}   
-#tableHeader td a:visited{text-decoration:underline; color:black;}   
-#tableHeader td a:hover{text-decoration:none; color:orange;}
-#tableHeader td a:active{text-decoration:underline; color:black;} 
-</style>
-<script src="prototype.js" type="text/javascript"></script>
-<script src="effects.js" type="text/javascript"></script>
-<script type="text/javascript">
-<!--
-function toggleThis(srcID, tgtID){
-  var el = document.getElementById(srcID);
-  if(document.getElementById(tgtID).style.display == 'none'){
-    new Effect.Appear(tgtID);
-    el.innerHTML = "Click to Hide";
-  }
-  else{
-     new Effect.DropOut(tgtID);
-     el.innerHTML = "Click to Show";
-  }
-}
-
-function toggleStuff(buttonID, tgtID){
-  var abutton = document.getElementById(buttonID);
-
-  if (document.getElementById(tgtID).style.display == 'none'){
-    new Effect.Appear(tgtID);
-
-    abutton.innerHTML = '<font color="red">[Hide]</font>';
-
-  }
-  else{
-    new Effect.DropOut(tgtID);
-    abutton.innerHTML = '[Show]';
-  }
-}
-// -->
-</script>
-</head>
-<body>
-<div id='admin' onclick="toggleThis('admin','topFrame');">Click to Show</div>
-<div id='topFrame' style="display:none">
-<script src="RTTpage1_LinksAtTopOfPage.js" type="text/JavaScript"></script>
-<script src="messageOfTheDay.js" type="text/JavaScript"></script>
-</div>
-<br />
-
-<?
-include("tools.php");
-
-// Grab the GET parameters in the URL
-$xmlFile = $_GET['xml'];
-
-if (!file_exists($xmlFile)){
-  echo "<br /><b><center>No data available yet.</center></b>";
-  return;
-}
-
-//Select specific packages to view
-$view = $_GET['viewBy'];
-$packagesToView = trim($_GET['package']);
-
-//Select specific states to view
-$packageStatesToView = trim($_GET['status']);
-
-//Open up the XML file
-$rootNode = getRootNodeFromXMLfile($xmlFile);
-$overviewNodes = $rootNode->get_elements_by_tagname('overview');
-$overviewNode = $overviewNodes[sizeOf($overviewNodes)-1]; //0th one is ATN summary, ignore
-
-//////////////////////////////////////////////////////////////////////////
-// SECTION 1: INFORMATION AT TOP OF THE PAGE
-//////////////////////////////////////////////////////////////////////////
-
-// Link to RTTSummary.txt
-$diagLink = explode('.xml', $xmlFile);
-$diagLink = $diagLink[0].".txt";
-echo "<a class='pageLink' href='".$diagLink."'>RTT developer run diagnostics</a>\n";
-echo "<br /><br />\n\n";
-
-
-$releaseName = getTagContent($overviewNode, 'releaseName');
-$runType     = getTagContent($overviewNode, 'runType');
-$platform    = getTagContent($overviewNode, 'targetCMTCONFIG');
-$topProj     = getTagContent($overviewNode, 'topProject');
-$otherProj   = getTagContent($overviewNode, 'otherProject');
-echo "<div class='pageHeader'>";
-if($otherProj==''){
-  echo $releaseName."/".$runType." --- ".$platform." --- ".$topProj;
-}
-else{
-  echo $releaseName."/".$runType." --- ".$platform." --- ".$topProj.'/'.$otherProj;
-}
-echo "</div>\n\n";
-echo "<br />\n\n";
-
-// Now the RTT versions at top of the page
-$rttSrc = getTagContent($overviewNode, 'rttSrcVersion');
-$rttLib = getTagContent($overviewNode, 'rttLibVersion');
-echo "<div class='sectionHeader'>";
-echo "RTT src code: ".$rttSrc;
-echo "<br />";
-echo "RTT library tools: ".$rttLib;
-echo "</div>\n\n";
-echo "<br />\n\n";
-
-// Find the package nodes, and count them
-$packages = $rootNode->get_elements_by_tagname('package');
-
-///////////////////////////////////////////////////////////
-////   FILTER THE PACKAGES
-///////////////////////////////////////////////////////////
-
-function filter($packages, $what, $allowedValues){
-  $acceptedPackages = array();
-  foreach($packages as $package){
-    $value = getTagContent($package, $what);
-    if (in_array($value, $allowedValues)){$acceptedPackages[]=$package;}
-  }
-  return $acceptedPackages;
-}
-
-// Does the user wish to see all or some packages?
-if ($packagesToView != ''){
-  $userPackages = split('::', $packagesToView);
-  $packages = filter($packages, 'packageName', $userPackages);
- }
-
-
-// Does the user wish to see some packages statii or all?
-if ($packageStatesToView != ''){
-  if ($packageStatesToView!='all'){
-    $states = split('::',$packageStatesToView);
-    $packages = filter($packages, 'packageStatus', $states);
-  }
-}
-else{ // default is to show only running and done states
-  $packages = filter($packages, 'packageStatus', array('running', 'done'));
-}
-
-//////////////////////////////////////////////////////////
-
-echo "<span class='pageHeader'>Packages found by the RTT: ".sizeOf($packages)."</span>\n";
-$currentURL = $_SERVER["HTTP_HOST"].$_SERVER["REQUEST_URI"];
-$currentURLtoks = explode('page1.php', $currentURL);
-$currentURL = 'page1.php'.$currentURLtoks[1];
-if($packageStatesToView=='all'){
-  echo "<a class='switchView' href='".$currentURL."&status=running::done'>[Show only running/done packages]</a>\n\n";
-}
-else{
-  echo "<a class='switchView' href='".$currentURL."&status=all'>[Show all packages]</a>\n\n";
-}
-echo "<br /><br />\n\n";
-
-$sText  = getTagContent($overviewNode, 'statusText');
-$jobsOK = extractFromStatusText($sText, 'ok');
-$done   = extractFromStatusText($sText, 'done');
-$total  = extractFromStatusText($sText, 'total');
-
-echo "<span class='stats'>[Jobs OK/Done/Total: ".$jobsOK."/".$done."/".$total."--- Pkgs with all jobs OK: ".getTagContent($overviewNode, 'nPackagesAllJobsOK')."/".getTagContent($overviewNode, 'nPackages')."]</span>";
-echo "<br /><br />";
-
-//////////////////////////////////////////////////////////////////////////
-// SECTION 2: LOAD THE VIEW
-//////////////////////////////////////////////////////////////////////////
-
-//set as default, if not set
-if ($view == '') $view = 'package'; 
-
-if ($view == 'cpcByPackage'){
-  // Show the package view options
-  echo "<div id='viewTitle'>CPC-by-Package View</div>";
-  echo " [<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=package'>Switch to package view</a>]";
-  echo "[<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=cpc'>Switch to CPC view</a>]<br /><br />";
-
-  // Show the package states options
-  echo "<div id='viewTitle'>View</div>";
-
-  doByCPCPackageView($packages);
-}
-else if($view == 'package'){
-  echo "<div id='viewTitle'>Package View</div>";
-  echo "[<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=cpc'>Switch to CPC view</a>]";
-  echo " [<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=cpcByPackage'>Switch to CPC-by-package view</a>]<br /><br />";
-  doByPackageView($packages);
-}
-else if($view == 'cpc'){
-  echo "<div id='viewTitle'>CPC View</div>\n";
-  echo "[<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=package'>Switch to package view</a>]\n";
-  echo "[<a class='switchView' href='page1.php?xml=".$xmlFile."&viewBy=cpcByPackage'>Switch to CPC-by-package view</a>]\n\n<br /><br />\n\n";
-  doByCPCView($packages);
-}
-
-///////////////////////////////////////////////////////////////////
-// VIEW 1: DO THE TRUE BY-CPC VIEW
-///////////////////////////////////////////////////////////////////
-function doByCPCView($packages){
-  global $xmlFile;
-  $sortBy = $_GET['sortBy'];;
-
-  function sortByCPC($packages){
-    $byCPC = array();
-    foreach($packages as $package){
-      $packageName = getTagContent($package, 'packageName');
-      $minders = $package->get_elements_by_tagname('minder');
-
-      foreach($minders as $minder){
-	$classList     = getMultiTagContent($minder, 'displayClass');
-	$processList   = getMultiTagContent($minder, 'displayProcess');
-	$componentList = getMultiTagContent($minder, 'displayComponent');
-
-	foreach($classList as $class){
-	  foreach($processList as $process){
-	    foreach($componentList as $component){
-
-	      $key = "[".$class."] [".$process."] [".$component."]";
-	      if (array_key_exists($key, $byCPC)){
-		$dict = $byCPC[$key]; // dictionary keyed on package name
-		if (array_key_exists($packageName, $dict)){
-		  $minderList = $dict[$packageName];
-		  $minderList[] = $minder;
-		  $dict[$packageName] = $minderList;
-		  $byCPC[$key] = $dict;
-		}
-		else{
-		  $dict[$packageName] = array($minder);
-		  $byCPC[$key] = $dict;
-		}
-	      }
-	      else{
-		$val = array($packageName => array($minder));
-		$byCPC[$key] = $val;
-	      }
-	    }
-	  }
-	}
-      }
-    }
-    
-    return $byCPC;
-  }
-
-  // --------------------------------------------------
-
-  function arrayOnlyHasSuccesses($theArray){
-    foreach($theArray as $entry){if (trim($entry)!='success') return false;}
-    return true;
-  }
-  
-  // --------------------------------------------------
-
-  function cpcStatus($dict){ // dict = {packageName:list of minders}
-    $totNumbJobs  = 0;
-    $totNumbPacks = 0;
-    $totNotOKJobs  = 0;
-    $statii = array();
-    foreach($dict as $pName => $minders){
-      $totNumbPacks++;
-      foreach($minders as $minder){
-	$status = getTagContent($minder, 'status');
-	$statii[] = $status;
-	$totNumbJobs++;
-	if ($status != 'success'){$totNotOKJobs++;}
-      }
-    }
-
-    if(in_array('running', $statii)) return array('status' => 'ongoing', 'jobsInCPC' => $totNumbJobs, 'packagesInCPC' => $totNumbPacks, 'notOKjobs' => $totNotOKJobs);
-    else if (arrayOnlyHasSuccesses($statii)) return array('status' => 'OK', 'jobsInCPC' => $totNumbJobs, 'packagesInCPC' => $totNumbPacks, 'notOKjobs' => 0);
-    return array('status' => 'FAIL', 'jobsInCPC' => $totNumbJobs, 'packagesInCPC' => $totNumbPacks, 'notOKjobs' => $totNotOKJobs);
-  }
-
-  // --------------------------------------------------
-
-  function colorizeCPCStatus($statusText){
-    if (trim($statusText)=='OK') return '<span style="color:#00ff00;font-weight:bold">OK</span>';
-    if (trim($statusText)=='FAIL') return '<span style="color:#ff0000;font-weight:bold">FAIL</span>';
-    return '<font color="black">ongoing</font>';
-  }
-
-  // --------------------------------------------------
-
-  $cpcCategories = sortByCPC($packages);
-
-  // Remove non-CPC jobs
-  $nonCPCKey = '[] [] []';
-  if(array_key_exists($nonCPCKey, $cpcCategories)){
-    unset($cpcCategories[$nonCPCKey]);
-  }
-
-
-  if (sizeOf(array_keys($cpcCategories)) == 0){
-    echo "<div style='font-weight:bold;'>There are no jobs with CPC values.</div>";
-    return;
-  }
-
-
-  // Now show the table of CPC values
-  echo "<table width='60%' border='0' align='center'>\n\n";
-
-  $colLink = "page1.php?xml=".$xmlFile."&viewBy=cpc&sortBy=";
-
-  echo "<tr id='tableHeader' align='center'>\n";
-  echo "   <td></td>\n";
-  echo "   <td><a href='".$colLink."class'>Class</a></td>\n";
-  echo "   <td><a href='".$colLink."process'>Process</a></td>\n";
-  echo "   <td><a href='".$colLink."component'>Component</a></td>\n";
-  echo "   <td>Status</td>\n";
-  echo "   <td>Summary</td>\n";
-  echo "</tr>\n\n";
-  
-  // Table Column sorting
-  $sortMap = array('class'     => 'sortCPCByClass',
-		   'process'   => 'sortCPCByProcess',
-		   'component' => 'sortCPCByComponent');
-  
-  //default is sort by release
-  if ($sortBy == ''){$sortBy='class';}
-  $cpcNames = array_keys($cpcCategories);
-  usort($cpcNames, $sortMap[$sortBy]);
-
-  // Now loop over the CPC categories (rows in table)
-  foreach($cpcNames as $cpcName){
-    $dict = $cpcCategories[$cpcName]; // dict = {packageName:list of minders}
-    //foreach($cpcCategories as $cpcName => $dict){ // dict = {packageName:list of minders}
-
-    $cpcToks = explode(' ', $cpcName);
-    $val = cpcStatus($dict);
-
-    $buttonID = $cpcName;
-    $jobsID   = "jobs_".$cpcName;
-
-    $show_Them = "<span style='text-decoration:underline' id='".$buttonID."' onclick='toggleStuff(\"".$buttonID."\",\"".$jobsID."\");'>[Show]</span>";
-    echo "<tr class='cpcTableHeader'>\n";
-    echo "   <td align='center'>".$show_Them."</td>\n";
-    echo "   <td>".$cpcToks[0]."</td>\n";
-    echo "   <td>".$cpcToks[1]."</td>\n";
-    echo "   <td>".$cpcToks[2]."</td>\n";
-    echo "   <td align='center'>".colorizeCPCStatus($val['status'])."</td>";
-
-    if($val['status'] == 'ongoing'){
-      echo "   <td class='statEntry'>".$val['jobsInCPC']." jobs total, ".$val['packagesInCPC']." packages total</td>\n";
-    }
-    else if($val['status'] == 'OK'){
-      echo "   <td class='statEntry'>".$val['jobsInCPC']."/".$val['jobsInCPC']." jobs OK; ".$val['packagesInCPC']." packages total</td>\n";
-    }
-    else{
-      echo "   <td class='statEntry'>".$val['notOKjobs']."/".$val['jobsInCPC']." jobs not OK; ".$val['packagesInCPC']." packages total</td>\n";
-    }
-
-    echo "</tr>\n\n";
-
-    echo "<tr>\n";
-    echo "   <td colspan='3'>\n";
-    echo "      <div  align='center' style='display:none' id='".$jobsID."'>\n";
-
-    echo "         <table width='80%'>\n";
-
-    foreach($dict as $pName => $minders){
-
-      echo "            <tr class='packageTableHeader'><td colspan='10'>".$pName."</td></tr>\n";
-      echo "            <tr class='tableRow2'><td>Name</td><td>JobGroup</td><td>Job status</td><td>Tests</td><td>Datasets</td><td>Hash</td></tr>";
-      foreach($minders as $minder){
-
-	$jobName = getTagContent($minder, 'jobName');
-	$escapedJobName = rawurlencode($jobName);
-	$displayName = getTagContent($minder, 'jobDisplayName');
-	if ($displayName == ''){$displayName = $jobName;}
-	$jobID   = getTagContent($minder, 'jobID');
-	$group   = getTagContent($minder, 'jobGroup');
-
-	// ==================================================
-	// Job doc/link stuff
-	// ==================================================
-	
-	/*
-	$jobDocString = getTagContent($minder, 'jobDocString');
-	$jobDocURL    = getTagContent($minder, 'jobDocURL');
-
-	if(sizeOf($jobDocString)!=0){
-	  $docString ='';
-	  if (!isEmptyTag($jobDocString)){
-	    $docString = trim($doc->get_content());
-	  }
-	}
-
-	if(sizeOf($jobDocURL)!=0){
-	  $doclink = $jobDocURL;
-	  $docURL = '';
-	  if (!isEmptyTag($doclink)){
-	    $docURL = "<a class='pageLink' href='".trim($doclink->get_content())."'>More info</a>";
-	  }
-	}
-	*/
-	// ==================================================
-	// Hash stuff
-	// ==================================================
-
-	$hash   = getTagContent($minder, 'hashString');
-	if ($hash==''){$hash = 'n/a';}
-	// Now add the datasets
-	$datasetsList = $minder->get_elements_by_tagname('datasets');
-	$datasetsTag  = $datasetsList[0];
-	$datasets_holder = '';
-	if (isEmptyTag($datasetsTag)){
-	  $datasets = "n/a";
-	}
-	else{
-	  $datasets = "";
-	  foreach($datasetsTag->get_elements_by_tagname('dataset') as $dataset){
-	    $datasets .= trim($dataset->get_content())."\n";
-	  }
-
-	  $datasetsButtonID = $jobID."_datasetsButtonID";
-	  $datasetsID = $jobID."_datasetsID";
-	  $datasets_holder = "<span style='text-decoration:underline' id='".$datasetsButtonID."' onclick='toggleStuff(\"".$datasetsButtonID."\",\"".$datasetsID."\");'>[Show]</span>";
-	}
-	
-	// Job status
-	$status = getTagContent($minder, 'status');
-	//$jobsStatii[] = $status; //append to array
-
-	$statusClass = 'statusBad';
-	if ($status == 'success') $statusClass = 'statusOK';
-	if ($status == 'running') $statusClass =  'statusRunning';
-	
-	// Post processing combined status
-	$postProcRes = getTagContent($minder, 'postProcessingResult');
-	$postProcClass = 'statusBad';
-	if ($postProcRes == 'success') $postProcClass = 'statusOK';
-	if ($postProcRes == 'running') $postProcClass =  'statusRunning';
-	if ($postProcRes == 'no tests') $postProcClass =  'statusNeutral';
-      
-	// Ref. release used
-	$refRel = getTagContent($minder, 'refRelease');
-
-	// =====================================================
-	// Now create the table row
-	// =====================================================
-
-	echo "\n<tr class='tableRow'>\n";
-
-	echo "<td valign='top'><a href='page2.php?xml=".$xmlFile."&package=".$pName."&job=".$escapedJobName."&id=".$jobID."'>".$displayName."</a><br /><br /><span>[id:".$jobID."]</span><br /><span>".$docString."</span></td>\n";
-	echo "<td valign='top'>".$group."</td>\n";
-	echo "<td valign='top' class='".$statusClass."'>".$status."</td>\n";
-	echo "<td valign='top' class='".$postProcClass."'>".$postProcRes."</td>\n";
-
-	if ($datasets_holder==''){
-	  echo "<td valign='top'>n/a</td>";
-	}
-	else{
-	  echo "<td valign='top'>";
-	  echo $datasets_holder;
-	  echo "<div align='center' style='display:none' id='".$datasetsID."'>".$datasets."</div>";
-	  echo "</td>\n";
-	}
-
-	echo "<td valign='top'>".$hash."</td>\n";
-	//echo "<td valign='top'>".$refRel."</td>\n";
-	echo "</tr>";
-      }
-      
-    }
-
-    echo "</table>\n"; // end of minder table      
-    echo "</div></td></tr>";
-
-  }
-
-  echo "</table>";
-
-}
-
-///////////////////////////////////////////////////////////////////
-// VIEW 2: DO THE BY-CPC (PACKAGE) VIEW
-///////////////////////////////////////////////////////////////////
-function doByCPCPackageView($packages){
-  global $xmlFile;
-
-  function sortByCPC($minders){
-    //Sort minders into CPC categories for this package
-    $cpcHash = array();
-    foreach($minders as $minder){
-      $class     = getTagContent($minder, 'displayClass');
-      $process   = getTagContent($minder, 'displayProcess');
-      $component = getTagContent($minder, 'displayComponent');
-      $key = "[".$class."] [".$process."] [".$component."]";
-      if (array_key_exists($key, $cpcHash)){
-	$val = $cpcHash[$key];
-	$val[] = $minder;
-	$cpcHash[$key] = $val;
-      }
-      else{
-	$cpcHash[$key] = array($minder);
-      }
-    }
-
-    return $cpcHash;
-  }
-
-  function sortByCPC2($minders){
-    //Sort minders into CPC categories for this package
-    $cpcHash = array();
-    foreach($minders as $minder){
-      $classTagsCont     = getMultiTagContent($minder, 'displayClass');
-      $processTagsCont   = getMultiTagContent($minder, 'displayProcess');
-      $componentTagsCont = getMultiTagContent($minder, 'displayComponent');
-      
-      foreach($classTagsCont as $class){
-	foreach($processTagsCont as $process){
-	  foreach($componentTagsCont as $component){
-	    $key = "[".$class."] [".$process."] [".$component."]";
-	    if (array_key_exists($key, $cpcHash)){
-	      $val = $cpcHash[$key];
-	      $val[] = $minder;
-	      $cpcHash[$key] = $val;
-	    }
-	    else{
-	      $cpcHash[$key] = array($minder);
-	    }
-	  } 
-	}
-      }
-    }
-
-    return $cpcHash;
-  }
-
-  function arrayOnlyHasSuccesses($theArray){
-    foreach($theArray as $entry){if (trim($entry)!='success') return false;}
-    return true;
-  }
-  
-  function cpcCategoryStatus($minders){
-    $statii = array();
-    foreach($minders as $minder){
-      $statii[] = getTagContent($minder, 'status');
-    }
-
-    if(in_array('running', $statii)) return 'ongoing';
-    else if (arrayOnlyHasSuccesses($statii)){ return 'OK';}
-    return 'FAIL';
-  }
-
-  function colorizeCPCStatus($statusText){
-    if (trim($statusText)=='OK') return '<span style="color:#00ff00;font-weight:bold">OK</span>';
-    if (trim($statusText)=='FAIL') return '<span style="color:#ff0000;font-weight:bold">FAIL</span>';
-    return '<font color="black">ongoing</font>';
-  }
-
-  function colorizePackageStat($stat){
-    if(trim($stat)!='running' && trim($stat)!='done'){
-      return '<span style="color:red">'.$stat.'</span>';
-    }
-    return $stat;
-  }
-
-  function countNotOKJobs($minders){
-    $tot = 0;
-    foreach($minders as $minder){
-      if(getTagContent($minder, 'status') != 'success'){$tot++;}
-    }
-    return $tot;
-  }
-
-  function countNotOKTests($minders){
-    $tot=0;
-    foreach($minders as $minder){
-      if(getTagContent($minder, 'postProcessingResult') != 'success'){$tot++;}
-    }
-    return $tot;
-  }
-
-  // Let's make the table
-  echo "\n<table align='center' border='0' cellspacing='3' cellpadding='3' width='60%'>";
-
-  //Sort and loop through the packages
-  usort($packages, "sortPackagesByName");
-
-  foreach($packages as $package){
-
-    $packageName = getTagContent($package, 'packageName');
-    $packageTag   = getTagContent($package, 'packageTag');
-    $packageStat  = getTagContent($package, 'packageStatus');
-
-    $tag = explode($packageName,$packageTag);
-    echo "\n<tr class='packageTableHeader'>";
-    echo "<td colspan='12'>".$packageName;
-    echo "<span style='font-weight:normal;font-size:0.9em'>[".$tag[1]."]</span>";
-    echo "<div style='font-weight:normal;font-size:0.8em;'>[".colorizePackageStat($packageStat)."]</div>";
-    echo "</td></tr>";
-
-    $minders = $package->get_elements_by_tagname('minder');
-
-    # $cpcCategories = sortByCPC($minders);
-    $cpcCategories = sortByCPC2($minders);
-
-    //Remove non-CPC jobs
-    $nonCPCKey = '[] [] []';
-    if(array_key_exists($nonCPCKey, $cpcCategories)){
-      unset($cpcCategories[$nonCPCKey]);
-    }
-
-    // Loop through the CPC categories
-    foreach($cpcCategories as $cpcName=>$mindersInCPC){ //cpcValue is an array of minder nodes
-
-      $buttonID = getTagContent($package, 'packageName')."_".$cpcName;
-      $jobsID   = "jobs_".getTagContent($package, 'packageName')."_".$cpcName;
-
-      // Print out the name of this CPC category
-      echo "<tr class='cpcTableHeader'>\n";
-      echo "<td style='text-decoration:underline' id='".$buttonID."' onclick='toggleStuff(\"".$buttonID."\",\"".$jobsID."\");'>[Show]</td>";
-      echo "<td>".$cpcName."</td>";
-
-      // Print out the status of this CPC category
-      $cpcStatus = cpcCategoryStatus($mindersInCPC);
-      echo "<td align='center'>".colorizeCPCStatus($cpcStatus)."</td>";
-
-      // Print out the number of jobs in this CPC category
-      $jobText = 'jobs';
-      if (sizeOf($mindersInCPC) == 1) $jobText = 'job';
-
-      if($cpcStatus == 'ongoing'){
-	echo "<td align='center'>ongoing</td>";
-      }
-      else if($cpcStatus == 'FAIL'){
-	$failedJobsInCPC = countNotOKJobs($mindersInCPC);	
-	echo "<td align='center'>".$failedJobsInCPC."/".sizeOf($mindersInCPC)." ".$jobText." not OK</td>";
-      }
-      else{
-	echo "<td align='center'>".sizeOf($mindersInCPC)."/".sizeOf($mindersInCPC)." ".$jobText." OK</td>";
-      }
-      
-      echo "</tr>";
-
-      // Print out a table of jobs in this CPC category
-      echo "<tr><td colspan='3'><div  align='center' style='display:none' id='".$jobsID."'>";
-
-      $jobsStatii = array();
-
-      echo "<table>\n";
-
-      echo "<tr class='tableRow2'><td>Name</td><td>JobGroup</td><td>Job status</td><td>Tests</td><td>Datasets</td><td>Hash</td></tr>";
-      foreach($mindersInCPC as $minder){
-	$jobName = getTagContent($minder, 'jobName');
-	$escapedJobName = rawurlencode($jobName);
-	$displayName = getTagContent($minder, 'jobDisplayName');
-	if ($displayName == ''){$displayName = $jobName;}
-	$jobID   = getTagContent($minder, 'jobID');
-	$group   = getTagContent($minder, 'jobGroup');
-	$hash   = getTagContent($minder, 'hashString');
-	if ($hash==''){$hash = 'n/a';}
-	// Now add the datasets
-	$datasetsList = $minder->get_elements_by_tagname('datasets');
-	$datasetsTag  = $datasetsList[0];
-	$datasets_holder = '';
-
-	if (isEmptyTag($datasetsTag)){
-	  $datasets = "n/a";
-	}
-	else{
-	  $datasets = "";
-	  foreach($datasetsTag->get_elements_by_tagname('dataset') as $dataset){
-	    $datasets .= trim($dataset->get_content())."\n";
-	  }
-
-	  $datasetsButtonID = $jobID."_datasetsButtonID";
-	  $datasetsID = $jobID."_datasetsID";
-	  $datasets_holder = "<span style='text-decoration:underline' id='".$datasetsButtonID."' onclick='toggleStuff(\"".$datasetsButtonID."\",\"".$datasetsID."\");'>[Show]</span>";
-	}
-
-	// Job status
-	$status = getTagContent($minder, 'status');
-	$jobsStatii[] = $status; //append to array
-
-	$statusClass = 'statusBad';
-	if ($status == 'success') $statusClass = 'statusOK';
-	if ($status == 'running') $statusClass =  'statusRunning';
-
-	// Post processing combined status
-	$postProcRes = getTagContent($minder, 'postProcessingResult');
-	$postProcClass = 'statusBad';
-	if ($postProcRes == 'success') $postProcClass = 'statusOK';
-	if ($postProcRes == 'running') $postProcClass =  'statusRunning';
-	if ($postProcRes == 'no tests') $postProcClass =  'statusNeutral';
-
-	// Ref. release used
-	$refRel = getTagContent($minder, 'refRelease');
-
-	// Now create the table row
-	echo "\n<tr class='tableRow'>\n";
-	//echo "<td valign='top'><a href='page2.php?xml=".$xmlFile."&package=".$packageName."&job=".$jobName."&id=".$jobID."'>".$displayName."</td>\n";
-	echo "<td valign='top'><a href='page2.php?xml=".$xmlFile."&package=".$packageName."&job=".$escapedJobName."&id=".$jobID."'>".$displayName."</a><br /><br /><span>[id:".$jobID."]</span></td>\n";
-
-	//echo "<td valign='top'>".$jobID."</td>\n";
-	echo "<td valign='top'>".$group."</td>\n";
-	echo "<td valign='top' class='".$statusClass."'>".$status."</td>\n";
-	echo "<td valign='top' class='".$postProcClass."'>".$postProcRes."</td>\n";
-
-	if ($datasets_holder == ''){
-	  echo "<td valign='top'>n/a</td>";
-	}
-	else{
-	  echo "<td valign='top'>";
-	  echo $datasets_holder;
-	  echo "<div align='center' style='display:none' id='".$datasetsID."'>".$datasets."</div>";
-	  echo "</td>\n";
-	}
-
-	echo "<td valign='top'>".$hash."</td>\n";
-	//echo "<td valign='top'>".$refRel."</td>\n";
-	echo "</tr>";
-
-      }
-
-      echo "</table>\n"; // end of minder table      
-      echo "</div></td></tr>";
-    }
-  }
-
-  echo "</table>";
-  
-}
-
-///////////////////////////////////////////////////////////////////
-// VIEW 3: DO THE BY-PACKAGE VIEW
-///////////////////////////////////////////////////////////////////
-function doByPackageView($packages){
-  global $xmlFile;
-
-  function sortByFramework($minders){
-    $atnMinders = array();
-    $rttMinders = array();
-    foreach($minders as $minder){
-      $group = getTagContent($minder, 'jobGroup');
-      if ($group=='AthenaATN'){$atnMinders[] = $minder;}
-      else {$rttMinders[] = $minder;}
-    }
-    return array($rttMinders, $atnMinders);
-  }
-
-  function countDoneJobs($minders){
-    $tot = 0;
-    foreach($minders as $minder){
-      if(hasChildTag($minder, 'done')){$tot++;}
-    }
-    return $tot;
-  }
-
-  function countOKJobs($minders){
-    $tot = 0;
-    foreach($minders as $minder){
-      if(getTagContent($minder, 'status') == 'success'){$tot++;}
-    }
-    return $tot;
-  }
-
-  function countBadJobs($minders){
-    $tot = 0;
-    foreach($minders as $minder){
-      if(getTagContent($minder, 'status') == 'error'){$tot++;}
-    }
-    return $tot;
-  }
-
-  function countOKTests($minders){
-    $tot = 0;
-    foreach($minders as $minder){
-      if(getTagContent($minder, 'postProcessingResult') == 'success'){$tot++;}
-    }
-    return $tot;
-  }
-
-  function colorizePackageStat($stat){
-    if(trim($stat)!='running' && trim($stat)!='done'){
-      return '<span style="color:#880000">'.$stat.'</span>';
-    }
-    return $stat;
-  }
-
-  echo "<table align='center' width='60%'>";
-
-  // Loop over each package
-  usort($packages, "sortPackagesByName");
-  foreach($packages as $package){
-    $packageName = getTagContent($package, 'packageName');
-    $packageCont = getTagContent($package, 'containerPackage');
-    $packageTag   = getTagContent($package, 'packageTag');
-    $packageStat  = getTagContent($package, 'packageStatus');
-
-    $cvsURL = cvsReposURL($packageCont, $packageTag);
-    $tag = explode($packageName,$packageTag);
-    echo "\n<tr class='packageTableHeader'><td colspan='12'>".$packageName;
-    echo " <span style='font-weight:normal;font-size:0.8em'>[".$tag[1]."]</span>";
-    echo "<span><a style='font-weight:normal;font-size:0.6em' href='".$cvsURL."'>[CVS]</a></span>";
-    echo " <div style='font-weight:normal;font-size:0.8em;'>[".colorizePackageStat($packageStat)."]</div></td></tr>";
-
-    // Get the jobs in this package
-    $minders = $package->get_elements_by_tagname('minder');
-  
-    // Split these jobs into ATN and RTT jobs
-    $sortedMinders = sortByFramework($minders);
-    $rttMinders = $sortedMinders[0];
-    $atnMinders = $sortedMinders[1];
-
-    $rttButtonID = "rttButton_".$packageName;
-    $atnButtonID = "atnButton_".$packageName;
-    $rttJobsID   = "rttJobs_".$packageName;
-    $atnJobsID   = "atnJobs_".$packageName;
-
-    $totRTTjobs  = sizeOf($rttMinders);
-    $doneRTTjobs = countDoneJobs($rttMinders);
-    $okRTTjobs   = countOKJobs($rttMinders);
-    $badRTTjobs  = countBadJobs($rttMinders);
-    $doneAthJobs = $okRTTjobs+$badRTTjobs;
-    $okRTTtests  = countOKTests($rttMinders);
-    $totRTTtests = $okRTTjobs;
-    echo "<tr class='frameworkHeader'>";
-    echo "<td>RTT jobs<br /><span id='".$rttButtonID."' onclick='toggleStuff(\"".$rttButtonID."\",\"".$rttJobsID."\");' class='showJobs'>[Show]</span></td>";
-    echo "<td class='stats'>";
-    echo "<table width='100%' border='0'>";
-    echo "<tr>";
-    echo "<td class='statHeader'>Run by RTT</td>";
-    echo "<td class='statTitle'>Jobs OK:&nbsp;</td><td class='statEntry'>".$okRTTjobs."/".$doneAthJobs."</td>";
-    echo "<td class='statTitle'>Jobs Done:&nbsp;</td><td class='statEntry'>".$doneRTTjobs."/".$totRTTjobs."</td>";
-    echo "<td class='statTitle'>Tests OK:&nbsp;</td><td class='statEntry'>".$okRTTtests."/".$totRTTtests."</td>";
-    echo "</tr>";
-    echo "</table>";
-    echo "</td>";
-    echo "</tr>";
-
-    // Print out a table of jobs in this CPC category
-    echo "<tr><td colspan='3'><div  align='center' style='display:none' id='".$rttJobsID."'>";
-
-    $jobsStatii = array();
-
-    echo "<table>\n";
-    echo "<tr class='tableRow2'><td>Name</td><td>JobGroup</td><td>Job status</td><td>PostProcessing status</td><td>Datasets</td><td>Hash</td></tr>";
-
-    foreach($rttMinders as $minder){
-      $jobName = getTagContent($minder, 'jobName');
-      $escapedJobName = rawurlencode($jobName);
-
-      $displayName = getTagContent($minder, 'jobDisplayName');
-      if ($displayName == ''){$displayName = $jobName;}
-      $jobID   = getTagContent($minder, 'jobID');
-      $group   = getTagContent($minder, 'jobGroup');
-      $hash    = getTagContent($minder, 'hashString');
-      if ($hash==''){$hash = 'n/a';}
-
-      $stateHistory = getTagContent($minder, 'stateHistory');
-      $currentState = getTagContent($minder, 'state');
-
-
-      // Now add the datasets
-      $datasetsList = $minder->get_elements_by_tagname('datasets');
-      $datasetsTag  = $datasetsList[0];
-      $datasets_holder = '';
-
-      if (isEmptyTag($datasetsTag)){
-	$datasets = "n/a";
-      }
-      else{
-	$datasets = "";
-	foreach($datasetsTag->get_elements_by_tagname('dataset') as $dataset){
-	  $datasets .= trim($dataset->get_content())."\n";
-	}
-
-	$datasetsButtonID = $jobID."_datasetsButtonID";
-	$datasetsID = $jobID."_datasetsID";
-	$datasets_holder = "<span style='text-decoration:underline' id='".$datasetsButtonID."' onclick='toggleStuff(\"".$datasetsButtonID."\",\"".$datasetsID."\");'>[Show]</span>";
-      }
-
-      // Job status
-      $status = getTagContent($minder, 'status');
-      $jobsStatii[] = $status; //append to array
-
-      $statusClass = 'statusBad';
-      if ($status == 'success') $statusClass = 'statusOK';
-      if ($status == 'running') $statusClass =  'statusRunning';
-
-      // Post processing combined status
-      $postProcRes = getTagContent($minder, 'postProcessingResult');
-      $postProcClass = 'statusBad';
-      if ($postProcRes == 'success') $postProcClass = 'statusOK';
-      if ($postProcRes == 'running') $postProcClass =  'statusRunning';
-      if ($postProcRes == 'no tests') $postProcClass =  'statusNeutral';
-      
-      // Ref. release used
-      $refRel = getTagContent($minder, 'refRelease');
-      
-      // Now create the table row
-      echo "\n<tr class='tableRow'>\n";
-      echo "<td valign='top'><a href='page2.php?xml=".$xmlFile."&package=".$packageName."&job=".$escapedJobName."&id=".$jobID."'>".$displayName."</a><br /><br /><span>[id:".$jobID."]</span></td>\n";
-      //echo "<td valign='top'>".$jobID."</td>\n";
-      echo "<td valign='top'>".$group."</td>\n";
-      echo "<td valign='top' class='".$statusClass."'>".$status."</td>\n";
-      echo "<td valign='top' class='".$postProcClass."'>".$postProcRes."</td>\n";
-
-      if($datasets_holder==''){
-	echo "<td valign='top'>n/a</td>\n";
-      }
-      else{
-	echo "<td valign='top'>";
-	echo $datasets_holder;
-	echo "<div align='center' style='display:none' id='".$datasetsID."'>".$datasets."</div>";
-	echo "</td>\n";
-      }
-
-      //echo "<td valign='top'>".$datasets."</td>\n";
-      echo "<td valign='top'>".$hash."</td>\n";
-      //echo "<td valign='top'>".$refRel."</td>\n";
-      echo "</tr>";
-      
-    }
-
-    echo "</table>\n"; // end of minder table      
-    echo "</div></td></tr>";
-
-    $totATNjobs  = sizeOf($atnMinders);
-    $doneATNjobs = countDoneJobs($atnMinders);
-    $okATNjobs   = countOKJobs($atnMinders);
-    if($totATNjobs != 0){
-      echo "<tr class='frameworkHeader'>";
-      echo "<td>ATN jobs<br /><span class='showJobs' id='".$atnButtonID."' onclick='toggleStuff(\"".$atnButtonID."\",\"".$atnJobsID."\");'>[Show]</span></td>";
-      echo "<td class='stats'>";
-      echo "<table width='100%' border='0'>";
-      echo "<tr>";
-      echo "<td class='statHeader'>Run by RTT</td>";
-      echo "<td class='statTitle'>Jobs OK:&nbsp;</td><td class='statEntry'>".$okATNjobs."/".$doneATNjobs."</td>";
-      echo "<td class='statTitle'>Jobs Done:&nbsp;</td><td class='statEntry'>".$doneATNjobs."/".$totATNjobs."</td>";
-      echo "</tr>";
-      echo "<tr>";
-      echo "<td class='statHeader'>Run by ATN</td>";
-      echo "<td class='statTitle'>Jobs OK:&nbsp;</td><td class='statEntry'>---</td>";
-      echo "<td class='statTitle'>Jobs Done:&nbsp;</td><td class='statEntry'>---</td>";
-      echo "</tr>";
-      echo "</table>";
-      echo "</td>";
-      echo "</tr>";
-
-      // Print out a table of jobs in this CPC category
-      echo "<tr><td colspan='3'><div  align='center' style='display:none' id='".$atnJobsID."'>";
-
-      $jobsStatii = array();
-
-      echo "<table>\n";
-
-      //echo "<tr class='tableRow2'><td>Name</td><td>ID</td><td>Hash</td><td>JobGroup</td><td>Datasets</td><td>Job status</td><td>Tests</td><td>Ref.Release</td></tr>";
-      echo "<tr class='tableRow2'><td>Name</td><td>JobGroup</td><td>Job status</td><td>Tests</td><td>Datasets</td><td>Hash</td></tr>";
-      foreach($atnMinders as $minder){
-	$jobName = getTagContent($minder, 'jobName');
-	$displayName = getTagContent($minder, 'jobDisplayName');
-	if ($displayName == ''){$displayName = $jobName;}
-	$jobID   = getTagContent($minder, 'jobID');
-	$hash   = getTagContent($minder, 'hashString');
-	if ($hash==''){$hash = 'n/a';}
-	$group   = getTagContent($minder, 'jobGroup');
-	$chainLength = getTagContent($minder, 'chainLength');
-	$chainSuccesses = getTagContent($minder, 'chainSuccesses');
-
-	// Now add the datasets
-	$datasetsList = $minder->get_elements_by_tagname('datasets');
-	$datasetsTag  = $datasetsList[0];
-	if (isEmptyTag($datasetsTag)){
-	  $datasets = "n/a";
-	}
-	else{
-	  $datasets = "";
-	  foreach($datasetsTag->get_elements_by_tagname('dataset') as $dataset){
-	    $datasets .= trim($dataset->get_content())."\n";
-	  }
-	}
-
-	// Job status
-	$status = getTagContent($minder, 'status');
-	$jobsStatii[] = $status; //append to array
-
-	$statusClass = 'statusBad';
-	if ($status == 'success') $statusClass = 'statusOK';
-	if ($status == 'running') $statusClass =  'statusRunning';
-
-	// Post processing combined status
-	$postProcRes = getTagContent($minder, 'postProcessingResult');
-	$postProcClass = 'statusBad';
-	if ($postProcRes == 'success') $postProcClass = 'statusOK';
-	if ($postProcRes == 'running') $postProcClass =  'statusRunning';
-	if ($postProcRes == 'no tests') $postProcClass =  'statusNeutral';
-      
-	// Ref. release used
-	$refRel = getTagContent($minder, 'refRelease');
-      
-	// Now create the table row
-	echo "\n<tr class='tableRow'>\n";
-	echo "<td valign='top'><a href='page2.php?xml=".$xmlFile."&package=".$packageName."&job=".$jobName."&id=".$jobID."'>".$displayName."</a><br /><br /><span>[id:".$jobID."]</span></td>\n";
-	//echo "<td>".$jobID."</td>\n";
-	echo "<td valign='top'>".$group."</td>\n";
-	echo "<td valign='top' class='".$statusClass."'>".$status."</td>\n";
-	echo "<td valign='top' class='".$postProcClass."'>".$postProcRes."</td>\n";
-	echo "<td>".$datasets."</td>\n";
-	echo "<td>".$hash."</td>\n";
-	//echo "<td valign='top'>".$refRel."</td>\n";
-	echo "</tr>";
-      }
-
-      echo "</table>\n"; // end of minder table      
-      echo "</div></td></tr>";
-
-    }
-
-  }
-
-  echo "</table>";
-}
-
-?>
-
-<br />
-<a class="pageLink" href="page0.php">Back to home page</a>
-<br />
-</body>
-</html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/page2.php b/Tools/RunTimeTester/share/web/page2.php
deleted file mode 100644
index 59e5d98795e523e180d710785178de509fc5ab62..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/page2.php
+++ /dev/null
@@ -1,331 +0,0 @@
-<?
-ob_start("ob_gzhandler");
-header("Content-type: text/html; charset=utf-8");
-?>
-
-<html>
-<head>
-<style>
-body{font-family:Trebuchet,sans-serif;background-color:#eeeeff;text-align:center;margin: 0 5 5 5;color:black}
-.pageHeader{font-size:1.8em;font-weight:bold;}
-.sectionHeader{font-size:1em;font-weight:bold;}
-.statusOK{color:green}
-.statusBad{color:red}
-.statusRunning{color:brown}
-#admin{font-size:0.7em;font-weight:bold;letter-spacing:2px;background-color:#8888df;color:white;font-family:verdana, sans-serif}
-a.pageLink{font-family:arial;font-size:0.8em;color:#777777;text-decoration:underline}
-a.pageLink:hover{font-family:arial;font-size:0.8em;color:black;text-decoration:none}
-a.normalKeepFile{font-family:arial;font-size:0.8em;color:black;text-decoration:underline;font-weight:bold}
-a.normalKeepFile:hover{font-family:arial;font-size:0.8em;color:orange;text-decoration:none}
-a.linkExpired{font-family:arial;font-size:0.8em;color:#999999;text-decoration:underline;font-style:italic}
-span.linkWillExpire{font-family:arial;font-size:0.7em;color:#777777;text-decoration:none;font-style:italic}
-table{background-color:#ddddff}
-.header{font-weight:bold;background-color:#aaaaff}
-.entry{font-size:0.85em;font-weight:normal}
-ul{list-style-type:none;padding:2px}
-#mainTable{padding:0;margin:auto;}
-</style>
-<script src="prototype.js" type="text/javascript"></script>
-<script src="effects.js" type="text/javascript"></script>
-<script type="text/javascript">
-<!--
-function toggleThis(srcID, tgtID){
-  var el = document.getElementById(srcID);
-  if(document.getElementById(tgtID).style.display == 'none'){
-    new Effect.Appear(tgtID);
-    el.innerHTML = "Click to Hide";
-  }
-  else{
-     new Effect.DropOut(tgtID);
-     el.innerHTML = "Click to Show";
-  }
-}
-
-// -->
-</script>
-</head>
-
-<body>
-<div id='admin' onclick="toggleThis('admin','topFrame');">Click to Show</div>
-<div id='topFrame' style="display:none">
-<script src="RTTpage1_LinksAtTopOfPage.js" type="text/JavaScript"></script>
-<script src="messageOfTheDay.js" type="text/JavaScript"></script>
-</div>
-<br />
-
-<!------------------------------------------------------------->
-<!-- START OF THE PHP SECTION                                -->
-<!------------------------------------------------------------->
-
-<?
-include("tools.php");
-
-$xmlFile = $_GET['xml'];
-
-if (!file_exists($xmlFile)){
-  echo "No data available yet.";
-  return;
-}
-
-$packageName = $_GET['package'];
-$jobName = $_GET['job'];
-$jobID = $_GET['id'];
-
-$rootNode = getRootNodeFromXMLfile($xmlFile);
-$overviewNodes = $rootNode->get_elements_by_tagname('overview');
-$overviewNode = $overviewNodes[sizeOf($overviewNodes)-1];
-
-// Find the corresponding minder node 
-$minder = getSingleTagWithChildTag($rootNode->get_elements_by_tagname('minder'), 'identifiedName', $jobName.$jobID);
-if (sizeOf($minder)==0){echo "No information available.";return;}
-
-$displayName = getTagContent($minder, 'jobDisplayName');
-if ($displayName == ''){$displayName = $jobName;}
-
-$hash   = getTagContent($minder, 'hashString');
-if ($hash==''){$hash = 'n/a'; }
-
-$jobGroup = getTagContent($minder, 'jobGroup');
-
-// Output the top of page overview info
-$releaseName = getTagContent($overviewNode, 'releaseName');
-$runType     = getTagContent($overviewNode, 'runType');
-$platform    = getTagContent($overviewNode, 'targetCMTCONFIG');
-$topProj     = getTagContent($overviewNode, 'topProject');
-$otherProj   = getTagContent($overviewNode, 'otherProject');
-echo "<div class='pageHeader'>";
-if($otherProj==''){
-  echo $releaseName."/".$runType." --- ".$platform." --- ".$topProj;
-}
-else{
-  echo $releaseName."/".$runType." --- ".$platform." --- ".$topProj.'/'.$otherProj;
-}
-echo "</div>";
-
-$allKeepFiles = $minder->get_elements_by_tagname('deliveredkeepfiles');
-
-$allKeepFilesTag  = $allKeepFiles[0];
-if (isEmptyTag($allKeepFilesTag)){
-  $keepFiles = array("None available");
-}
-else{
-  $keepFiles = $allKeepFilesTag->get_elements_by_tagname('keepfile_newstyle');
-}
-
-echo "<br />";
-
-//Is this an error minder with content in <errors> tag?
-$errors = getTagContent($minder, 'errors');
-if($errors==''){
-  // store for later the job tests
-  $allTests = $minder->get_elements_by_tagname('testResults');
-  $allTestsTag  = $allTests[0];
-  $jobTests = $allTestsTag->get_elements_by_tagname('testResult');
-}
-
-?>
-
-<!------------------------------------------------------------->
-<!-- END OF THE PHP SECTION                                  -->
-<!------------------------------------------------------------->
-
-<table id='mainTable' cellpadding="0" cellspacing="0">
-   <tr>
-     <td valign="top">
-        <table>
-           <tr class='header'><td>Overview</td></tr>
-           <tr>
-              <td>
-                  <table>
-                     <tr class='entry'><td align="right">Package: </td><? echo "<td class='entry'>".$packageName."</td>";?></tr>
-                     <tr class='entry'><td align="right">Job name: </td><? echo "<td class='entry'>".$displayName."</td>";?></tr>
-                     <tr class='entry'><td align="right">Jobgroup: </td><? echo "<td class='entry'>".$jobGroup."</td>";?></tr>
-                     <tr class='entry'><td align="right">Job ID: </td><? echo "<td class='entry'>".$jobID."</td>";?></tr>
-                     <tr class='entry'><td align="right">Hash ID: </td><? echo "<td class='entry'>".$hash."</td>";?></tr>
-                  </table>
-              </td>
-           </tr>
-
-           <tr class='header'><td>Results</td></tr>
-           <tr>
-              <td>
-                  <table>
-		     <?
-                      $jobStat = trim(getTagContent($minder, 'status'));
-                      if ($jobStat == 'success'){$jobStatClass = 'statusOK';}
-                      else if ($jobStat == 'running'){$jobStatClass = 'statusRunning';}
-                      else {$jobStatClass = 'statusBad';}
-                     ?>
-                     <tr class='entry'><td align="right">Job Status: </td><? echo "<td class='".$jobStatClass."'>".$jobStat."</td>";?></tr>
-
-		     <?
-		     if (!isset($jobTests)){
-		       echo "<tr class='entry'><td align='right'>Tests: </td><td>n/a</td></tr>";
-		     } 
-                     else if (sizeOf($jobTests) == 0){
-		       echo "<tr class='entry'><td align='right'>Tests: </td><td>No tests</td></tr>";
-		     }
-                     else{ 
-		       foreach ($jobTests as $test){
-			 echo "<tr>";
-			 if(trim(getTagContent($test, 'testStatus')) == 0){
-			   $testStatus = "success";
-			   $testClass  = "statusOK";
-			 }
-			 else{
-			   $testStatus = "error";
-			   $testClass  = "statusBad";
-			 }
-			 echo "<td class='entry'>".trim(getTagContent($test, 'testName')).": </td>";
-			 echo "<td class='entry ".$testClass."'>".$testStatus."</td>";
-			 echo "</tr>";
-		       }
-		     }
-                     ?>
-                  </table>
-              </td>
-           </tr>
-
-           <tr class='header'><td>Classification</td></tr>
-           <tr>
-              <td>
-		     <table class='entry'>
-		     <? 
-		     $displayClassList = $minder->get_elements_by_tagname('displayClass');
-                     if (sizeOf($displayClassList) == 0){
-		       echo "<tr><td>None available</td></tr>";
-                     }
-                     else{
-                        $displayClass = $displayClassList[0];
-			$displayProcessList = $minder->get_elements_by_tagname('displayProcess');
-			//$displayProcess = $displayProcessList[0];
-			$displayCompList = $minder->get_elements_by_tagname('displayComponent');
-			//$displayComp = $displayCompList[0];
-                     
-			if (isEmptyTag($displayClass)){
-			  echo "<tr><td>None available</td></tr>";
-			}
-			else{
-			  echo "<tr><td align='right'>Class: </td><td>".trim($displayClass->get_content())."</td></tr>";
-			  foreach($displayProcessList as $dp){
-			    echo "<tr><td align='right'>Process: </td><td>".trim($dp->get_content())."</td></tr>";
-			  }
-			  foreach($displayCompList as $dc){
-			    echo "<tr><td align='right'>Component: </td><td>".trim($dc->get_content())."</td></tr>";
-			  }
-			}
-		     }
-                     ?>
-                   </table>
-              </td>
-           </tr>
-           <tr class='header'><td>Documentation</td></tr>
-           <tr>
-              <td>
-		   <table class='entry'>
-		   <?
-		   $docList = $minder->get_elements_by_tagname('jobDocString');
-                   if(sizeOf($docList)==0){
-		     echo "<tr><td style='font-style:italic'>[No doc available]</td></tr>";
-		   }
-                   else{
-		     $doc = $docList[0];
-		     if (isEmptyTag($doc)){
-		       echo "<tr><td style='font-style:italic'>[No doc available]</td></tr>";
-		     }
-		     else{
-		       $docContent = trim($doc->get_content());
-		       echo "<tr><td align='left' style='width:300px'>".$docContent."</td></tr>";
-		     }
-		   }
-
-		   $docLinkList = $minder->get_elements_by_tagname('jobDocURL');
-                   if(sizeOf($docLinkList)==0){
-		     echo "<tr><td style='font-style:italic'>[No doclink available]</td></tr>";
-		   }
-                   else{
-		     $doclink = $docLinkList[0];
-		     if (isEmptyTag($doclink)){
-		       echo "<tr><td style='font-style:italic'>[No doclink available]</td></tr>";
-		     }
-		     else{
-		       echo "<tr><td align='left' style='width:300px'><a class='pageLink' href='".trim($doclink->get_content())."'>More info</a></td></tr>";
-		     }
-		   }
-		   ?>
-		   </table>
-              </td>
-           </tr>
-           <tr class='header'><td>Auditor Stats</td></tr>
-           <tr>
-               <td>
-                  <table class='entry'>
-                     <tr><td align="right">CPU time (s): </td><td><?echo trim(getTagContent($minder, 'cpuTime'));?></td></tr>
-                     <tr><td align="right">CPU time 2000 (s): </td><td><?echo trim(getTagContent($minder, 'cpuTime2000'));?></td></tr>
-                     <tr><td align="right">Walltime (s): </td><td><?echo trim(getTagContent($minder, 'wallTime'));?></td></tr>
-                     <tr><td align="right">Memory: </td><td><?echo trim(getTagContent($minder, 'mem'));?></td></tr>
-                     <tr><td align="right">V-Memory: </td><td><?echo trim(getTagContent($minder, 'vmem'));?></td></tr>
-                     <tr><td align="right">Batch Status: </td><td><?echo trim(getTagContent($minder, 'batchStatus'));?></td></tr>
-                  </table>
-               </td>
-           </tr>
-        </table>
-     </td>
-     <td valign="top">
-        <ul>
-          <?
-
-		     //$base = getTagContent($minder,'resultsPath');
-if(trim($jobStat)=='batchError'){
-  echo "Job had batch error.";
-}
-else if ($errors != ''){
-  echo "RTT rejected this job. Here's the traceback:<br />";
-  echo "<font color='red'><pre>".$errors."</pre></font>";
-}
-else{
-  usort($keepFiles, "sortKeepFiles");
-  if (sizeOf($keepFiles)==1 && $keepFiles[0]=='None available'){
-    echo "<li>None available</li>\n";
-  }
-  else{
-    foreach($keepFiles as $keepFile){
-      $linkColor    = getAttributeValue($keepFile, 'displayColor');
-      $md5sum       = getAttributeValue($keepFile, 'md5sum');
-      $link_info    = getAttributeValue($keepFile, 'info');
-      $keepFileName = trim($keepFile->get_content());
-      $linkTgt      = $keepFileName;
-
-      if ($md5sum != ''){
-	if (file_exists($linkTgt)){
-	  $md5sumNow = md5_file($linkTgt);
-	  if ($md5sum != $md5sumNow){
-	    echo "<li><a class='linkExpired'>".basename($keepFileName)." [LINK EXPIRED]</a></li>\n";
-	    continue;
-	  }
-	  else{
-	    echo "<li><a class='normalKeepFile' title='".$link_info."' style='color:".$linkColor."' href='".$linkTgt."'>".basename($keepFileName)."</a><span class='linkWillExpire'>&nbsp;[LINK WILL EXPIRE]</span></li>\n";
-	    continue;
-	  }
-	}
-	else{
-	  echo "<li><a class='linkExpired'>".basename($keepFileName)." [LINK EXPIRED]</a></li>\n";
-	  continue;
-	}
-      }
-
-      echo "<li><a class='normalKeepFile' title='".$link_info."' style='color:".$linkColor."' href='".$linkTgt."'>".basename($keepFileName)."</a></li>\n";
-    }
-  }
-}
-?>
-          </ul>
-     </td>
-   </tr>
-</table>
-
-<br />
-<a class="pageLink" href="page0.php">Back to home page</a>
-<br />
-</body>
-</html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/prototype.js b/Tools/RunTimeTester/share/web/prototype.js
deleted file mode 100644
index 0476b8fdccc0717b3be623ff464748d1c5e751e6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/prototype.js
+++ /dev/null
@@ -1,2515 +0,0 @@
-/*  Prototype JavaScript framework, version 1.5.0
- *  (c) 2005-2007 Sam Stephenson
- *
- *  Prototype is freely distributable under the terms of an MIT-style license.
- *  For details, see the Prototype web site: http://prototype.conio.net/
- *
-/*--------------------------------------------------------------------------*/
-
-var Prototype = {
-  Version: '1.5.0',
-  BrowserFeatures: {
-    XPath: !!document.evaluate
-  },
-
-  ScriptFragment: '(?:<script.*?>)((\n|\r|.)*?)(?:<\/script>)',
-  emptyFunction: function() {},
-  K: function(x) { return x }
-}
-
-var Class = {
-  create: function() {
-    return function() {
-      this.initialize.apply(this, arguments);
-    }
-  }
-}
-
-var Abstract = new Object();
-
-Object.extend = function(destination, source) {
-  for (var property in source) {
-    destination[property] = source[property];
-  }
-  return destination;
-}
-
-Object.extend(Object, {
-  inspect: function(object) {
-    try {
-      if (object === undefined) return 'undefined';
-      if (object === null) return 'null';
-      return object.inspect ? object.inspect() : object.toString();
-    } catch (e) {
-      if (e instanceof RangeError) return '...';
-      throw e;
-    }
-  },
-
-  keys: function(object) {
-    var keys = [];
-    for (var property in object)
-      keys.push(property);
-    return keys;
-  },
-
-  values: function(object) {
-    var values = [];
-    for (var property in object)
-      values.push(object[property]);
-    return values;
-  },
-
-  clone: function(object) {
-    return Object.extend({}, object);
-  }
-});
-
-Function.prototype.bind = function() {
-  var __method = this, args = $A(arguments), object = args.shift();
-  return function() {
-    return __method.apply(object, args.concat($A(arguments)));
-  }
-}
-
-Function.prototype.bindAsEventListener = function(object) {
-  var __method = this, args = $A(arguments), object = args.shift();
-  return function(event) {
-    return __method.apply(object, [( event || window.event)].concat(args).concat($A(arguments)));
-  }
-}
-
-Object.extend(Number.prototype, {
-  toColorPart: function() {
-    var digits = this.toString(16);
-    if (this < 16) return '0' + digits;
-    return digits;
-  },
-
-  succ: function() {
-    return this + 1;
-  },
-
-  times: function(iterator) {
-    $R(0, this, true).each(iterator);
-    return this;
-  }
-});
-
-var Try = {
-  these: function() {
-    var returnValue;
-
-    for (var i = 0, length = arguments.length; i < length; i++) {
-      var lambda = arguments[i];
-      try {
-        returnValue = lambda();
-        break;
-      } catch (e) {}
-    }
-
-    return returnValue;
-  }
-}
-
-/*--------------------------------------------------------------------------*/
-
-var PeriodicalExecuter = Class.create();
-PeriodicalExecuter.prototype = {
-  initialize: function(callback, frequency) {
-    this.callback = callback;
-    this.frequency = frequency;
-    this.currentlyExecuting = false;
-
-    this.registerCallback();
-  },
-
-  registerCallback: function() {
-    this.timer = setInterval(this.onTimerEvent.bind(this), this.frequency * 1000);
-  },
-
-  stop: function() {
-    if (!this.timer) return;
-    clearInterval(this.timer);
-    this.timer = null;
-  },
-
-  onTimerEvent: function() {
-    if (!this.currentlyExecuting) {
-      try {
-        this.currentlyExecuting = true;
-        this.callback(this);
-      } finally {
-        this.currentlyExecuting = false;
-      }
-    }
-  }
-}
-String.interpret = function(value){
-  return value == null ? '' : String(value);
-}
-
-Object.extend(String.prototype, {
-  gsub: function(pattern, replacement) {
-    var result = '', source = this, match;
-    replacement = arguments.callee.prepareReplacement(replacement);
-
-    while (source.length > 0) {
-      if (match = source.match(pattern)) {
-        result += source.slice(0, match.index);
-        result += String.interpret(replacement(match));
-        source  = source.slice(match.index + match[0].length);
-      } else {
-        result += source, source = '';
-      }
-    }
-    return result;
-  },
-
-  sub: function(pattern, replacement, count) {
-    replacement = this.gsub.prepareReplacement(replacement);
-    count = count === undefined ? 1 : count;
-
-    return this.gsub(pattern, function(match) {
-      if (--count < 0) return match[0];
-      return replacement(match);
-    });
-  },
-
-  scan: function(pattern, iterator) {
-    this.gsub(pattern, iterator);
-    return this;
-  },
-
-  truncate: function(length, truncation) {
-    length = length || 30;
-    truncation = truncation === undefined ? '...' : truncation;
-    return this.length > length ?
-      this.slice(0, length - truncation.length) + truncation : this;
-  },
-
-  strip: function() {
-    return this.replace(/^\s+/, '').replace(/\s+$/, '');
-  },
-
-  stripTags: function() {
-    return this.replace(/<\/?[^>]+>/gi, '');
-  },
-
-  stripScripts: function() {
-    return this.replace(new RegExp(Prototype.ScriptFragment, 'img'), '');
-  },
-
-  extractScripts: function() {
-    var matchAll = new RegExp(Prototype.ScriptFragment, 'img');
-    var matchOne = new RegExp(Prototype.ScriptFragment, 'im');
-    return (this.match(matchAll) || []).map(function(scriptTag) {
-      return (scriptTag.match(matchOne) || ['', ''])[1];
-    });
-  },
-
-  evalScripts: function() {
-    return this.extractScripts().map(function(script) { return eval(script) });
-  },
-
-  escapeHTML: function() {
-    var div = document.createElement('div');
-    var text = document.createTextNode(this);
-    div.appendChild(text);
-    return div.innerHTML;
-  },
-
-  unescapeHTML: function() {
-    var div = document.createElement('div');
-    div.innerHTML = this.stripTags();
-    return div.childNodes[0] ? (div.childNodes.length > 1 ?
-      $A(div.childNodes).inject('',function(memo,node){ return memo+node.nodeValue }) :
-      div.childNodes[0].nodeValue) : '';
-  },
-
-  toQueryParams: function(separator) {
-    var match = this.strip().match(/([^?#]*)(#.*)?$/);
-    if (!match) return {};
-
-    return match[1].split(separator || '&').inject({}, function(hash, pair) {
-      if ((pair = pair.split('='))[0]) {
-        var name = decodeURIComponent(pair[0]);
-        var value = pair[1] ? decodeURIComponent(pair[1]) : undefined;
-
-        if (hash[name] !== undefined) {
-          if (hash[name].constructor != Array)
-            hash[name] = [hash[name]];
-          if (value) hash[name].push(value);
-        }
-        else hash[name] = value;
-      }
-      return hash;
-    });
-  },
-
-  toArray: function() {
-    return this.split('');
-  },
-
-  succ: function() {
-    return this.slice(0, this.length - 1) +
-      String.fromCharCode(this.charCodeAt(this.length - 1) + 1);
-  },
-
-  camelize: function() {
-    var parts = this.split('-'), len = parts.length;
-    if (len == 1) return parts[0];
-
-    var camelized = this.charAt(0) == '-'
-      ? parts[0].charAt(0).toUpperCase() + parts[0].substring(1)
-      : parts[0];
-
-    for (var i = 1; i < len; i++)
-      camelized += parts[i].charAt(0).toUpperCase() + parts[i].substring(1);
-
-    return camelized;
-  },
-
-  capitalize: function(){
-    return this.charAt(0).toUpperCase() + this.substring(1).toLowerCase();
-  },
-
-  underscore: function() {
-    return this.gsub(/::/, '/').gsub(/([A-Z]+)([A-Z][a-z])/,'#{1}_#{2}').gsub(/([a-z\d])([A-Z])/,'#{1}_#{2}').gsub(/-/,'_').toLowerCase();
-  },
-
-  dasherize: function() {
-    return this.gsub(/_/,'-');
-  },
-
-  inspect: function(useDoubleQuotes) {
-    var escapedString = this.replace(/\\/g, '\\\\');
-    if (useDoubleQuotes)
-      return '"' + escapedString.replace(/"/g, '\\"') + '"';
-    else
-      return "'" + escapedString.replace(/'/g, '\\\'') + "'";
-  }
-});
-
-String.prototype.gsub.prepareReplacement = function(replacement) {
-  if (typeof replacement == 'function') return replacement;
-  var template = new Template(replacement);
-  return function(match) { return template.evaluate(match) };
-}
-
-String.prototype.parseQuery = String.prototype.toQueryParams;
-
-var Template = Class.create();
-Template.Pattern = /(^|.|\r|\n)(#\{(.*?)\})/;
-Template.prototype = {
-  initialize: function(template, pattern) {
-    this.template = template.toString();
-    this.pattern  = pattern || Template.Pattern;
-  },
-
-  evaluate: function(object) {
-    return this.template.gsub(this.pattern, function(match) {
-      var before = match[1];
-      if (before == '\\') return match[2];
-      return before + String.interpret(object[match[3]]);
-    });
-  }
-}
-
-var $break    = new Object();
-var $continue = new Object();
-
-var Enumerable = {
-  each: function(iterator) {
-    var index = 0;
-    try {
-      this._each(function(value) {
-        try {
-          iterator(value, index++);
-        } catch (e) {
-          if (e != $continue) throw e;
-        }
-      });
-    } catch (e) {
-      if (e != $break) throw e;
-    }
-    return this;
-  },
-
-  eachSlice: function(number, iterator) {
-    var index = -number, slices = [], array = this.toArray();
-    while ((index += number) < array.length)
-      slices.push(array.slice(index, index+number));
-    return slices.map(iterator);
-  },
-
-  all: function(iterator) {
-    var result = true;
-    this.each(function(value, index) {
-      result = result && !!(iterator || Prototype.K)(value, index);
-      if (!result) throw $break;
-    });
-    return result;
-  },
-
-  any: function(iterator) {
-    var result = false;
-    this.each(function(value, index) {
-      if (result = !!(iterator || Prototype.K)(value, index))
-        throw $break;
-    });
-    return result;
-  },
-
-  collect: function(iterator) {
-    var results = [];
-    this.each(function(value, index) {
-      results.push((iterator || Prototype.K)(value, index));
-    });
-    return results;
-  },
-
-  detect: function(iterator) {
-    var result;
-    this.each(function(value, index) {
-      if (iterator(value, index)) {
-        result = value;
-        throw $break;
-      }
-    });
-    return result;
-  },
-
-  findAll: function(iterator) {
-    var results = [];
-    this.each(function(value, index) {
-      if (iterator(value, index))
-        results.push(value);
-    });
-    return results;
-  },
-
-  grep: function(pattern, iterator) {
-    var results = [];
-    this.each(function(value, index) {
-      var stringValue = value.toString();
-      if (stringValue.match(pattern))
-        results.push((iterator || Prototype.K)(value, index));
-    })
-    return results;
-  },
-
-  include: function(object) {
-    var found = false;
-    this.each(function(value) {
-      if (value == object) {
-        found = true;
-        throw $break;
-      }
-    });
-    return found;
-  },
-
-  inGroupsOf: function(number, fillWith) {
-    fillWith = fillWith === undefined ? null : fillWith;
-    return this.eachSlice(number, function(slice) {
-      while(slice.length < number) slice.push(fillWith);
-      return slice;
-    });
-  },
-
-  inject: function(memo, iterator) {
-    this.each(function(value, index) {
-      memo = iterator(memo, value, index);
-    });
-    return memo;
-  },
-
-  invoke: function(method) {
-    var args = $A(arguments).slice(1);
-    return this.map(function(value) {
-      return value[method].apply(value, args);
-    });
-  },
-
-  max: function(iterator) {
-    var result;
-    this.each(function(value, index) {
-      value = (iterator || Prototype.K)(value, index);
-      if (result == undefined || value >= result)
-        result = value;
-    });
-    return result;
-  },
-
-  min: function(iterator) {
-    var result;
-    this.each(function(value, index) {
-      value = (iterator || Prototype.K)(value, index);
-      if (result == undefined || value < result)
-        result = value;
-    });
-    return result;
-  },
-
-  partition: function(iterator) {
-    var trues = [], falses = [];
-    this.each(function(value, index) {
-      ((iterator || Prototype.K)(value, index) ?
-        trues : falses).push(value);
-    });
-    return [trues, falses];
-  },
-
-  pluck: function(property) {
-    var results = [];
-    this.each(function(value, index) {
-      results.push(value[property]);
-    });
-    return results;
-  },
-
-  reject: function(iterator) {
-    var results = [];
-    this.each(function(value, index) {
-      if (!iterator(value, index))
-        results.push(value);
-    });
-    return results;
-  },
-
-  sortBy: function(iterator) {
-    return this.map(function(value, index) {
-      return {value: value, criteria: iterator(value, index)};
-    }).sort(function(left, right) {
-      var a = left.criteria, b = right.criteria;
-      return a < b ? -1 : a > b ? 1 : 0;
-    }).pluck('value');
-  },
-
-  toArray: function() {
-    return this.map();
-  },
-
-  zip: function() {
-    var iterator = Prototype.K, args = $A(arguments);
-    if (typeof args.last() == 'function')
-      iterator = args.pop();
-
-    var collections = [this].concat(args).map($A);
-    return this.map(function(value, index) {
-      return iterator(collections.pluck(index));
-    });
-  },
-
-  size: function() {
-    return this.toArray().length;
-  },
-
-  inspect: function() {
-    return '#<Enumerable:' + this.toArray().inspect() + '>';
-  }
-}
-
-Object.extend(Enumerable, {
-  map:     Enumerable.collect,
-  find:    Enumerable.detect,
-  select:  Enumerable.findAll,
-  member:  Enumerable.include,
-  entries: Enumerable.toArray
-});
-var $A = Array.from = function(iterable) {
-  if (!iterable) return [];
-  if (iterable.toArray) {
-    return iterable.toArray();
-  } else {
-    var results = [];
-    for (var i = 0, length = iterable.length; i < length; i++)
-      results.push(iterable[i]);
-    return results;
-  }
-}
-
-Object.extend(Array.prototype, Enumerable);
-
-if (!Array.prototype._reverse)
-  Array.prototype._reverse = Array.prototype.reverse;
-
-Object.extend(Array.prototype, {
-  _each: function(iterator) {
-    for (var i = 0, length = this.length; i < length; i++)
-      iterator(this[i]);
-  },
-
-  clear: function() {
-    this.length = 0;
-    return this;
-  },
-
-  first: function() {
-    return this[0];
-  },
-
-  last: function() {
-    return this[this.length - 1];
-  },
-
-  compact: function() {
-    return this.select(function(value) {
-      return value != null;
-    });
-  },
-
-  flatten: function() {
-    return this.inject([], function(array, value) {
-      return array.concat(value && value.constructor == Array ?
-        value.flatten() : [value]);
-    });
-  },
-
-  without: function() {
-    var values = $A(arguments);
-    return this.select(function(value) {
-      return !values.include(value);
-    });
-  },
-
-  indexOf: function(object) {
-    for (var i = 0, length = this.length; i < length; i++)
-      if (this[i] == object) return i;
-    return -1;
-  },
-
-  reverse: function(inline) {
-    return (inline !== false ? this : this.toArray())._reverse();
-  },
-
-  reduce: function() {
-    return this.length > 1 ? this : this[0];
-  },
-
-  uniq: function() {
-    return this.inject([], function(array, value) {
-      return array.include(value) ? array : array.concat([value]);
-    });
-  },
-
-  clone: function() {
-    return [].concat(this);
-  },
-
-  size: function() {
-    return this.length;
-  },
-
-  inspect: function() {
-    return '[' + this.map(Object.inspect).join(', ') + ']';
-  }
-});
-
-Array.prototype.toArray = Array.prototype.clone;
-
-function $w(string){
-  string = string.strip();
-  return string ? string.split(/\s+/) : [];
-}
-
-if(window.opera){
-  Array.prototype.concat = function(){
-    var array = [];
-    for(var i = 0, length = this.length; i < length; i++) array.push(this[i]);
-    for(var i = 0, length = arguments.length; i < length; i++) {
-      if(arguments[i].constructor == Array) {
-        for(var j = 0, arrayLength = arguments[i].length; j < arrayLength; j++)
-          array.push(arguments[i][j]);
-      } else {
-        array.push(arguments[i]);
-      }
-    }
-    return array;
-  }
-}
-var Hash = function(obj) {
-  Object.extend(this, obj || {});
-};
-
-Object.extend(Hash, {
-  toQueryString: function(obj) {
-    var parts = [];
-
-	  this.prototype._each.call(obj, function(pair) {
-      if (!pair.key) return;
-
-      if (pair.value && pair.value.constructor == Array) {
-        var values = pair.value.compact();
-        if (values.length < 2) pair.value = values.reduce();
-        else {
-        	key = encodeURIComponent(pair.key);
-          values.each(function(value) {
-            value = value != undefined ? encodeURIComponent(value) : '';
-            parts.push(key + '=' + encodeURIComponent(value));
-          });
-          return;
-        }
-      }
-      if (pair.value == undefined) pair[1] = '';
-      parts.push(pair.map(encodeURIComponent).join('='));
-	  });
-
-    return parts.join('&');
-  }
-});
-
-Object.extend(Hash.prototype, Enumerable);
-Object.extend(Hash.prototype, {
-  _each: function(iterator) {
-    for (var key in this) {
-      var value = this[key];
-      if (value && value == Hash.prototype[key]) continue;
-
-      var pair = [key, value];
-      pair.key = key;
-      pair.value = value;
-      iterator(pair);
-    }
-  },
-
-  keys: function() {
-    return this.pluck('key');
-  },
-
-  values: function() {
-    return this.pluck('value');
-  },
-
-  merge: function(hash) {
-    return $H(hash).inject(this, function(mergedHash, pair) {
-      mergedHash[pair.key] = pair.value;
-      return mergedHash;
-    });
-  },
-
-  remove: function() {
-    var result;
-    for(var i = 0, length = arguments.length; i < length; i++) {
-      var value = this[arguments[i]];
-      if (value !== undefined){
-        if (result === undefined) result = value;
-        else {
-          if (result.constructor != Array) result = [result];
-          result.push(value)
-        }
-      }
-      delete this[arguments[i]];
-    }
-    return result;
-  },
-
-  toQueryString: function() {
-    return Hash.toQueryString(this);
-  },
-
-  inspect: function() {
-    return '#<Hash:{' + this.map(function(pair) {
-      return pair.map(Object.inspect).join(': ');
-    }).join(', ') + '}>';
-  }
-});
-
-function $H(object) {
-  if (object && object.constructor == Hash) return object;
-  return new Hash(object);
-};
-ObjectRange = Class.create();
-Object.extend(ObjectRange.prototype, Enumerable);
-Object.extend(ObjectRange.prototype, {
-  initialize: function(start, end, exclusive) {
-    this.start = start;
-    this.end = end;
-    this.exclusive = exclusive;
-  },
-
-  _each: function(iterator) {
-    var value = this.start;
-    while (this.include(value)) {
-      iterator(value);
-      value = value.succ();
-    }
-  },
-
-  include: function(value) {
-    if (value < this.start)
-      return false;
-    if (this.exclusive)
-      return value < this.end;
-    return value <= this.end;
-  }
-});
-
-var $R = function(start, end, exclusive) {
-  return new ObjectRange(start, end, exclusive);
-}
-
-var Ajax = {
-  getTransport: function() {
-    return Try.these(
-      function() {return new XMLHttpRequest()},
-      function() {return new ActiveXObject('Msxml2.XMLHTTP')},
-      function() {return new ActiveXObject('Microsoft.XMLHTTP')}
-    ) || false;
-  },
-
-  activeRequestCount: 0
-}
-
-Ajax.Responders = {
-  responders: [],
-
-  _each: function(iterator) {
-    this.responders._each(iterator);
-  },
-
-  register: function(responder) {
-    if (!this.include(responder))
-      this.responders.push(responder);
-  },
-
-  unregister: function(responder) {
-    this.responders = this.responders.without(responder);
-  },
-
-  dispatch: function(callback, request, transport, json) {
-    this.each(function(responder) {
-      if (typeof responder[callback] == 'function') {
-        try {
-          responder[callback].apply(responder, [request, transport, json]);
-        } catch (e) {}
-      }
-    });
-  }
-};
-
-Object.extend(Ajax.Responders, Enumerable);
-
-Ajax.Responders.register({
-  onCreate: function() {
-    Ajax.activeRequestCount++;
-  },
-  onComplete: function() {
-    Ajax.activeRequestCount--;
-  }
-});
-
-Ajax.Base = function() {};
-Ajax.Base.prototype = {
-  setOptions: function(options) {
-    this.options = {
-      method:       'post',
-      asynchronous: true,
-      contentType:  'application/x-www-form-urlencoded',
-      encoding:     'UTF-8',
-      parameters:   ''
-    }
-    Object.extend(this.options, options || {});
-
-    this.options.method = this.options.method.toLowerCase();
-    if (typeof this.options.parameters == 'string')
-      this.options.parameters = this.options.parameters.toQueryParams();
-  }
-}
-
-Ajax.Request = Class.create();
-Ajax.Request.Events =
-  ['Uninitialized', 'Loading', 'Loaded', 'Interactive', 'Complete'];
-
-Ajax.Request.prototype = Object.extend(new Ajax.Base(), {
-  _complete: false,
-
-  initialize: function(url, options) {
-    this.transport = Ajax.getTransport();
-    this.setOptions(options);
-    this.request(url);
-  },
-
-  request: function(url) {
-    this.url = url;
-    this.method = this.options.method;
-    var params = this.options.parameters;
-
-    if (!['get', 'post'].include(this.method)) {
-      // simulate other verbs over post
-      params['_method'] = this.method;
-      this.method = 'post';
-    }
-
-    params = Hash.toQueryString(params);
-    if (params && /Konqueror|Safari|KHTML/.test(navigator.userAgent)) params += '&_='
-
-    // when GET, append parameters to URL
-    if (this.method == 'get' && params)
-      this.url += (this.url.indexOf('?') > -1 ? '&' : '?') + params;
-
-    try {
-      Ajax.Responders.dispatch('onCreate', this, this.transport);
-
-      this.transport.open(this.method.toUpperCase(), this.url,
-        this.options.asynchronous);
-
-      if (this.options.asynchronous)
-        setTimeout(function() { this.respondToReadyState(1) }.bind(this), 10);
-
-      this.transport.onreadystatechange = this.onStateChange.bind(this);
-      this.setRequestHeaders();
-
-      var body = this.method == 'post' ? (this.options.postBody || params) : null;
-
-      this.transport.send(body);
-
-      /* Force Firefox to handle ready state 4 for synchronous requests */
-      if (!this.options.asynchronous && this.transport.overrideMimeType)
-        this.onStateChange();
-
-    }
-    catch (e) {
-      this.dispatchException(e);
-    }
-  },
-
-  onStateChange: function() {
-    var readyState = this.transport.readyState;
-    if (readyState > 1 && !((readyState == 4) && this._complete))
-      this.respondToReadyState(this.transport.readyState);
-  },
-
-  setRequestHeaders: function() {
-    var headers = {
-      'X-Requested-With': 'XMLHttpRequest',
-      'X-Prototype-Version': Prototype.Version,
-      'Accept': 'text/javascript, text/html, application/xml, text/xml, */*'
-    };
-
-    if (this.method == 'post') {
-      headers['Content-type'] = this.options.contentType +
-        (this.options.encoding ? '; charset=' + this.options.encoding : '');
-
-      /* Force "Connection: close" for older Mozilla browsers to work
-       * around a bug where XMLHttpRequest sends an incorrect
-       * Content-length header. See Mozilla Bugzilla #246651.
-       */
-      if (this.transport.overrideMimeType &&
-          (navigator.userAgent.match(/Gecko\/(\d{4})/) || [0,2005])[1] < 2005)
-            headers['Connection'] = 'close';
-    }
-
-    // user-defined headers
-    if (typeof this.options.requestHeaders == 'object') {
-      var extras = this.options.requestHeaders;
-
-      if (typeof extras.push == 'function')
-        for (var i = 0, length = extras.length; i < length; i += 2)
-          headers[extras[i]] = extras[i+1];
-      else
-        $H(extras).each(function(pair) { headers[pair.key] = pair.value });
-    }
-
-    for (var name in headers)
-      this.transport.setRequestHeader(name, headers[name]);
-  },
-
-  success: function() {
-    return !this.transport.status
-        || (this.transport.status >= 200 && this.transport.status < 300);
-  },
-
-  respondToReadyState: function(readyState) {
-    var state = Ajax.Request.Events[readyState];
-    var transport = this.transport, json = this.evalJSON();
-
-    if (state == 'Complete') {
-      try {
-        this._complete = true;
-        (this.options['on' + this.transport.status]
-         || this.options['on' + (this.success() ? 'Success' : 'Failure')]
-         || Prototype.emptyFunction)(transport, json);
-      } catch (e) {
-        this.dispatchException(e);
-      }
-
-      if ((this.getHeader('Content-type') || 'text/javascript').strip().
-        match(/^(text|application)\/(x-)?(java|ecma)script(;.*)?$/i))
-          this.evalResponse();
-    }
-
-    try {
-      (this.options['on' + state] || Prototype.emptyFunction)(transport, json);
-      Ajax.Responders.dispatch('on' + state, this, transport, json);
-    } catch (e) {
-      this.dispatchException(e);
-    }
-
-    if (state == 'Complete') {
-      // avoid memory leak in MSIE: clean up
-      this.transport.onreadystatechange = Prototype.emptyFunction;
-    }
-  },
-
-  getHeader: function(name) {
-    try {
-      return this.transport.getResponseHeader(name);
-    } catch (e) { return null }
-  },
-
-  evalJSON: function() {
-    try {
-      var json = this.getHeader('X-JSON');
-      return json ? eval('(' + json + ')') : null;
-    } catch (e) { return null }
-  },
-
-  evalResponse: function() {
-    try {
-      return eval(this.transport.responseText);
-    } catch (e) {
-      this.dispatchException(e);
-    }
-  },
-
-  dispatchException: function(exception) {
-    (this.options.onException || Prototype.emptyFunction)(this, exception);
-    Ajax.Responders.dispatch('onException', this, exception);
-  }
-});
-
-Ajax.Updater = Class.create();
-
-Object.extend(Object.extend(Ajax.Updater.prototype, Ajax.Request.prototype), {
-  initialize: function(container, url, options) {
-    this.container = {
-      success: (container.success || container),
-      failure: (container.failure || (container.success ? null : container))
-    }
-
-    this.transport = Ajax.getTransport();
-    this.setOptions(options);
-
-    var onComplete = this.options.onComplete || Prototype.emptyFunction;
-    this.options.onComplete = (function(transport, param) {
-      this.updateContent();
-      onComplete(transport, param);
-    }).bind(this);
-
-    this.request(url);
-  },
-
-  updateContent: function() {
-    var receiver = this.container[this.success() ? 'success' : 'failure'];
-    var response = this.transport.responseText;
-
-    if (!this.options.evalScripts) response = response.stripScripts();
-
-    if (receiver = $(receiver)) {
-      if (this.options.insertion)
-        new this.options.insertion(receiver, response);
-      else
-        receiver.update(response);
-    }
-
-    if (this.success()) {
-      if (this.onComplete)
-        setTimeout(this.onComplete.bind(this), 10);
-    }
-  }
-});
-
-Ajax.PeriodicalUpdater = Class.create();
-Ajax.PeriodicalUpdater.prototype = Object.extend(new Ajax.Base(), {
-  initialize: function(container, url, options) {
-    this.setOptions(options);
-    this.onComplete = this.options.onComplete;
-
-    this.frequency = (this.options.frequency || 2);
-    this.decay = (this.options.decay || 1);
-
-    this.updater = {};
-    this.container = container;
-    this.url = url;
-
-    this.start();
-  },
-
-  start: function() {
-    this.options.onComplete = this.updateComplete.bind(this);
-    this.onTimerEvent();
-  },
-
-  stop: function() {
-    this.updater.options.onComplete = undefined;
-    clearTimeout(this.timer);
-    (this.onComplete || Prototype.emptyFunction).apply(this, arguments);
-  },
-
-  updateComplete: function(request) {
-    if (this.options.decay) {
-      this.decay = (request.responseText == this.lastText ?
-        this.decay * this.options.decay : 1);
-
-      this.lastText = request.responseText;
-    }
-    this.timer = setTimeout(this.onTimerEvent.bind(this),
-      this.decay * this.frequency * 1000);
-  },
-
-  onTimerEvent: function() {
-    this.updater = new Ajax.Updater(this.container, this.url, this.options);
-  }
-});
-function $(element) {
-  if (arguments.length > 1) {
-    for (var i = 0, elements = [], length = arguments.length; i < length; i++)
-      elements.push($(arguments[i]));
-    return elements;
-  }
-  if (typeof element == 'string')
-    element = document.getElementById(element);
-  return Element.extend(element);
-}
-
-if (Prototype.BrowserFeatures.XPath) {
-  document._getElementsByXPath = function(expression, parentElement) {
-    var results = [];
-    var query = document.evaluate(expression, $(parentElement) || document,
-      null, XPathResult.ORDERED_NODE_SNAPSHOT_TYPE, null);
-    for (var i = 0, length = query.snapshotLength; i < length; i++)
-      results.push(query.snapshotItem(i));
-    return results;
-  };
-}
-
-document.getElementsByClassName = function(className, parentElement) {
-  if (Prototype.BrowserFeatures.XPath) {
-    var q = ".//*[contains(concat(' ', @class, ' '), ' " + className + " ')]";
-    return document._getElementsByXPath(q, parentElement);
-  } else {
-    var children = ($(parentElement) || document.body).getElementsByTagName('*');
-    var elements = [], child;
-    for (var i = 0, length = children.length; i < length; i++) {
-      child = children[i];
-      if (Element.hasClassName(child, className))
-        elements.push(Element.extend(child));
-    }
-    return elements;
-  }
-};
-
-/*--------------------------------------------------------------------------*/
-
-if (!window.Element)
-  var Element = new Object();
-
-Element.extend = function(element) {
-  if (!element || _nativeExtensions || element.nodeType == 3) return element;
-
-  if (!element._extended && element.tagName && element != window) {
-    var methods = Object.clone(Element.Methods), cache = Element.extend.cache;
-
-    if (element.tagName == 'FORM')
-      Object.extend(methods, Form.Methods);
-    if (['INPUT', 'TEXTAREA', 'SELECT'].include(element.tagName))
-      Object.extend(methods, Form.Element.Methods);
-
-    Object.extend(methods, Element.Methods.Simulated);
-
-    for (var property in methods) {
-      var value = methods[property];
-      if (typeof value == 'function' && !(property in element))
-        element[property] = cache.findOrStore(value);
-    }
-  }
-
-  element._extended = true;
-  return element;
-};
-
-Element.extend.cache = {
-  findOrStore: function(value) {
-    return this[value] = this[value] || function() {
-      return value.apply(null, [this].concat($A(arguments)));
-    }
-  }
-};
-
-Element.Methods = {
-  visible: function(element) {
-    return $(element).style.display != 'none';
-  },
-
-  toggle: function(element) {
-    element = $(element);
-    Element[Element.visible(element) ? 'hide' : 'show'](element);
-    return element;
-  },
-
-  hide: function(element) {
-    $(element).style.display = 'none';
-    return element;
-  },
-
-  show: function(element) {
-    $(element).style.display = '';
-    return element;
-  },
-
-  remove: function(element) {
-    element = $(element);
-    element.parentNode.removeChild(element);
-    return element;
-  },
-
-  update: function(element, html) {
-    html = typeof html == 'undefined' ? '' : html.toString();
-    $(element).innerHTML = html.stripScripts();
-    setTimeout(function() {html.evalScripts()}, 10);
-    return element;
-  },
-
-  replace: function(element, html) {
-    element = $(element);
-    html = typeof html == 'undefined' ? '' : html.toString();
-    if (element.outerHTML) {
-      element.outerHTML = html.stripScripts();
-    } else {
-      var range = element.ownerDocument.createRange();
-      range.selectNodeContents(element);
-      element.parentNode.replaceChild(
-        range.createContextualFragment(html.stripScripts()), element);
-    }
-    setTimeout(function() {html.evalScripts()}, 10);
-    return element;
-  },
-
-  inspect: function(element) {
-    element = $(element);
-    var result = '<' + element.tagName.toLowerCase();
-    $H({'id': 'id', 'className': 'class'}).each(function(pair) {
-      var property = pair.first(), attribute = pair.last();
-      var value = (element[property] || '').toString();
-      if (value) result += ' ' + attribute + '=' + value.inspect(true);
-    });
-    return result + '>';
-  },
-
-  recursivelyCollect: function(element, property) {
-    element = $(element);
-    var elements = [];
-    while (element = element[property])
-      if (element.nodeType == 1)
-        elements.push(Element.extend(element));
-    return elements;
-  },
-
-  ancestors: function(element) {
-    return $(element).recursivelyCollect('parentNode');
-  },
-
-  descendants: function(element) {
-    return $A($(element).getElementsByTagName('*'));
-  },
-
-  immediateDescendants: function(element) {
-    if (!(element = $(element).firstChild)) return [];
-    while (element && element.nodeType != 1) element = element.nextSibling;
-    if (element) return [element].concat($(element).nextSiblings());
-    return [];
-  },
-
-  previousSiblings: function(element) {
-    return $(element).recursivelyCollect('previousSibling');
-  },
-
-  nextSiblings: function(element) {
-    return $(element).recursivelyCollect('nextSibling');
-  },
-
-  siblings: function(element) {
-    element = $(element);
-    return element.previousSiblings().reverse().concat(element.nextSiblings());
-  },
-
-  match: function(element, selector) {
-    if (typeof selector == 'string')
-      selector = new Selector(selector);
-    return selector.match($(element));
-  },
-
-  up: function(element, expression, index) {
-    return Selector.findElement($(element).ancestors(), expression, index);
-  },
-
-  down: function(element, expression, index) {
-    return Selector.findElement($(element).descendants(), expression, index);
-  },
-
-  previous: function(element, expression, index) {
-    return Selector.findElement($(element).previousSiblings(), expression, index);
-  },
-
-  next: function(element, expression, index) {
-    return Selector.findElement($(element).nextSiblings(), expression, index);
-  },
-
-  getElementsBySelector: function() {
-    var args = $A(arguments), element = $(args.shift());
-    return Selector.findChildElements(element, args);
-  },
-
-  getElementsByClassName: function(element, className) {
-    return document.getElementsByClassName(className, element);
-  },
-
-  readAttribute: function(element, name) {
-    element = $(element);
-    if (document.all && !window.opera) {
-      var t = Element._attributeTranslations;
-      if (t.values[name]) return t.values[name](element, name);
-      if (t.names[name])  name = t.names[name];
-      var attribute = element.attributes[name];
-      if(attribute) return attribute.nodeValue;
-    }
-    return element.getAttribute(name);
-  },
-
-  getHeight: function(element) {
-    return $(element).getDimensions().height;
-  },
-
-  getWidth: function(element) {
-    return $(element).getDimensions().width;
-  },
-
-  classNames: function(element) {
-    return new Element.ClassNames(element);
-  },
-
-  hasClassName: function(element, className) {
-    if (!(element = $(element))) return;
-    var elementClassName = element.className;
-    if (elementClassName.length == 0) return false;
-    if (elementClassName == className ||
-        elementClassName.match(new RegExp("(^|\\s)" + className + "(\\s|$)")))
-      return true;
-    return false;
-  },
-
-  addClassName: function(element, className) {
-    if (!(element = $(element))) return;
-    Element.classNames(element).add(className);
-    return element;
-  },
-
-  removeClassName: function(element, className) {
-    if (!(element = $(element))) return;
-    Element.classNames(element).remove(className);
-    return element;
-  },
-
-  toggleClassName: function(element, className) {
-    if (!(element = $(element))) return;
-    Element.classNames(element)[element.hasClassName(className) ? 'remove' : 'add'](className);
-    return element;
-  },
-
-  observe: function() {
-    Event.observe.apply(Event, arguments);
-    return $A(arguments).first();
-  },
-
-  stopObserving: function() {
-    Event.stopObserving.apply(Event, arguments);
-    return $A(arguments).first();
-  },
-
-  // removes whitespace-only text node children
-  cleanWhitespace: function(element) {
-    element = $(element);
-    var node = element.firstChild;
-    while (node) {
-      var nextNode = node.nextSibling;
-      if (node.nodeType == 3 && !/\S/.test(node.nodeValue))
-        element.removeChild(node);
-      node = nextNode;
-    }
-    return element;
-  },
-
-  empty: function(element) {
-    return $(element).innerHTML.match(/^\s*$/);
-  },
-
-  descendantOf: function(element, ancestor) {
-    element = $(element), ancestor = $(ancestor);
-    while (element = element.parentNode)
-      if (element == ancestor) return true;
-    return false;
-  },
-
-  scrollTo: function(element) {
-    element = $(element);
-    var pos = Position.cumulativeOffset(element);
-    window.scrollTo(pos[0], pos[1]);
-    return element;
-  },
-
-  getStyle: function(element, style) {
-    element = $(element);
-    if (['float','cssFloat'].include(style))
-      style = (typeof element.style.styleFloat != 'undefined' ? 'styleFloat' : 'cssFloat');
-    style = style.camelize();
-    var value = element.style[style];
-    if (!value) {
-      if (document.defaultView && document.defaultView.getComputedStyle) {
-        var css = document.defaultView.getComputedStyle(element, null);
-        value = css ? css[style] : null;
-      } else if (element.currentStyle) {
-        value = element.currentStyle[style];
-      }
-    }
-
-    if((value == 'auto') && ['width','height'].include(style) && (element.getStyle('display') != 'none'))
-      value = element['offset'+style.capitalize()] + 'px';
-
-    if (window.opera && ['left', 'top', 'right', 'bottom'].include(style))
-      if (Element.getStyle(element, 'position') == 'static') value = 'auto';
-    if(style == 'opacity') {
-      if(value) return parseFloat(value);
-      if(value = (element.getStyle('filter') || '').match(/alpha\(opacity=(.*)\)/))
-        if(value[1]) return parseFloat(value[1]) / 100;
-      return 1.0;
-    }
-    return value == 'auto' ? null : value;
-  },
-
-  setStyle: function(element, style) {
-    element = $(element);
-    for (var name in style) {
-      var value = style[name];
-      if(name == 'opacity') {
-        if (value == 1) {
-          value = (/Gecko/.test(navigator.userAgent) &&
-            !/Konqueror|Safari|KHTML/.test(navigator.userAgent)) ? 0.999999 : 1.0;
-          if(/MSIE/.test(navigator.userAgent) && !window.opera)
-            element.style.filter = element.getStyle('filter').replace(/alpha\([^\)]*\)/gi,'');
-        } else if(value === '') {
-          if(/MSIE/.test(navigator.userAgent) && !window.opera)
-            element.style.filter = element.getStyle('filter').replace(/alpha\([^\)]*\)/gi,'');
-        } else {
-          if(value < 0.00001) value = 0;
-          if(/MSIE/.test(navigator.userAgent) && !window.opera)
-            element.style.filter = element.getStyle('filter').replace(/alpha\([^\)]*\)/gi,'') +
-              'alpha(opacity='+value*100+')';
-        }
-      } else if(['float','cssFloat'].include(name)) name = (typeof element.style.styleFloat != 'undefined') ? 'styleFloat' : 'cssFloat';
-      element.style[name.camelize()] = value;
-    }
-    return element;
-  },
-
-  getDimensions: function(element) {
-    element = $(element);
-    var display = $(element).getStyle('display');
-    if (display != 'none' && display != null) // Safari bug
-      return {width: element.offsetWidth, height: element.offsetHeight};
-
-    // All *Width and *Height properties give 0 on elements with display none,
-    // so enable the element temporarily
-    var els = element.style;
-    var originalVisibility = els.visibility;
-    var originalPosition = els.position;
-    var originalDisplay = els.display;
-    els.visibility = 'hidden';
-    els.position = 'absolute';
-    els.display = 'block';
-    var originalWidth = element.clientWidth;
-    var originalHeight = element.clientHeight;
-    els.display = originalDisplay;
-    els.position = originalPosition;
-    els.visibility = originalVisibility;
-    return {width: originalWidth, height: originalHeight};
-  },
-
-  makePositioned: function(element) {
-    element = $(element);
-    var pos = Element.getStyle(element, 'position');
-    if (pos == 'static' || !pos) {
-      element._madePositioned = true;
-      element.style.position = 'relative';
-      // Opera returns the offset relative to the positioning context, when an
-      // element is position relative but top and left have not been defined
-      if (window.opera) {
-        element.style.top = 0;
-        element.style.left = 0;
-      }
-    }
-    return element;
-  },
-
-  undoPositioned: function(element) {
-    element = $(element);
-    if (element._madePositioned) {
-      element._madePositioned = undefined;
-      element.style.position =
-        element.style.top =
-        element.style.left =
-        element.style.bottom =
-        element.style.right = '';
-    }
-    return element;
-  },
-
-  makeClipping: function(element) {
-    element = $(element);
-    if (element._overflow) return element;
-    element._overflow = element.style.overflow || 'auto';
-    if ((Element.getStyle(element, 'overflow') || 'visible') != 'hidden')
-      element.style.overflow = 'hidden';
-    return element;
-  },
-
-  undoClipping: function(element) {
-    element = $(element);
-    if (!element._overflow) return element;
-    element.style.overflow = element._overflow == 'auto' ? '' : element._overflow;
-    element._overflow = null;
-    return element;
-  }
-};
-
-Object.extend(Element.Methods, {childOf: Element.Methods.descendantOf});
-
-Element._attributeTranslations = {};
-
-Element._attributeTranslations.names = {
-  colspan:   "colSpan",
-  rowspan:   "rowSpan",
-  valign:    "vAlign",
-  datetime:  "dateTime",
-  accesskey: "accessKey",
-  tabindex:  "tabIndex",
-  enctype:   "encType",
-  maxlength: "maxLength",
-  readonly:  "readOnly",
-  longdesc:  "longDesc"
-};
-
-Element._attributeTranslations.values = {
-  _getAttr: function(element, attribute) {
-    return element.getAttribute(attribute, 2);
-  },
-
-  _flag: function(element, attribute) {
-    return $(element).hasAttribute(attribute) ? attribute : null;
-  },
-
-  style: function(element) {
-    return element.style.cssText.toLowerCase();
-  },
-
-  title: function(element) {
-    var node = element.getAttributeNode('title');
-    return node.specified ? node.nodeValue : null;
-  }
-};
-
-Object.extend(Element._attributeTranslations.values, {
-  href: Element._attributeTranslations.values._getAttr,
-  src:  Element._attributeTranslations.values._getAttr,
-  disabled: Element._attributeTranslations.values._flag,
-  checked:  Element._attributeTranslations.values._flag,
-  readonly: Element._attributeTranslations.values._flag,
-  multiple: Element._attributeTranslations.values._flag
-});
-
-Element.Methods.Simulated = {
-  hasAttribute: function(element, attribute) {
-    var t = Element._attributeTranslations;
-    attribute = t.names[attribute] || attribute;
-    return $(element).getAttributeNode(attribute).specified;
-  }
-};
-
-// IE is missing .innerHTML support for TABLE-related elements
-if (document.all && !window.opera){
-  Element.Methods.update = function(element, html) {
-    element = $(element);
-    html = typeof html == 'undefined' ? '' : html.toString();
-    var tagName = element.tagName.toUpperCase();
-    if (['THEAD','TBODY','TR','TD'].include(tagName)) {
-      var div = document.createElement('div');
-      switch (tagName) {
-        case 'THEAD':
-        case 'TBODY':
-          div.innerHTML = '<table><tbody>' +  html.stripScripts() + '</tbody></table>';
-          depth = 2;
-          break;
-        case 'TR':
-          div.innerHTML = '<table><tbody><tr>' +  html.stripScripts() + '</tr></tbody></table>';
-          depth = 3;
-          break;
-        case 'TD':
-          div.innerHTML = '<table><tbody><tr><td>' +  html.stripScripts() + '</td></tr></tbody></table>';
-          depth = 4;
-      }
-      $A(element.childNodes).each(function(node){
-        element.removeChild(node)
-      });
-      depth.times(function(){ div = div.firstChild });
-
-      $A(div.childNodes).each(
-        function(node){ element.appendChild(node) });
-    } else {
-      element.innerHTML = html.stripScripts();
-    }
-    setTimeout(function() {html.evalScripts()}, 10);
-    return element;
-  }
-};
-
-Object.extend(Element, Element.Methods);
-
-var _nativeExtensions = false;
-
-if(/Konqueror|Safari|KHTML/.test(navigator.userAgent))
-  ['', 'Form', 'Input', 'TextArea', 'Select'].each(function(tag) {
-    var className = 'HTML' + tag + 'Element';
-    if(window[className]) return;
-    var klass = window[className] = {};
-    klass.prototype = document.createElement(tag ? tag.toLowerCase() : 'div').__proto__;
-  });
-
-Element.addMethods = function(methods) {
-  Object.extend(Element.Methods, methods || {});
-
-  function copy(methods, destination, onlyIfAbsent) {
-    onlyIfAbsent = onlyIfAbsent || false;
-    var cache = Element.extend.cache;
-    for (var property in methods) {
-      var value = methods[property];
-      if (!onlyIfAbsent || !(property in destination))
-        destination[property] = cache.findOrStore(value);
-    }
-  }
-
-  if (typeof HTMLElement != 'undefined') {
-    copy(Element.Methods, HTMLElement.prototype);
-    copy(Element.Methods.Simulated, HTMLElement.prototype, true);
-    copy(Form.Methods, HTMLFormElement.prototype);
-    [HTMLInputElement, HTMLTextAreaElement, HTMLSelectElement].each(function(klass) {
-      copy(Form.Element.Methods, klass.prototype);
-    });
-    _nativeExtensions = true;
-  }
-}
-
-var Toggle = new Object();
-Toggle.display = Element.toggle;
-
-/*--------------------------------------------------------------------------*/
-
-Abstract.Insertion = function(adjacency) {
-  this.adjacency = adjacency;
-}
-
-Abstract.Insertion.prototype = {
-  initialize: function(element, content) {
-    this.element = $(element);
-    this.content = content.stripScripts();
-
-    if (this.adjacency && this.element.insertAdjacentHTML) {
-      try {
-        this.element.insertAdjacentHTML(this.adjacency, this.content);
-      } catch (e) {
-        var tagName = this.element.tagName.toUpperCase();
-        if (['TBODY', 'TR'].include(tagName)) {
-          this.insertContent(this.contentFromAnonymousTable());
-        } else {
-          throw e;
-        }
-      }
-    } else {
-      this.range = this.element.ownerDocument.createRange();
-      if (this.initializeRange) this.initializeRange();
-      this.insertContent([this.range.createContextualFragment(this.content)]);
-    }
-
-    setTimeout(function() {content.evalScripts()}, 10);
-  },
-
-  contentFromAnonymousTable: function() {
-    var div = document.createElement('div');
-    div.innerHTML = '<table><tbody>' + this.content + '</tbody></table>';
-    return $A(div.childNodes[0].childNodes[0].childNodes);
-  }
-}
-
-var Insertion = new Object();
-
-Insertion.Before = Class.create();
-Insertion.Before.prototype = Object.extend(new Abstract.Insertion('beforeBegin'), {
-  initializeRange: function() {
-    this.range.setStartBefore(this.element);
-  },
-
-  insertContent: function(fragments) {
-    fragments.each((function(fragment) {
-      this.element.parentNode.insertBefore(fragment, this.element);
-    }).bind(this));
-  }
-});
-
-Insertion.Top = Class.create();
-Insertion.Top.prototype = Object.extend(new Abstract.Insertion('afterBegin'), {
-  initializeRange: function() {
-    this.range.selectNodeContents(this.element);
-    this.range.collapse(true);
-  },
-
-  insertContent: function(fragments) {
-    fragments.reverse(false).each((function(fragment) {
-      this.element.insertBefore(fragment, this.element.firstChild);
-    }).bind(this));
-  }
-});
-
-Insertion.Bottom = Class.create();
-Insertion.Bottom.prototype = Object.extend(new Abstract.Insertion('beforeEnd'), {
-  initializeRange: function() {
-    this.range.selectNodeContents(this.element);
-    this.range.collapse(this.element);
-  },
-
-  insertContent: function(fragments) {
-    fragments.each((function(fragment) {
-      this.element.appendChild(fragment);
-    }).bind(this));
-  }
-});
-
-Insertion.After = Class.create();
-Insertion.After.prototype = Object.extend(new Abstract.Insertion('afterEnd'), {
-  initializeRange: function() {
-    this.range.setStartAfter(this.element);
-  },
-
-  insertContent: function(fragments) {
-    fragments.each((function(fragment) {
-      this.element.parentNode.insertBefore(fragment,
-        this.element.nextSibling);
-    }).bind(this));
-  }
-});
-
-/*--------------------------------------------------------------------------*/
-
-Element.ClassNames = Class.create();
-Element.ClassNames.prototype = {
-  initialize: function(element) {
-    this.element = $(element);
-  },
-
-  _each: function(iterator) {
-    this.element.className.split(/\s+/).select(function(name) {
-      return name.length > 0;
-    })._each(iterator);
-  },
-
-  set: function(className) {
-    this.element.className = className;
-  },
-
-  add: function(classNameToAdd) {
-    if (this.include(classNameToAdd)) return;
-    this.set($A(this).concat(classNameToAdd).join(' '));
-  },
-
-  remove: function(classNameToRemove) {
-    if (!this.include(classNameToRemove)) return;
-    this.set($A(this).without(classNameToRemove).join(' '));
-  },
-
-  toString: function() {
-    return $A(this).join(' ');
-  }
-};
-
-Object.extend(Element.ClassNames.prototype, Enumerable);
-var Selector = Class.create();
-Selector.prototype = {
-  initialize: function(expression) {
-    this.params = {classNames: []};
-    this.expression = expression.toString().strip();
-    this.parseExpression();
-    this.compileMatcher();
-  },
-
-  parseExpression: function() {
-    function abort(message) { throw 'Parse error in selector: ' + message; }
-
-    if (this.expression == '')  abort('empty expression');
-
-    var params = this.params, expr = this.expression, match, modifier, clause, rest;
-    while (match = expr.match(/^(.*)\[([a-z0-9_:-]+?)(?:([~\|!]?=)(?:"([^"]*)"|([^\]\s]*)))?\]$/i)) {
-      params.attributes = params.attributes || [];
-      params.attributes.push({name: match[2], operator: match[3], value: match[4] || match[5] || ''});
-      expr = match[1];
-    }
-
-    if (expr == '*') return this.params.wildcard = true;
-
-    while (match = expr.match(/^([^a-z0-9_-])?([a-z0-9_-]+)(.*)/i)) {
-      modifier = match[1], clause = match[2], rest = match[3];
-      switch (modifier) {
-        case '#':       params.id = clause; break;
-        case '.':       params.classNames.push(clause); break;
-        case '':
-        case undefined: params.tagName = clause.toUpperCase(); break;
-        default:        abort(expr.inspect());
-      }
-      expr = rest;
-    }
-
-    if (expr.length > 0) abort(expr.inspect());
-  },
-
-  buildMatchExpression: function() {
-    var params = this.params, conditions = [], clause;
-
-    if (params.wildcard)
-      conditions.push('true');
-    if (clause = params.id)
-      conditions.push('element.readAttribute("id") == ' + clause.inspect());
-    if (clause = params.tagName)
-      conditions.push('element.tagName.toUpperCase() == ' + clause.inspect());
-    if ((clause = params.classNames).length > 0)
-      for (var i = 0, length = clause.length; i < length; i++)
-        conditions.push('element.hasClassName(' + clause[i].inspect() + ')');
-    if (clause = params.attributes) {
-      clause.each(function(attribute) {
-        var value = 'element.readAttribute(' + attribute.name.inspect() + ')';
-        var splitValueBy = function(delimiter) {
-          return value + ' && ' + value + '.split(' + delimiter.inspect() + ')';
-        }
-
-        switch (attribute.operator) {
-          case '=':       conditions.push(value + ' == ' + attribute.value.inspect()); break;
-          case '~=':      conditions.push(splitValueBy(' ') + '.include(' + attribute.value.inspect() + ')'); break;
-          case '|=':      conditions.push(
-                            splitValueBy('-') + '.first().toUpperCase() == ' + attribute.value.toUpperCase().inspect()
-                          ); break;
-          case '!=':      conditions.push(value + ' != ' + attribute.value.inspect()); break;
-          case '':
-          case undefined: conditions.push('element.hasAttribute(' + attribute.name.inspect() + ')'); break;
-          default:        throw 'Unknown operator ' + attribute.operator + ' in selector';
-        }
-      });
-    }
-
-    return conditions.join(' && ');
-  },
-
-  compileMatcher: function() {
-    this.match = new Function('element', 'if (!element.tagName) return false; \
-      element = $(element); \
-      return ' + this.buildMatchExpression());
-  },
-
-  findElements: function(scope) {
-    var element;
-
-    if (element = $(this.params.id))
-      if (this.match(element))
-        if (!scope || Element.childOf(element, scope))
-          return [element];
-
-    scope = (scope || document).getElementsByTagName(this.params.tagName || '*');
-
-    var results = [];
-    for (var i = 0, length = scope.length; i < length; i++)
-      if (this.match(element = scope[i]))
-        results.push(Element.extend(element));
-
-    return results;
-  },
-
-  toString: function() {
-    return this.expression;
-  }
-}
-
-Object.extend(Selector, {
-  matchElements: function(elements, expression) {
-    var selector = new Selector(expression);
-    return elements.select(selector.match.bind(selector)).map(Element.extend);
-  },
-
-  findElement: function(elements, expression, index) {
-    if (typeof expression == 'number') index = expression, expression = false;
-    return Selector.matchElements(elements, expression || '*')[index || 0];
-  },
-
-  findChildElements: function(element, expressions) {
-    return expressions.map(function(expression) {
-      return expression.match(/[^\s"]+(?:"[^"]*"[^\s"]+)*/g).inject([null], function(results, expr) {
-        var selector = new Selector(expr);
-        return results.inject([], function(elements, result) {
-          return elements.concat(selector.findElements(result || element));
-        });
-      });
-    }).flatten();
-  }
-});
-
-function $$() {
-  return Selector.findChildElements(document, $A(arguments));
-}
-var Form = {
-  reset: function(form) {
-    $(form).reset();
-    return form;
-  },
-
-  serializeElements: function(elements, getHash) {
-    var data = elements.inject({}, function(result, element) {
-      if (!element.disabled && element.name) {
-        var key = element.name, value = $(element).getValue();
-        if (value != undefined) {
-          if (result[key]) {
-            if (result[key].constructor != Array) result[key] = [result[key]];
-            result[key].push(value);
-          }
-          else result[key] = value;
-        }
-      }
-      return result;
-    });
-
-    return getHash ? data : Hash.toQueryString(data);
-  }
-};
-
-Form.Methods = {
-  serialize: function(form, getHash) {
-    return Form.serializeElements(Form.getElements(form), getHash);
-  },
-
-  getElements: function(form) {
-    return $A($(form).getElementsByTagName('*')).inject([],
-      function(elements, child) {
-        if (Form.Element.Serializers[child.tagName.toLowerCase()])
-          elements.push(Element.extend(child));
-        return elements;
-      }
-    );
-  },
-
-  getInputs: function(form, typeName, name) {
-    form = $(form);
-    var inputs = form.getElementsByTagName('input');
-
-    if (!typeName && !name) return $A(inputs).map(Element.extend);
-
-    for (var i = 0, matchingInputs = [], length = inputs.length; i < length; i++) {
-      var input = inputs[i];
-      if ((typeName && input.type != typeName) || (name && input.name != name))
-        continue;
-      matchingInputs.push(Element.extend(input));
-    }
-
-    return matchingInputs;
-  },
-
-  disable: function(form) {
-    form = $(form);
-    form.getElements().each(function(element) {
-      element.blur();
-      element.disabled = 'true';
-    });
-    return form;
-  },
-
-  enable: function(form) {
-    form = $(form);
-    form.getElements().each(function(element) {
-      element.disabled = '';
-    });
-    return form;
-  },
-
-  findFirstElement: function(form) {
-    return $(form).getElements().find(function(element) {
-      return element.type != 'hidden' && !element.disabled &&
-        ['input', 'select', 'textarea'].include(element.tagName.toLowerCase());
-    });
-  },
-
-  focusFirstElement: function(form) {
-    form = $(form);
-    form.findFirstElement().activate();
-    return form;
-  }
-}
-
-Object.extend(Form, Form.Methods);
-
-/*--------------------------------------------------------------------------*/
-
-Form.Element = {
-  focus: function(element) {
-    $(element).focus();
-    return element;
-  },
-
-  select: function(element) {
-    $(element).select();
-    return element;
-  }
-}
-
-Form.Element.Methods = {
-  serialize: function(element) {
-    element = $(element);
-    if (!element.disabled && element.name) {
-      var value = element.getValue();
-      if (value != undefined) {
-        var pair = {};
-        pair[element.name] = value;
-        return Hash.toQueryString(pair);
-      }
-    }
-    return '';
-  },
-
-  getValue: function(element) {
-    element = $(element);
-    var method = element.tagName.toLowerCase();
-    return Form.Element.Serializers[method](element);
-  },
-
-  clear: function(element) {
-    $(element).value = '';
-    return element;
-  },
-
-  present: function(element) {
-    return $(element).value != '';
-  },
-
-  activate: function(element) {
-    element = $(element);
-    element.focus();
-    if (element.select && ( element.tagName.toLowerCase() != 'input' ||
-      !['button', 'reset', 'submit'].include(element.type) ) )
-      element.select();
-    return element;
-  },
-
-  disable: function(element) {
-    element = $(element);
-    element.disabled = true;
-    return element;
-  },
-
-  enable: function(element) {
-    element = $(element);
-    element.blur();
-    element.disabled = false;
-    return element;
-  }
-}
-
-Object.extend(Form.Element, Form.Element.Methods);
-var Field = Form.Element;
-var $F = Form.Element.getValue;
-
-/*--------------------------------------------------------------------------*/
-
-Form.Element.Serializers = {
-  input: function(element) {
-    switch (element.type.toLowerCase()) {
-      case 'checkbox':
-      case 'radio':
-        return Form.Element.Serializers.inputSelector(element);
-      default:
-        return Form.Element.Serializers.textarea(element);
-    }
-  },
-
-  inputSelector: function(element) {
-    return element.checked ? element.value : null;
-  },
-
-  textarea: function(element) {
-    return element.value;
-  },
-
-  select: function(element) {
-    return this[element.type == 'select-one' ?
-      'selectOne' : 'selectMany'](element);
-  },
-
-  selectOne: function(element) {
-    var index = element.selectedIndex;
-    return index >= 0 ? this.optionValue(element.options[index]) : null;
-  },
-
-  selectMany: function(element) {
-    var values, length = element.length;
-    if (!length) return null;
-
-    for (var i = 0, values = []; i < length; i++) {
-      var opt = element.options[i];
-      if (opt.selected) values.push(this.optionValue(opt));
-    }
-    return values;
-  },
-
-  optionValue: function(opt) {
-    // extend element because hasAttribute may not be native
-    return Element.extend(opt).hasAttribute('value') ? opt.value : opt.text;
-  }
-}
-
-/*--------------------------------------------------------------------------*/
-
-Abstract.TimedObserver = function() {}
-Abstract.TimedObserver.prototype = {
-  initialize: function(element, frequency, callback) {
-    this.frequency = frequency;
-    this.element   = $(element);
-    this.callback  = callback;
-
-    this.lastValue = this.getValue();
-    this.registerCallback();
-  },
-
-  registerCallback: function() {
-    setInterval(this.onTimerEvent.bind(this), this.frequency * 1000);
-  },
-
-  onTimerEvent: function() {
-    var value = this.getValue();
-    var changed = ('string' == typeof this.lastValue && 'string' == typeof value
-      ? this.lastValue != value : String(this.lastValue) != String(value));
-    if (changed) {
-      this.callback(this.element, value);
-      this.lastValue = value;
-    }
-  }
-}
-
-Form.Element.Observer = Class.create();
-Form.Element.Observer.prototype = Object.extend(new Abstract.TimedObserver(), {
-  getValue: function() {
-    return Form.Element.getValue(this.element);
-  }
-});
-
-Form.Observer = Class.create();
-Form.Observer.prototype = Object.extend(new Abstract.TimedObserver(), {
-  getValue: function() {
-    return Form.serialize(this.element);
-  }
-});
-
-/*--------------------------------------------------------------------------*/
-
-Abstract.EventObserver = function() {}
-Abstract.EventObserver.prototype = {
-  initialize: function(element, callback) {
-    this.element  = $(element);
-    this.callback = callback;
-
-    this.lastValue = this.getValue();
-    if (this.element.tagName.toLowerCase() == 'form')
-      this.registerFormCallbacks();
-    else
-      this.registerCallback(this.element);
-  },
-
-  onElementEvent: function() {
-    var value = this.getValue();
-    if (this.lastValue != value) {
-      this.callback(this.element, value);
-      this.lastValue = value;
-    }
-  },
-
-  registerFormCallbacks: function() {
-    Form.getElements(this.element).each(this.registerCallback.bind(this));
-  },
-
-  registerCallback: function(element) {
-    if (element.type) {
-      switch (element.type.toLowerCase()) {
-        case 'checkbox':
-        case 'radio':
-          Event.observe(element, 'click', this.onElementEvent.bind(this));
-          break;
-        default:
-          Event.observe(element, 'change', this.onElementEvent.bind(this));
-          break;
-      }
-    }
-  }
-}
-
-Form.Element.EventObserver = Class.create();
-Form.Element.EventObserver.prototype = Object.extend(new Abstract.EventObserver(), {
-  getValue: function() {
-    return Form.Element.getValue(this.element);
-  }
-});
-
-Form.EventObserver = Class.create();
-Form.EventObserver.prototype = Object.extend(new Abstract.EventObserver(), {
-  getValue: function() {
-    return Form.serialize(this.element);
-  }
-});
-if (!window.Event) {
-  var Event = new Object();
-}
-
-Object.extend(Event, {
-  KEY_BACKSPACE: 8,
-  KEY_TAB:       9,
-  KEY_RETURN:   13,
-  KEY_ESC:      27,
-  KEY_LEFT:     37,
-  KEY_UP:       38,
-  KEY_RIGHT:    39,
-  KEY_DOWN:     40,
-  KEY_DELETE:   46,
-  KEY_HOME:     36,
-  KEY_END:      35,
-  KEY_PAGEUP:   33,
-  KEY_PAGEDOWN: 34,
-
-  element: function(event) {
-    return event.target || event.srcElement;
-  },
-
-  isLeftClick: function(event) {
-    return (((event.which) && (event.which == 1)) ||
-            ((event.button) && (event.button == 1)));
-  },
-
-  pointerX: function(event) {
-    return event.pageX || (event.clientX +
-      (document.documentElement.scrollLeft || document.body.scrollLeft));
-  },
-
-  pointerY: function(event) {
-    return event.pageY || (event.clientY +
-      (document.documentElement.scrollTop || document.body.scrollTop));
-  },
-
-  stop: function(event) {
-    if (event.preventDefault) {
-      event.preventDefault();
-      event.stopPropagation();
-    } else {
-      event.returnValue = false;
-      event.cancelBubble = true;
-    }
-  },
-
-  // find the first node with the given tagName, starting from the
-  // node the event was triggered on; traverses the DOM upwards
-  findElement: function(event, tagName) {
-    var element = Event.element(event);
-    while (element.parentNode && (!element.tagName ||
-        (element.tagName.toUpperCase() != tagName.toUpperCase())))
-      element = element.parentNode;
-    return element;
-  },
-
-  observers: false,
-
-  _observeAndCache: function(element, name, observer, useCapture) {
-    if (!this.observers) this.observers = [];
-    if (element.addEventListener) {
-      this.observers.push([element, name, observer, useCapture]);
-      element.addEventListener(name, observer, useCapture);
-    } else if (element.attachEvent) {
-      this.observers.push([element, name, observer, useCapture]);
-      element.attachEvent('on' + name, observer);
-    }
-  },
-
-  unloadCache: function() {
-    if (!Event.observers) return;
-    for (var i = 0, length = Event.observers.length; i < length; i++) {
-      Event.stopObserving.apply(this, Event.observers[i]);
-      Event.observers[i][0] = null;
-    }
-    Event.observers = false;
-  },
-
-  observe: function(element, name, observer, useCapture) {
-    element = $(element);
-    useCapture = useCapture || false;
-
-    if (name == 'keypress' &&
-        (navigator.appVersion.match(/Konqueror|Safari|KHTML/)
-        || element.attachEvent))
-      name = 'keydown';
-
-    Event._observeAndCache(element, name, observer, useCapture);
-  },
-
-  stopObserving: function(element, name, observer, useCapture) {
-    element = $(element);
-    useCapture = useCapture || false;
-
-    if (name == 'keypress' &&
-        (navigator.appVersion.match(/Konqueror|Safari|KHTML/)
-        || element.detachEvent))
-      name = 'keydown';
-
-    if (element.removeEventListener) {
-      element.removeEventListener(name, observer, useCapture);
-    } else if (element.detachEvent) {
-      try {
-        element.detachEvent('on' + name, observer);
-      } catch (e) {}
-    }
-  }
-});
-
-/* prevent memory leaks in IE */
-if (navigator.appVersion.match(/\bMSIE\b/))
-  Event.observe(window, 'unload', Event.unloadCache, false);
-var Position = {
-  // set to true if needed, warning: firefox performance problems
-  // NOT neeeded for page scrolling, only if draggable contained in
-  // scrollable elements
-  includeScrollOffsets: false,
-
-  // must be called before calling withinIncludingScrolloffset, every time the
-  // page is scrolled
-  prepare: function() {
-    this.deltaX =  window.pageXOffset
-                || document.documentElement.scrollLeft
-                || document.body.scrollLeft
-                || 0;
-    this.deltaY =  window.pageYOffset
-                || document.documentElement.scrollTop
-                || document.body.scrollTop
-                || 0;
-  },
-
-  realOffset: function(element) {
-    var valueT = 0, valueL = 0;
-    do {
-      valueT += element.scrollTop  || 0;
-      valueL += element.scrollLeft || 0;
-      element = element.parentNode;
-    } while (element);
-    return [valueL, valueT];
-  },
-
-  cumulativeOffset: function(element) {
-    var valueT = 0, valueL = 0;
-    do {
-      valueT += element.offsetTop  || 0;
-      valueL += element.offsetLeft || 0;
-      element = element.offsetParent;
-    } while (element);
-    return [valueL, valueT];
-  },
-
-  positionedOffset: function(element) {
-    var valueT = 0, valueL = 0;
-    do {
-      valueT += element.offsetTop  || 0;
-      valueL += element.offsetLeft || 0;
-      element = element.offsetParent;
-      if (element) {
-        if(element.tagName=='BODY') break;
-        var p = Element.getStyle(element, 'position');
-        if (p == 'relative' || p == 'absolute') break;
-      }
-    } while (element);
-    return [valueL, valueT];
-  },
-
-  offsetParent: function(element) {
-    if (element.offsetParent) return element.offsetParent;
-    if (element == document.body) return element;
-
-    while ((element = element.parentNode) && element != document.body)
-      if (Element.getStyle(element, 'position') != 'static')
-        return element;
-
-    return document.body;
-  },
-
-  // caches x/y coordinate pair to use with overlap
-  within: function(element, x, y) {
-    if (this.includeScrollOffsets)
-      return this.withinIncludingScrolloffsets(element, x, y);
-    this.xcomp = x;
-    this.ycomp = y;
-    this.offset = this.cumulativeOffset(element);
-
-    return (y >= this.offset[1] &&
-            y <  this.offset[1] + element.offsetHeight &&
-            x >= this.offset[0] &&
-            x <  this.offset[0] + element.offsetWidth);
-  },
-
-  withinIncludingScrolloffsets: function(element, x, y) {
-    var offsetcache = this.realOffset(element);
-
-    this.xcomp = x + offsetcache[0] - this.deltaX;
-    this.ycomp = y + offsetcache[1] - this.deltaY;
-    this.offset = this.cumulativeOffset(element);
-
-    return (this.ycomp >= this.offset[1] &&
-            this.ycomp <  this.offset[1] + element.offsetHeight &&
-            this.xcomp >= this.offset[0] &&
-            this.xcomp <  this.offset[0] + element.offsetWidth);
-  },
-
-  // within must be called directly before
-  overlap: function(mode, element) {
-    if (!mode) return 0;
-    if (mode == 'vertical')
-      return ((this.offset[1] + element.offsetHeight) - this.ycomp) /
-        element.offsetHeight;
-    if (mode == 'horizontal')
-      return ((this.offset[0] + element.offsetWidth) - this.xcomp) /
-        element.offsetWidth;
-  },
-
-  page: function(forElement) {
-    var valueT = 0, valueL = 0;
-
-    var element = forElement;
-    do {
-      valueT += element.offsetTop  || 0;
-      valueL += element.offsetLeft || 0;
-
-      // Safari fix
-      if (element.offsetParent==document.body)
-        if (Element.getStyle(element,'position')=='absolute') break;
-
-    } while (element = element.offsetParent);
-
-    element = forElement;
-    do {
-      if (!window.opera || element.tagName=='BODY') {
-        valueT -= element.scrollTop  || 0;
-        valueL -= element.scrollLeft || 0;
-      }
-    } while (element = element.parentNode);
-
-    return [valueL, valueT];
-  },
-
-  clone: function(source, target) {
-    var options = Object.extend({
-      setLeft:    true,
-      setTop:     true,
-      setWidth:   true,
-      setHeight:  true,
-      offsetTop:  0,
-      offsetLeft: 0
-    }, arguments[2] || {})
-
-    // find page position of source
-    source = $(source);
-    var p = Position.page(source);
-
-    // find coordinate system to use
-    target = $(target);
-    var delta = [0, 0];
-    var parent = null;
-    // delta [0,0] will do fine with position: fixed elements,
-    // position:absolute needs offsetParent deltas
-    if (Element.getStyle(target,'position') == 'absolute') {
-      parent = Position.offsetParent(target);
-      delta = Position.page(parent);
-    }
-
-    // correct by body offsets (fixes Safari)
-    if (parent == document.body) {
-      delta[0] -= document.body.offsetLeft;
-      delta[1] -= document.body.offsetTop;
-    }
-
-    // set position
-    if(options.setLeft)   target.style.left  = (p[0] - delta[0] + options.offsetLeft) + 'px';
-    if(options.setTop)    target.style.top   = (p[1] - delta[1] + options.offsetTop) + 'px';
-    if(options.setWidth)  target.style.width = source.offsetWidth + 'px';
-    if(options.setHeight) target.style.height = source.offsetHeight + 'px';
-  },
-
-  absolutize: function(element) {
-    element = $(element);
-    if (element.style.position == 'absolute') return;
-    Position.prepare();
-
-    var offsets = Position.positionedOffset(element);
-    var top     = offsets[1];
-    var left    = offsets[0];
-    var width   = element.clientWidth;
-    var height  = element.clientHeight;
-
-    element._originalLeft   = left - parseFloat(element.style.left  || 0);
-    element._originalTop    = top  - parseFloat(element.style.top || 0);
-    element._originalWidth  = element.style.width;
-    element._originalHeight = element.style.height;
-
-    element.style.position = 'absolute';
-    element.style.top    = top + 'px';
-    element.style.left   = left + 'px';
-    element.style.width  = width + 'px';
-    element.style.height = height + 'px';
-  },
-
-  relativize: function(element) {
-    element = $(element);
-    if (element.style.position == 'relative') return;
-    Position.prepare();
-
-    element.style.position = 'relative';
-    var top  = parseFloat(element.style.top  || 0) - (element._originalTop || 0);
-    var left = parseFloat(element.style.left || 0) - (element._originalLeft || 0);
-
-    element.style.top    = top + 'px';
-    element.style.left   = left + 'px';
-    element.style.height = element._originalHeight;
-    element.style.width  = element._originalWidth;
-  }
-}
-
-// Safari returns margins on body which is incorrect if the child is absolutely
-// positioned.  For performance reasons, redefine Position.cumulativeOffset for
-// KHTML/WebKit only.
-if (/Konqueror|Safari|KHTML/.test(navigator.userAgent)) {
-  Position.cumulativeOffset = function(element) {
-    var valueT = 0, valueL = 0;
-    do {
-      valueT += element.offsetTop  || 0;
-      valueL += element.offsetLeft || 0;
-      if (element.offsetParent == document.body)
-        if (Element.getStyle(element, 'position') == 'absolute') break;
-
-      element = element.offsetParent;
-    } while (element);
-
-    return [valueL, valueT];
-  }
-}
-
-Element.addMethods();
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/tools.php b/Tools/RunTimeTester/share/web/tools.php
deleted file mode 100644
index 9f5f95bcebb021acfafbc3dfcd04eb4cf48987c1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/tools.php
+++ /dev/null
@@ -1,281 +0,0 @@
-<?
-
-function extractFromStatusText($statusText, $whatToExtract){
-  if ($statusText == '---'){return "---";}
-
-  $toks = split(':', $statusText);
-
-  $numbers = $toks[3];
-  $nToks = split('/', $numbers);
-  if ($whatToExtract=='ok') {
-    return $nToks[0];
-  }
-  if ($whatToExtract=='done') {
-    return $nToks[1];
-  } 
-  if ($whatToExtract=='total') {
-    return $nToks[2];
-  } 
-}
-
-function cvsReposURL($fullPackagePath, $packageTag){
-  // fullpackagePath is path from offline/
-  $baseURL = "http://atlas-sw.cern.ch/cgi-bin/viewcvs-atlas.cgi/offline/";
-  $URL = $baseURL.$fullPackagePath.'/?pathrev='.$packageTag;
-  if ($packageTag == ''){
-    $URL = $baseURL.$fullPackagePath;
-  }
-
-  return $URL;
-}
-
-function openAndReadSummaryFilePaths($pathToFile){
-  if(!file_exists($pathToFile)){return NULL;}
-  $handle = fopen($pathToFile, 'r');
-  $lines = array();
-  while (!feof($handle)){
-    $line = trim(fgets($handle));
-    if (!feof($handle) && sizeOf($line) > 0){
-      if (!in_array($line, $lines)){
-	$lines[] = $line;
-      }
-    }
-  }
-
-  $listOfSummFiles = array();
-  foreach($lines as $line){
-    if(!file_exists($line) || filesize($line)==0){continue;}
-    $listOfSummFiles[] = $line;
-  }
-
-  fclose($handle);
-
-  return $listOfSummFiles;
-}
-
-function getTagContent($parentNode, $tagName){
-  $thing = $parentNode->get_elements_by_tagname($tagName);
-  if (sizeOf($thing)==0) return "";
-  $thing = $thing[0]->get_content();
-  return trim($thing);
-}
-
-function getMultiTagContent($parentNode, $tagName){
-  $things = $parentNode->get_elements_by_tagname($tagName);
-  if (sizeOf($things)==0) return array("");
-
-  $contents = array();
-  foreach($things as $thing){
-    $contents[] = trim($thing->get_content());
-  }
-  
-  return $contents;
-}
-
-function getRootNodeFromXMLfile($theFile){
-  //$xmlDoc = xmldocfile($theFile);
-  $xmlDoc = domxml_open_file($theFile);
-  $rootNode = $xmlDoc->document_element();
-  $xmlDoc->unlink();
-  unset($xmlDoc);
-  return $rootNode;
-}
-
-function getAttributeValue($element, $attributeName){
-  if ($element->has_attributes()){
-    foreach ($element->attributes() as $attr){
-      if($attr->name == $attributeName) return trim($attr->value());
-    }
-  }
-
-  return '';
-}
-
-function hasChildTag($parentNode, $childTagName){
-  return sizeOf($parentNode->get_elements_by_tagname($childTagName)) > 0;
-}
-
-function isEmptyTag($element){
-  return ($element->is_blank_node() || strlen(trim($element->get_content())) == 0);
-}
-
-function getSingleTagWithChildTag($parentNodes, $childName, $childValue){
-  $tags = getTagsWithChildTags($parentNodes, $childName, $childValue);
-  return $tags[0];
-}
-
-function getTagsWithChildTags($parentNodes, $childTagName, $childTagValue, $inverse = FALSE){
-  $index = 0;
-  $matches = array();
-  foreach ($parentNodes as $parentNode){
-    $val = getTagContent($parentNode, $childTagName);
-    if ($val == $childTagValue){
-      $matches[$index] = $parentNode;
-      $index++;
-    }
-  }
-
-  return $matches;
-}
-
-////////////////////////////////////////////////////////////
-// SORT SUMMARY FILES
-////////////////////////////////////////////////////////////
-
-function getOverviewNode($summaryFilePath){
-  $rootNode = getRootNodeFromXMLfile($summaryFilePath);
-  $overviewNodes = $rootNode->get_elements_by_tagname('overview');
-  $oNode = $overviewNodes[sizeOf($overviewNodes)-1]; //0th one is ATN summary, ignore if present
-  return $oNode;
-}
-
-function sortSummaryFilesByRelease($a, $b){
-  $A = $a['release'];
-  $B = $b['release'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortSummaryFilesByBranch($a, $b){
-  $A = $a['branch'];
-  $B = $b['branch'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortSummaryFilesByPlatform($a, $b){
-  $A = $a['platform'];
-  $B = $b['platform'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortSummaryFilesByProject($a, $b){
-  $A = $a['topProject'];
-  $B = $b['topProject'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortSummaryFilesByStarted($a, $b){
-  $A = $a['started'];
-  $B = $b['started'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortSummaryFilesByCompleted($a, $b){
-  $A = $a['completed'];
-  $B = $b['completed'];
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-// ==================================================================
-
-function sortPackagesByName($a, $b){
-  $nameA    = strtolower(trim(getTagContent($a, 'packageName')));
-  $nameB    = strtolower(trim(getTagContent($b, 'packageName')));
-  if ($nameA == $nameB) {
-    return 0;
-  }
-  return ($nameA < $nameB) ? -1 : 1;
-}
-
-function sortPackagesByStatus($a, $b){
-  $nameA    = getTagContent($a, 'packageStatus');
-  $nameB    = getTagContent($b, 'packageStatus');
-  if ($nameA == $nameB) {
-    return 0;
-  }
-  return ($nameA < $nameB) ? -1 : 1;
-}
-
-function sortJobsByJob($a, $b){
-  $nameA    = getTagContent($a, 'identifiedName');
-  $nameB    = getTagContent($b, 'identifiedName');
-  if ($nameA == $nameB) {
-    return 0;
-  }
-  return ($nameA < $nameB) ? -1 : 1;
-}
-
-function sortJobsByGroup($a, $b){
-  $nameA    = getTagContent($a, 'jobGroup');
-  $nameB    = getTagContent($b, 'jobGroup');
-  if ($nameA == $nameB) {
-    return 0;
-  }
-  return ($nameA < $nameB) ? -1 : 1;
-}
-
-function sortKeepFiles($a, $b){
-  $keepA    = strtolower(trim($a->get_content()));
-  $keepB    = strtolower(trim($b->get_content()));
-  if ($keepA == $keepB) {
-    return 0;
-  }
-  return ($keepA < $keepB) ? -1 : 1;
-}
-
-//=====================================================
-// CPC sorting for the CPC view
-//=====================================================
-function sortCPCByClass($cpcKeyA, $cpcKeyB){
-  $A = explode(' ', $cpcKeyA);
-  $A = trim($A[0]);
-  $A = substr($A, 1, -1);
-
-  $B = explode(' ', $cpcKeyB);
-  $B = trim($B[0]);
-  $B = substr($B, 1, -1);
-
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortCPCByProcess($cpcKeyA, $cpcKeyB){
-  $A = explode(' ', $cpcKeyA);
-  $A = trim($A[1]);
-  $A = substr($A, 1, -1);
-
-  $B = explode(' ', $cpcKeyB);
-  $B = trim($B[1]);
-  $B = substr($B, 1, -1);
-
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-function sortCPCByComponent($cpcKeyA, $cpcKeyB){
-  $A = explode(' ', $cpcKeyA);
-  $A = trim($A[2]);
-  $A = substr($A, 1, -1);
-
-  $B = explode(' ', $cpcKeyB);
-  $B = trim($B[2]);
-  $B = substr($B, 1, -1);
-
-  if ($A == $B) {
-    return 0;
-  }
-  return ($A < $B) ? -1 : 1;
-}
-
-?>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/share/web/unifiedTestConfiguration.dtd b/Tools/RunTimeTester/share/web/unifiedTestConfiguration.dtd
deleted file mode 100755
index e1b04cbec545989337bfc8160635883c713630f3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/share/web/unifiedTestConfiguration.dtd
+++ /dev/null
@@ -1,171 +0,0 @@
-<!ELEMENT unifiedTestConfiguration ((doc?, atn,  kv?, rtt?) |
-                                    (doc?, atn?, kv,  rtt?) |
-                                    (doc?, atn?, kv?, rtt))>
-
-<!ELEMENT classification ((none) | (displayClass, displayProcess+, displayComponent+))?>
-<!ELEMENT displayClass (#PCDATA)>
-<!ELEMENT displayProcess (#PCDATA)>
-<!ELEMENT displayComponent (#PCDATA)>
-<!ELEMENT none EMPTY>
-
-
-<!ELEMENT atn         (classification?, TEST*)>
-  <!ELEMENT TEST     (doc?, classification?, options_atn*, furtherOptions_atn*, timelimit*, author*, mailto*, prescript*, postscript*, package_atn*, rundir*, expectations*)>
-  <!ATTLIST TEST   name       CDATA                    #IMPLIED>
-  <!ATTLIST TEST   type       (script|athena|kv|rtt|makecheck)   #REQUIRED>
-  <!ATTLIST TEST   suite      CDATA                    "general">
-  <!ELEMENT options_atn     (#PCDATA | rttSpecs | kvSpecs)*>
-      <!ELEMENT rttSpecs  ANY>
-      <!ELEMENT kvSpecs   ANY>
-  <!ELEMENT furtherOptions_atn (#PCDATA)>
-  <!ELEMENT timelimit          (#PCDATA)>
-  <!ELEMENT author             (#PCDATA)>
-  <!ELEMENT mailto             (#PCDATA)>
-  <!ELEMENT prescript          (#PCDATA)>
-  <!ELEMENT postscript         (#PCDATA)>
-  <!ELEMENT package_atn        (#PCDATA)>
-  <!ELEMENT rundir             (#PCDATA)>
-  <!ELEMENT expectations       (errorMessage*, warningMessage*, successMessage*, returnValue*)>
-     <!ELEMENT errorMessage    (#PCDATA)>
-     <!ELEMENT warningMessage  (#PCDATA)>
-     <!ELEMENT successMessage  (#PCDATA)>
-     <!ELEMENT returnValue     (#PCDATA)>
-
-<!ELEMENT kv                (kvtest*)>
-    <!ELEMENT kvtest (release,priority?,kvsuite,trf,desc,author?,version?,
-                    inpath,infile,inpool,outpath,outfile,outntuple?,logfile,signature,copyfiles,checkfiles)>
-    <!ATTLIST kvtest                name    CDATA #REQUIRED>
-    <!ATTLIST kvtest                enabled CDATA #REQUIRED>
-
-    <!ELEMENT release     (#PCDATA)> 
-    <!ELEMENT priority    (#PCDATA)> 
-    <!ELEMENT kvsuite       (#PCDATA)> 
-    <!ELEMENT trf         (#PCDATA)> 
-    <!ELEMENT desc        (#PCDATA)> 
-    <!ELEMENT version     (#PCDATA)> 
-    <!ELEMENT inpath      (#PCDATA)> 
-    <!ELEMENT infile      (#PCDATA)> 
-    <!ELEMENT inpool      (#PCDATA)> 
-    <!ELEMENT outpath     (#PCDATA)> 
-    <!ELEMENT outfile     (#PCDATA)> 
-    <!ELEMENT outntuple   (#PCDATA)> 
-    <!ELEMENT logfile     (#PCDATA)> 
-    <!ELEMENT signature   (#PCDATA)> 
-    <!ELEMENT copyfiles   (#PCDATA)> 
-    <!ELEMENT checkfiles  (#PCDATA)> 
-
-
-<!ELEMENT rtt              ((EMPTY) | (rttContactPerson*,mailto?,refRelease?, jobList, jobGroups?) | (jobGroups))?>
-   <!ELEMENT refRelease    (#PCDATA)>   
-   <!ELEMENT jobList       (classification?, athena*, rulechecker*, 
-                            qametrics*,  kitvalidation*, rttmonitor*,
-                            rttselftest*, jobTransform*)>
-      <!ELEMENT athena     ( doc?, doclink?, (numberedOnly |  limitedNightlies)?, rttKitTestJob?, 
-                            classification?, displayName?, commandLineFlags?, commandLineString?, preconfigOptions?, options, furtherOptions?,
-                            package?, group, confederation?, 
-                            queue?,  dataset*, dataset_info*, dq2*, asciiRefFile*, binRefFile*,
-                            storeInDatabasePattern*, action*, monAction*, 
-                            test*, athenaFlavour*)>
-
-	 <!ELEMENT doc                    (#PCDATA)>
-	 <!ELEMENT doclink                (#PCDATA)>
-	 <!ELEMENT displayName            (#PCDATA)>
-	 <!ELEMENT rttContactPerson       (#PCDATA)>
-	 <!ELEMENT numberedOnly              EMPTY>
-	 <!ELEMENT limitedNightlies          EMPTY>
-	 <!ELEMENT limitedNightly         (nightly+)>
-         <!ELEMENT rttKitTestJob          EMPTY>
-         <!ELEMENT commandLineFlags       (#PCDATA)>
-         <!ELEMENT commandLineString      (#PCDATA)>
-         <!ELEMENT preconfigOptions       (#PCDATA)>
-         <!ELEMENT options                (#PCDATA)>
-          <!ELEMENT furtherOptions        (jobOptionItem+)>
-             <!ELEMENT jobOptionItem      (#PCDATA)>
-	 <!ELEMENT package                (#PCDATA)>
-	 <!ELEMENT group                  (#PCDATA)>
-	 <!ELEMENT confederation          (#PCDATA)>
-	 <!ELEMENT queue                  (#PCDATA)>
-	 <!ELEMENT nightly                (#PCDATA)>
-
- 	 <!-- The old dataset tag -->
-         <!ELEMENT dataset                  (#PCDATA)>	
-
- 	 <!-- The new dataset tag along with the definition of its children elements -->
-
-	 <!ELEMENT dataset_info                ((bytestream | dc2 | dc3 | jobTransformData), datasetName,datasetRepeat?)>
-	 <!ELEMENT bytestream                 EMPTY>
-	 <!ELEMENT dc2                        EMPTY>
-	 <!ELEMENT dc3                        EMPTY>
-	 <!ELEMENT jobTransformData           EMPTY>
-	 <!ELEMENT datasetName            (#PCDATA)>
-	 <!ELEMENT datasetRepeat          (#PCDATA)>
-         <!ELEMENT dq2                    (dq2site, dq2dataset,datatype, dq2file+)>
-         <!ELEMENT dq2site                (#PCDATA)>
-         <!ELEMENT dq2dataset             (#PCDATA)>
-         <!ELEMENT datatype               (#PCDATA)>
-         <!ELEMENT dq2file                (#PCDATA)>
-	 <!ELEMENT asciiRefFile           (#PCDATA)>
-	 <!ELEMENT storeInDatabasePattern (#PCDATA)>
-	 <!ELEMENT athenaFlavour          (#PCDATA)>
-
-      <!ELEMENT rulechecker ( (numberedOnly | limitedNightlies)?, 
-                             package, queue)>
-      <!ELEMENT qametrics   ( (numberedOnly | limitedNightlies)?,
-                             package, queue)>
-      <!ELEMENT kitvalidation ((EMPTY) | ( (numberedOnly | limitedNightlies)?, queue))?>
-      <!ELEMENT rttmonitor EMPTY>
-      <!ELEMENT rttselftest EMPTY>
-      <!ELEMENT jobTransform ((numberedOnly |  limitedNightlies)?, 
-                              doc?, doclink?, classification?, rttKitTestJob?, displayName?, jobTransformJobName, 
-                              jobTransformCmd, group, dataset_info*, queue?)>
-
-         <!ELEMENT jobTransformJobName (#PCDATA)>
-         <!ELEMENT jobTransformCmd (#PCDATA)>
-
-
-   <!ELEMENT jobGroups (jobGroup+)>
-      <!ELEMENT jobGroup (keepFilePattern*,   asciiRefFile*, 
-                          binRefFile*,        storeInDatabasePattern*, 
-                          checker*,           
-                          postScript*,        auxFilePattern*, 
-                          action*,            monAction*,test*,   stringToFile*,     
-                          summary*,           logChecker*)>
-
-            <!ELEMENT keepFilePattern (#PCDATA)>
-	    <!ELEMENT binRefFile      (#PCDATA)>
-            <!ELEMENT checker         ((searchString*, vetoString*) | (comparisonType?))>
-	       <!ELEMENT searchString     (#PCDATA)>
-	       <!ELEMENT vetoString       (#PCDATA)>
-               <!ELEMENT comparisonType   (#PCDATA)>
-            <!ELEMENT postScript          EMPTY>
-	    <!ELEMENT auxFilePattern      (#PCDATA)>
-            <!ELEMENT action             (modulename,testname, arg*, listarg*)>
-            <!ELEMENT monAction          (modulename,testname, arg*, listarg*)>
-            <!ELEMENT test               (modulename,testname, arg*, listarg*)>
-               <!ELEMENT modulename       (#PCDATA)>
-	       <!ELEMENT testname         (#PCDATA)>
-               <!ELEMENT arg              (argname, argvalue)>
-	          <!ELEMENT argname       (#PCDATA)>
-		  <!ELEMENT argvalue      (#PCDATA)>
-               <!ELEMENT listarg          (listargname, listargvalue+)>
-	          <!ELEMENT listargname   (#PCDATA)>
-		  <!ELEMENT listargvalue  (#PCDATA)>
-            <!ELEMENT stringToFile        (line+)>
-	       <!ELEMENT line             (#PCDATA)>
-            <!ELEMENT logChecker          ((searchString+, vetoString*) | 
-	                                   (searchString*, vetoString+))>
-            <!ELEMENT summary             (#PCDATA)>
-
-<!ATTLIST rtt               xmlns  CDATA #REQUIRED>
-<!ATTLIST jobGroup          name   CDATA #REQUIRED>
-<!ATTLIST jobGroup          parent CDATA #REQUIRED>
-<!ATTLIST class             name   CDATA #REQUIRED>
-<!ATTLIST action            position  CDATA "">
-<!ATTLIST test              position  CDATA "">
-<!ATTLIST keepFilePattern   info   CDATA "">
-<!ATTLIST keepFilePattern   displayColor   CDATA "black">
-<!ATTLIST checker           class  CDATA #REQUIRED>
-<!ATTLIST logChecker        class  CDATA #REQUIRED>
-<!ATTLIST postScript        class  CDATA #REQUIRED>
-<!ATTLIST stringToFile      file   CDATA #REQUIRED>
-
diff --git a/Tools/RunTimeTester/src/ATNSummaryRetriever.py b/Tools/RunTimeTester/src/ATNSummaryRetriever.py
deleted file mode 100755
index 791da114ece2a45894388f872d967a700b418f3c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ATNSummaryRetriever.py
+++ /dev/null
@@ -1,82 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import urllib, urlparse, xml.dom.minidom
-from Evaluate import Evaluate
-
-class ATNSummaryRetriever:
-    def __init__(self, cmtConfig, originalBranch, release, logger):
-
-        pathFrags = {
-            ('i686-slc3-gcc323-opt',  'bugfix'): 'nicos_web_areaBF32BS3ProdOpt',
-            ('i686-slc4-gcc34-opt',  'bugfix'): 'nicos_web_areaBF32BS4ProdOpt',
-            ('i686-slc3-gcc323-opt',  'dev'):    'nicos_web_areaDev32BS3ProdOpt',
-            ('i686-slc4-gcc34-debug', 'dev'):    'nicos_web_areaDev32BS4ProdDbg',
-            ('i686-slc4-gcc34-opt',   'dev'):    'nicos_web_areaDev32BS4ProdOpt',
-            ('i686-slc4-gcc34-dbg',   'val'):    'VALWebArea/nicos_web_areaVAL32BS4ProdDbg',
-            ('i686-slc4-gcc34-opt',   'val'):    'VALWebArea/nicos_web_areaVAL32BS4ProdOpt',
-            ('x86_64-slc4-gcc34-opt', 'dev'):    'nicos_web_areaExp64BS4ProdOpt',
-            ('x86_64-slc4-gcc34-debug','dev'):   'nicos_web_areaExp64BS4ProdDbg'
-            }
-
-        cfg = (cmtConfig, originalBranch)
-        try:
-            pathFrag =  pathFrags[cfg]
-        except:
-            self.dom = None
-            logger.warning('Cannot handle config:  %s %s' % cfg)
-            return
-
-        
-        
-        self.logger = logger
-        nameLookUp = {
-            'rel_0':  'ATNComboSummary_0.xml',
-            'rel_1':  'ATNComboSummary_1.xml',
-            'rel_2':  'ATNComboSummary_2.xml',
-            'rel_3':  'ATNComboSummary_3.xml',
-            'rel_4':  'ATNComboSummary_4.xml',
-            'rel_5':  'ATNComboSummary_5.xml',
-            'rel_6':  'ATNComboSummary_6.xml'}
-
-        
-
-        servername = 'atlas-computing.web.cern.ch'
-        try:
-            localFile  = nameLookUp[release]
-        except:
-            self.logger.error('Unknown release: %s' % str(release))
-            self.dom = None
-            return
-
-        addOn = 'projects'
-        if originalBranch == 'val': addOn = ''
-        filename   = '/atlas-computing/links/distDirectory/nightlies/%s/%s/%s' % (addOn, pathFrag, localFile)
-
-        
-        remoteAddr = 'http://%s%s' % (servername, filename)
-        try:
-            urllib.urlretrieve(remoteAddr, localFile)
-            self.logger.info('Retrieved ATN summary: %s' %  remoteAddr)
-        except:
-            self.logger.warning('Could not retrieve ATN summary: %s' %  remoteAddr)
-            self.dom = None
-
-        try:
-            self.dom = xml.dom.minidom.parse(localFile)
-        except:
-            self.logger.warning('Could not parse ATN summary: %s' %  localFile)
-            self.dom = None
-
-        if self.dom: self.massageDom()
-
-    def massageDom(self):
-
-        de = self.dom.documentElement
-        tests = Evaluate('test', de)
-
-        nonAthenaTests = [t for t in tests if Evaluate('testType/text()', t) != 'athena']
-            
-        [de.removeChild(t) for t  in nonAthenaTests]
-        [n.unlink() for n in nonAthenaTests]
-        
-    def getDom(self): return self.dom
diff --git a/Tools/RunTimeTester/src/Archiver.py b/Tools/RunTimeTester/src/Archiver.py
deleted file mode 100644
index 2a8132ff3fb1d75426775264c9fb8e90e8b9f18f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Archiver.py
+++ /dev/null
@@ -1,312 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-import os
-import fnmatch
-import shutil
-
-from RTTSException     import RTTInputError, RTTTimeOut
-from Factory_BaseClass import Factory
-from Tools2            import getFileOrDirDateStamp
-from Factory_IOUtils   import IOUtilsFactory
-from Tools             import timed_cmd
-
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-
-class ArchivePathSelectorFactory(Factory):
-    def __init__(self, logger):
-        Factory.__init__(self, logger)        
-
-    def create(self, type_, base_path, narchives):
-        dict = {'afs': AFSArchivePathSelector, 'castor':CastorArchivePathSelector}
-        try:
-            return dict[type_](self.logger, base_path, narchives)
-        except KeyError:
-            m = 'Unknown archive <type> %s, must be one of %s' % (type_, dict.keys())
-            raise RTTInputError(m)
-        
-
-# -----------------------------------------------------------------
-
-class ArchivePathSelector:
-    def __init__(self, logger, base_path, narchives):
-        self.logger      = logger
-        self.io          = self.createUtils()
-        shouldExist      = [os.path.join(base_path, str(i)) for i in range(int(narchives))]
-        dontExist        = [s for s in shouldExist if not self.io.exists(s)]
-        doExist          = [s for s in shouldExist if s not in dontExist]
-        doExistButNotDir = [s for s in shouldExist if s in doExist and not self.io.isdir(s)]
-        if doExistButNotDir:
-            m = ''
-            for thing in doExistButNotDir:
-                m  += "%s: exists but is not a directory. Cannot archive there.\n" % thing
-            raise RTTInputError(m)
-
-        self.base_path = base_path
-        self.available = doExist
-        self.narchives = narchives
-        if self.makeDirs(dontExist): self.available.extend(dontExist)
-        
-    def cmpTimeStamps(self, a, b):
-        aStamp = self.io.lastmodtime(a, asSecsSinceEpoch=True)
-        bStamp = self.io.lastmodtime(b, asSecsSinceEpoch=True)
-        cmp = 0
-        if aStamp < bStamp: cmp = -1
-        if aStamp > bStamp: cmp = 1
-        return cmp
-
-    def makeDirs(self, dirsToMake):
-        return False not in [self.io.mkdir(d) for d in dirsToMake]
-
-    def select_archive(self, run_frag): # run_frag = nightly/dev/build/i686-slc4-gcc34-opt/offline
-        self.available.sort()
-        timestamps = [os.path.join(a, run_frag, 'time_stamp') for a in self.available]
-        notUsedYet = [o for o in timestamps if not self.io.exists(o)]
-        if notUsedYet:
-            path_ = os.path.dirname(notUsedYet[0])
-            m = 'Choosing unused archive: %s' % path_
-            self.logger.debug(m)
-            self.makeTimeStamp(path_)
-            return path_ # just take the first one
-        
-        # Have to compare timestamps
-        timestamps.sort(self.cmpTimeStamps)
-        path_ = os.path.dirname(timestamps[0])
-        # Delete everything in that dir
-        self.logger.debug('Choosing previously used archive: %s' % path_)
-        self.logger.info('Preparing to delete everything inside: %s' % path_)
-        self.io.deleteStuffIn(path_)
-        self.logger.info('Have deleted everything inside: %s' % path_)
-        self.makeTimeStamp(path_)
-        return path_
-    
-# -----------------------------------------------------------------
-
-class AFSArchivePathSelector(ArchivePathSelector):
-    def __init__(self, logger, base_path, narchives):      
-        ArchivePathSelector.__init__(self, logger, base_path, narchives) 
-
-    def createUtils(self):
-        return IOUtilsFactory(self.logger).create('afs')
-
-    def makeTimeStamp(self, path_):
-        ts = os.path.join(path_, 'time_stamp')
-        timed_cmd('mkdir -p %s; touch %s' % (path_, ts))
-
-# -----------------------------------------------------------------
-    
-class CastorArchivePathSelector(ArchivePathSelector):
-    def __init__(self, logger, base_path, narchives):
-        ArchivePathSelector.__init__(self, logger, base_path, narchives)
-
-    def createUtils(self):
-        return IOUtilsFactory(self.logger).create('castor')
-
-    def makeTimeStamp(self, path_):
-        ts = os.path.join(path_, 'time_stamp')
-        m = 'Making time stamp: %s' % ts
-        self.logger.debug(m)
-        cmd = 'touch time_stamp;rfmkdir -p %s;rfcp time_stamp %s;nsls -l %s;rm -f time_stamp;' % (path_, path_, ts)
-        self.logger.debug("Exec command: %s" % cmd)
-        m = 'Reply from cmd was:\n%s' % str(timed_cmd(cmd, 30))
-        self.logger.debug(m)
-        
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-
-class ArchivingCondition:
-    class ParentDir:
-        def __init__(self, dirDict):
-            self.is_    = dirDict.get('is_', [])
-            self.isnot_ = dirDict.get('isnot_', [])
-            
-        def ismatch(self, filepath):
-            dirs     = [f.strip() for f in os.path.dirname(filepath).split('/') if f.strip()]
-            if self.isnot_ and True in [len(fnmatch.filter(dirs, isnot_))>0 for isnot_ in self.isnot_]: return False
-            if not self.is_: return True
-
-            return [len(fnmatch.filter(dirs, is_))>0 for is_ in self.is_].count(True)==len(self.is_)
-
-    class Size:
-        def __init__(self, sizeDict):
-            self.gt = sizeDict.get('gt', 0)
-            self.lt = sizeDict.get('lt', 0)
-
-        def ismatch(self,filepath):
-            if self.gt and not self.lt:
-                return os.path.getsize(filepath)>self.gt
-            elif self.lt and not self.gt:
-                return os.path.getsize(filepath)<self.lt
-            else:
-                return os.path.getsize(filepath)<self.lt and os.path.getsize(filepath)>self.gt
-                
-    class Name:
-        def __init__(self, nameDict):
-            self.is_    = nameDict.get('is_', [])
-            self.isnot_ = nameDict.get('isnot_', [])
-                
-        def ismatch(self,filepath):
-            filename = os.path.basename(filepath)
-            m_   = True in [fnmatch.fnmatch(filename, is_) for is_ in self.is_]
-            nm_  = True in [fnmatch.fnmatch(filename, isnot_) for isnot_ in self.isnot_]
-            if self.isnot_:
-                if self.is_:
-                    value = m_ and not nm_
-                    return value
-                value = not nm_
-                return value
-            return m_
-
-    def __init__(self, conditionsList):
-        self.conditions = conditionsList
-    def ismatch(self, f):
-        # All these conditions must match
-        # This is an A _and_ B _and_ C type condition
-        return False not in [c.ismatch(f) for c in self.conditions]
-
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-
-class ArchiverFactory(Factory):
-    def __init__(self, logger):
-        Factory.__init__(self, logger)
-
-    def create(self, type_, base_path, run_frag, conditions):
-        return {'afs': AFSArchiver, 'castor':CastorArchiver}[type_](self.logger, base_path, run_frag, conditions)
-
-# -----------------------------------------------------------------
-
-class Archiver:
-    def __init__(self, logger, base_path, run_frag, conditions=[]):
-        self.logger      = logger
-        self.conditions  = conditions
-        self.basepath    = base_path
-        self.copiedOK    = []
-        self.notCopiedOK = []
-        self.run_frag    = run_frag # e.g. nightly/dev/build/i686-slc4-gcc34-opt/offline
-        self.archivepath = os.path.join(base_path, run_frag) # full path down to the heirarchy
-        self.io = self.createUtils()
-        
-    def get(self, runpath):
-        candidateFiles = []
-        for root, dirs, files in os.walk(runpath):
-            candidateFiles.extend([os.path.join(root,f) for f in files])
-        return candidateFiles
-
-    def archive(self, runpath, dontArchive=[]):
-        try:
-            matches = [m for m in self.findmatches(runpath) if m not in dontArchive]
-            self.copy(matches, runpath)
-        except Exception, e:
-            self.logger.error('Exception thrown while archiving files from run dir: %s' % runpath)
-            self.logger.error('Traceback: \n%s' % str(e))
-            
-    def copy(self, matchingfiles, runpath):
-        if not matchingfiles: return
-
-        statii = []
-        for mf in matchingfiles:
-            status = self.copyfile(mf)
-            if status[0] is None: continue
-            statii.append(status)
-            
-        self.copiedOK    = [s for s in statii if s[0]]
-        self.notCopiedOK = [s for s in statii if not s[0]]
-        self.logger.info("Copied %d files ok, %d files not ok" % (len(self.copiedOK),
-                                                                  len(self.notCopiedOK)))
-
-    def copiedFiles(self):
-        return [s[1] for s in self.copiedOK]
-        
-    def copyfile(self, filepath):
-        if not os.path.isfile(filepath): return (None, None, None) # no links or dirs
-        try:
-            fileRelToHierarchy = filepath.split(self.run_frag[self.run_frag.find('/'):])[1]
-        except:
-            raise
-        if fileRelToHierarchy.startswith('/'): fileRelToHierarchy = fileRelToHierarchy[1:]
-        
-        src     = filepath
-        dirname = os.path.dirname(fileRelToHierarchy)
-        dst     = os.path.join(self.archivepath, dirname)
-
-        self.io.mkdir(dst)
-        self.io.copyfile(src,dst)
-        return (self.io.exists(os.path.join(dst,os.path.basename(filepath))), src, dst)
-
-    def findmatches(self, runpath):        
-        matching = []
-        for f in self.get(runpath):
-            if [c for c in self.conditions if c.ismatch(f)]: matching.append(f)
-        return matching
-
-# -----------------------------------------------------------------
-
-class AFSArchiver(Archiver):
-    def __init__(self, logger, base_path, run_frag, conditions=[]):
-        Archiver.__init__(self, logger, base_path, run_frag, conditions)
-        self.type = 'afs'
-        
-    def createUtils(self):
-        return IOUtilsFactory(self.logger).create('afs')
-
-    def duplicate(self):
-        return AFSArchiver(self.logger, self.basepath, self.run_frag, self.conditions[:])
-
-# -----------------------------------------------------------------
-        
-class CastorArchiver(Archiver):
-    def __init__(self, logger, base_path, run_frag, conditions=[]):
-        Archiver.__init__(self, logger, base_path, run_frag, conditions)
-        self.type = 'castor'
-
-    def createUtils(self):
-        return IOUtilsFactory(self.logger).create('castor')
-
-    def duplicate(self):
-        return CastorArchiver(self.logger, self.basepath, self.run_frag, self.conditions[:])
-
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-
-if __name__ == '__main__':
-    class Logger:
-        def debug(self, s): pass
-        def info(self, s): pass
-        def warning(self, s): pass
-        def error(self, s): pass
-
-    import sys
-    # basepath = '/castor/cern.ch/grid/atlas/atlasgroupdisk/proj-sit/rtt/tct/archive/15.2.X.Y/'
-    basepath = '/castor/cern.ch/grid/atlas/atlasgroupdisk/proj-sit/rtt/brinick/archive'
-    run_frag = 'nightly/devval/build/i686-slc4-gcc34-opt/offline'
-    aps = CastorArchivePathSelector(Logger(), basepath, 2)
-    where =  aps.select_archive(run_frag)
-    print where
-    sys.exit(1)
-    
-    # apsf = ArchivePathSelectorFactory(Logger())
-    # sel = apsf.create('afs', '/afs/cern.ch/atlas/project/RTT/data/brinick/archive',2)
-    # run_frag = 'nightly/dev/build/i686-slc4-gcc34-opt/offline'
-    # where = sel.select_archive(run_frag)
-    archive_base = where.split(run_frag)[0].strip()
-    af = ArchiverFactory(Logger())
-    ac = ArchivingCondition(
-        [
-        # ArchivingCondition.Size({'lt':500.0}),
-        ArchivingCondition.ParentDir({'isnot_':['DBR*ase']}),
-        # ArchivingCondition.Name({'is_':['*.root', '*.data']})
-        ]
-        )
-    
-    archiver = af.create('castor', archive_base, run_frag, [ac])
-    rp = '/afs/cern.ch/atlas/offline/external/FullChainTest/tier0/rtt/FCT.Results/rel_0/15.2.X.Y/build/i686-slc4-gcc34-opt/offline/FullChainTests/FCTTransform/Ztautau_csc_recoAOD_trf/46'
-    matches =  archiver.findmatches(rp)
-    print len(matches)
-    for m in matches: print m
-    # archiver.archive(rp)
diff --git a/Tools/RunTimeTester/src/AthenaJobDescriptor.py b/Tools/RunTimeTester/src/AthenaJobDescriptor.py
deleted file mode 100755
index e32bb00ce8363d7bdb7002b46c7a07fab848d0da..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/AthenaJobDescriptor.py
+++ /dev/null
@@ -1,451 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""
-Class that stocks information about Athena jobs.
-Responsible for setting up the run directory for a job. Eventually becomes a data member
-of JobMinder.
-"""
-import os, string, copy, shutil
-
-from Paths               import Paths
-from os.path             import join, exists
-from WorkerJobDescriptor import WorkerJobDescriptor
-from Tools2              import searchPaths
-from Tools               import mkdirIfNew
-from formatCollection    import formatCollection
-from ShellCommand        import ShellCommand
-from RTTSException       import RTTInputError
-
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class FileLocator:
-    """Finds a file on a search path."""
-    def __init__(self, searchPath, logger):
-        self.paths  = searchPath['path'].split(searchPath['separater'])
-        self.logger = logger
-
-    def locateFile(self, fn):
-        for p in self.paths:
-            fpath = os.path.join(p, fn)
-            if os.path.exists(fpath):
-                self.logger.info('Found file: %s' % fpath)
-                return fpath
-            else:
-                self.logger.debug('%s: inexistant.' % fpath)
-        m = 'File %s not found in paths %s' % (fn, str(self.paths))
-        self.logger.error(m)
-        return None
-
-class AthenaJobDescriptor(WorkerJobDescriptor):    
-    
-    def __init__(self, paths, logger, jobSerialNumber, argDict):
-
-        # init the base class
-
-        self.commandLineFlags        = '' # corresponds to -c athena option
-        self.commandLineString       = '' # an ad verbatim string of whatever options are required
-        self.name                    = '' # filled in latter
-        self.suggestedQueue          = ''
-        self.confederation           = ''
-        self.jobGroup                = ''
-        # self.datasets                = {}
-        self.mode                    = ''
-        self.preconfigJobOptions     = []
-        self.jobOptions              = []
-        self.displayClass            = []
-        self.displayProcess          = []
-        self.displayComponent        = []
-        self.defaultAthenaExecutable = 'athena.py'
-        self.athenaFlavour           = self.defaultAthenaExecutable
-        self.athenaCommonFlags       = False
-        
-        # list of scripts to run before the athena command.
-        # 2/10/06: this list is provided by AthenaATNJobDescriptor
-        # and is passed here through argDict
-        self.preScripts = []
-
-        # argDict contains data read from the config files -
-        # this is the way variables such as jobOptions are set.
-        WorkerJobDescriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-        
-        self.package = paths.package
-        
-        self.fixJobOptions()
-
-        # recalculate the run and results paths (was done in Descriptor)
-        # note that self.name is set in self.fixName()
-        pathExtensions = [
-            ('package',    self.package),
-            ('jobGroup',   self.jobGroup),
-            ('jobName',    self.name),
-            ('SN',         str(self.jobSerialNumber))
-            ]
-
-        self.setRunPath(pathExtensions)
-        self.setKeepFiles()
-        # self.logger.debug('AthenaJobDescriptor.__init__ end')
-
-    # ------------------------------------------------------------------------
-
-    def pruneAttrs(self):
-        return ['package']
-
-    # ------------------------------------------------------------------------
-
-    def  setKeepFiles(self):
-        # copy the job options from the install area.
-        kfdict = {}
-        for jo in self.jobOptions:
-            srcFile     = join(self.runPath, jo)
-            destFile    = join(self.resPath, jo)
-            infoString  = "Job options called with the athena command"
-            displayColor= "#cc3333"
-            self.keepFiles[srcFile] = {'keepFileString':destFile,
-                                       'infoString':infoString,
-                                       'displayColor':displayColor}
-
-    # --------------------------------------------------------------
-
-    def fixJobOptions(self):
-        def fixExtension(jo):
-            # name: need to come up with a name that be used to mkdir a
-            # directory when (added) concatenated with a base path
-            ext =jo[-3:]
-            if ext !='.py':
-                m = "Adding .py extension to jobOptions" + jo
-                self.logger.debug(m)
-                jo += '.py'
-                
-            return jo
-        
-        self.jobOptions = [fixExtension(jo) for jo in self.jobOptions]
-
-        def fixPackage(jo):
-            tokens = jo.split('/')
-            if len(tokens) == 0: return jo
-            if len(tokens) == 1: 
-                jo = os.path.join(self.package, jo)
-                return jo
-            if len(tokens) == 2: return jo
-            if len(tokens) > 2:
-                return os.path.join(tokens[-2], tokens[-1])
-
-        self.jobOptions = [fixPackage(jo) for jo in self.jobOptions]
-
-    # --------------------------------------------------------------
-
-    def fixName(self):
-        # provide a name that be used to mkdir a directory
-        # when (added) concatenated with a base path
-
-        self.name,ext         = os.path.splitext(self.jobOptions[0])
-        self.name             = os.path.basename(self.name)
-        self.identifiedName   = self.name+str(self.jobSerialNumber)
-        self.log              = self.identifiedName+"_log"
-        # self.elog             = self.identifiedName+"_elog"
-
-    # --------------------------------------------------------------
-
-    def isDefaultExecutable(self):
-        return self.defaultAthenaExecutable == self.athenaFlavour
-
-    # --------------------------------------------------------------
-        
-    def athenaCmdOptions(self):
-        batchFlag = '-b'
-        
-        if self.commandLineString:
-            self.commandLineFlags = '' # can't have both, override
-            
-            if self.isDefaultExecutable() and self.commandLineString.find(batchFlag)==-1: # impose -b option
-                self.commandLineString = '%s %s' % (batchFlag, self.commandLineString)
-            return self.commandLineString
-
-        preconfig = ''
-        for p in self.preconfigJobOptions:
-            preconfig += ' %s' % p
-
-        if preconfig:
-            preconfig = '-p %s' % preconfig
-        
-        if not self.isDefaultExecutable():
-            batchFlag = '' # doesn't necessarily exist for other athena flavours
-            
-        if self.commandLineFlags:
-            self.commandLineFlags = self.commandLineFlags.replace("'",'"')
-            return "%s -c '%s' %s" % (preconfig, self.commandLineFlags, batchFlag)
-        else:
-            return '%s %s' % (preconfig, batchFlag)
-
-    # --------------------------------------------------------------
-
-    def athenaCommand(self):
-
-        cmd = ''
-        
-        for ps in self.preScripts:
-            cmd += 'source %s\n' % (ps)
-
-        njob = 0
-        for jo in self.jobOptions:
-            cmd += 'echo "Starting job %d"\n' % njob
-            cmd += '%s %s %s <&-\n' % (self.athenaFlavour,
-                                       self.athenaCmdOptions(),
-                                       jo)
-            cmd += 'echo "RTT determined exit status to be: " $?'
-            njob += 1
-        return cmd
-            
-    # --------------------------------------------------------------
-
-    def __str__(self):
-        """Dumps out descriptor"""
-
-        s = WorkerJobDescriptor.__str__(self)
-        for k in self.__dict__.keys():
-            line = 'job %s:' % k
-            line.ljust(24)
-            line += str(self.__dict__[k])+'\n'
-            s += line
-                        
-        return s
-
-    # --------------------------------------------------------------
-
-    def dump(self):
-        """Dumps out descriptor"""
-
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info('| Athena jobdescriptor  dump                |')
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info(self.__str__())
-
-    # --------------------------------------------------------------
-    def locateJobOptions(self, jo):
-        """
-        Use the joboptions search path from Paths and the list of jobOptions
-        (elements of the form <package>/<options> to find the full path to the
-        jobOptions files.
-        """
-        
-        class JOLocator:
-            def __init__(self, searchPath, logger):
-                self.paths  = searchPath.split(',')
-                self.logger = logger
-            def locateJO(self, jo):
-                for p in self.paths:
-                    fpath = os.path.join(p, jo)
-                    if os.path.exists(fpath): return fpath
-                m = 'JobOptions %s not found in paths %s' % (
-                    jo, str(self.paths))
-                self.logger.error(m)
-                return None
-
-
-        locator = JOLocator(self.paths.jobOptionsSearchPath, self.logger)
-        return locator.locateJO(jo)
-
-    # --------------------------------------------------------------
-
-    def setupRunDir(self):
-        """
-        copy job options to a subdirectory to <rundir>/<package>/<options>
-        """
-        class FullPathJOMaker:
-            def __init__(self, searchPath, logger):
-                self.locator = FileLocator(searchPath, logger)
-            def __call__(self, jo) :
-                return self.locator.locateFile(jo)
-
-        fpJOMaker = FullPathJOMaker(self.paths.searchPaths['jobOptionsSearchPath'], self.logger)
-
-        missingJobOptions = [jo for jo in self.jobOptions if not fpJOMaker(jo)]
-
-        if missingJobOptions:
-            msg = 'The following job option files could not be found in %s\n' % self.paths.searchPaths['jobOptionsSearchPath']
-            for jo in missingJobOptions: msg += '%s\n' % jo
-            raise RTTInputError(msg)
-            
-        def makeDirAndCopy(dir, jo, logger):
-            # jo of the form <package>/<options>.py
-            localPackageDir = os.path.join(dir, os.path.dirname(jo))
-            optionsFile     = os.path.basename(jo)
-            mkdirIfNew(localPackageDir)
-
-            fpJO = fpJOMaker(jo)
-            logger.debug('JobOption Copy: \n%s\n to %s' % (fpJO, localPackageDir))
-            shutil.copy(fpJO, localPackageDir)
-            
-        [makeDirAndCopy(self.runPath, jo, self.logger) for jo in self.jobOptions]
-
-        def addDataSets(dir, datasets, jo):
-            # calculate the string to add to the end of the joboptions to specifiy the data sets
-            def getdataset():
-                for datasetType in datasets.keys():
-                    for ds in datasets[datasetType]:
-                        yield ds
-            def clean(text):
-                # convert first '+=' to '+'    
-                return text.replace('+=', '=', 1)
-                        
-            bottomJOText = ''
-            athenaCommonFlagsText  = 'from AthenaCommon.AthenaCommonFlags import athenaCommonFlags\n'
-
-            for ds in getdataset():            
-                dataset = '["%s"]\n' % ds.physical.replace('/castor', '%s/castor' % ds.castorProtocol())
-                athenaCommonFlagsText += 'athenaCommonFlags.FilesInput += %s' % dataset
-                bottomJOText += '%s += %s\n' % (ds.whichAthenaList(), dataset)
-
-            bottomJOText = clean(bottomJOText)    
-            athenaCommonFlagsText = clean(athenaCommonFlagsText)
-            if not bottomJOText: return # no datasets
-            
-            # append the dataset string to the end of the job options
-            file_ = open(os.path.join(dir, jo))
-            contents = file_.read()
-            file_.close()
-
-            if self.athenaCommonFlags:
-                contents_  = '# The following athenaCommonFlags text was pre-pended by the RTT.\n\n' 
-                contents_ += athenaCommonFlagsText
-                contents_ += '# End of RTT pre-pending.\n'
-                contents   = contents_ + contents
-                
-            contents += '\n\n'
-            contents += '# Everything below this line was appended by the RTT.'
-            contents += '\n\n'
-            contents += bottomJOText
-
-            # Write out the new contents
-            file_ = open(os.path.join(dir, jo),'w')
-            file_.write(contents)
-            file_.close()
-                       
-        # only add datasets for athena, not athenaMT/PT (they add data via command line)
-        if self.hasData() and self.isDefaultExecutable(): 
-            [addDataSets(self.runPath, self.datasets, jo) for jo in self.jobOptions]
-            
-
-    def dataForXMLNode(self):
-        names = ['displayComponent', 'displayClass', 'displayProcess']
-        dict = WorkerJobDescriptor.dataForXMLNode(self)
-        for n in names: dict[n] = self.__dict__[n]
-        return dict
-
-    def deferredMinderUpdate(self):
-       WorkerJobDescriptor.deferredDataSetUpdate(self) 
-        
-class AthenaPilotJobDescriptor(AthenaJobDescriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict):
-
-        AthenaJobDescriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-        self.logger.debug('AthenaPilotJobDescriptor.__init__ start')
-        self.rttPilotJob = True
-        self.suggestedQueue = 'short'
-        self.logger.debug('AthenaPilotJobDescriptor.__init__ end %s' % (self.name))
-
-    def fixName(self):
-        AthenaJobDescriptor.fixName(self)
-        self.name += 'Pilot'
-        self.identifiedName   = self.name+str(self.jobSerialNumber)
-        self.log              = self.identifiedName+"_log"
-        # self.elog             = self.identifiedName+"_elog"
-
-
-    def setupRunDir(self):
-        """
-        use the base class methodf to copy job options.
-        this method modifies them to run one event.
-        """
-
-        AthenaJobDescriptor.setupRunDir(self)
-        
-        def limitEvents(dir, nevents, jo):
-
-            # append the dataset string to the end of the job options
-            fn = os.path.join(dir, jo)
-            opts = open(fn, 'r')
-            contents = opts.read()
-            opts.close()
-            contents += 'theApp.EvtMax = %d\n' % nevents
-            opts = open(fn, 'w')
-            opts.write(contents)
-            opts.close()
-                       
-
-        [limitEvents(self.runPath, 1, jo) for jo in self.jobOptions]
-
-
-
-class AthenaATNJobDescriptor(AthenaJobDescriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict):
-
-        self.optionsAndScripts = argDict.pop('optionsAndScripts', [])
-        self.atnName           = argDict.pop('atnName', '')
-
-        AthenaJobDescriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-        self.logger.debug('AthenaATNJobDescriptor.__init__ start')
-        self.rttATNJob = True
-
-        # need to recalculate work and results derectories
-        pathExtensions = [
-            ('package',    self.package),
-            ('jobGroup',   self.jobGroup),
-            ('jobName',    self.name),
-            ('SN',         str(self.jobSerialNumber))
-            ]
- 
-        self.setRunPath(pathExtensions)
-        self.logger.debug('AthenaATNJobDescriptor.__init__ end %s' % self.name)
-
-        
-    def fixJobOptions(self): pass
-    
-    def setupRunDir(self):
-        """
-        Copy the presripts passed through the ATN configuration file to the
-        run directory
-        """
-                
-        locator = FileLocator(self.paths.searchPaths['PATH'], self.logger)
-        preScripts =  [locator.locateFile(ps) for ps in self.preScripts]
-        preScripts =  [ps for ps in preScripts if not  ps is None]
-        
-
-        [shutil.copy(s, self.runPath) for s in preScripts]
-
-    def fixName(self):
-        AthenaJobDescriptor.fixName(self)
-        self.name             = self.atnName
-        self.identifiedName   = self.name+str(self.jobSerialNumber)
-        self.log              = self.identifiedName+"_log"
-        # self.elog             = self.identifiedName+"_elog"
-
-            
-    # --------------------------------------------------------------
-
-    def athenaCommand(self):
-        """
-        ATN delivers the job options and the command line flags in a
-        single string.
-        """
-
-        cmd = ''
-
-        for item in self.optionsAndScripts:                    
-            for ps in item[1]:
-                cmd += 'source %s\n' % (ps)
-
-            njob = 0
-            for jo in item[0]:
-                cmd += 'echo "Starting job %s"\n' % njob
-                cmd += '%s %s<&-\n' % (self.athenaFlavour, jo)
-                cmd += 'echo "RTT determined exit status to be: " $?'
-        return cmd
-            
-        
-                 
diff --git a/Tools/RunTimeTester/src/Bag.py.obsolete b/Tools/RunTimeTester/src/Bag.py.obsolete
deleted file mode 100644
index 2f63ab92cb96719a7cb12b21800cd3b7120abbfb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Bag.py.obsolete
+++ /dev/null
@@ -1,2 +0,0 @@
-class Bag:
-    pass
diff --git a/Tools/RunTimeTester/src/BaseDirectoryMaker.py b/Tools/RunTimeTester/src/BaseDirectoryMaker.py
deleted file mode 100755
index b7837f0a84449a702dccba958625f61235fffb83..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BaseDirectoryMaker.py
+++ /dev/null
@@ -1,67 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools        import mkdirIfNew, emptyDirMkDir
-
-#===== RTT MODULES =====
-from exc2string2              import exc2string2
-#===== PYTHON MODULES ==
-import sys, os
-
-"""
-A class to make Work, Results and Run directories.
-
-User             calls: makeDirs
-makeDirs         calls: makeBaseDirs - sets up non minder dependent parts of
-                                work and results directory tree
-"""
-
-class BaseDirectoryMaker:
-    def __init__(self, paths, logger):
-        self.paths  = paths
-        self.logger = logger
-                
-
-    def makeDir(self,parentDir):
-        try:
-            # create base directories for work directory and results
-            # directory trees
-            if os.path.exists(parentDir):
-                self.logger.debug('Emptying base directory %s' % parentDir)
-                emptyDirMkDir(parentDir)
-            else:
-                self.logger.debug('Creating base directory %s' % parentDir)
-                mkdirIfNew(parentDir)
-
-            # to permit signal in Tester to know if directories have been made
-            self.status =  True
-            msg = self.logger.debug('Base directory %s ready' % parentDir)
-        except Exception, e:
-            msg = 'Error handling base directory %s\nException%s\nTraceback\n%s' % (parentDir, str(e), exc2string2())
-            self.logger.error(msg)            
-
-    def makeDirs(self):
-        resDir = self.paths.resultsDirs['packages']
-        wrkDir = self.paths.workDirs['packages']
-
-        self.makeDir(wrkDir)
-        if resDir != wrkDir:
-                self.makeDir(resDir)
-
-
-class CERNSLC4BaseDirectoryMaker:
-    'do nothing - all the base directories are pre-made'
-    def __init__(self, paths, logger):
-        self.status =  True
-        self.logger = logger
-        self.paths  = paths
-        
-    def makeDirs(self):
-        resDir = self.paths.resultsDirs['packages']
-        wrkDir = self.paths.workDirs['packages']
-        
-        emptyDirMkDir(wrkDir)
-        if resDir != wrkDir:
-            emptyDirMkDir(resDir)
-            
-        self.logger.debug('Emptied the directories %s and %s' % (self.paths.resultsDirs['packages'],
-                                                                 self.paths.workDirs['packages']))
diff --git a/Tools/RunTimeTester/src/BaseXMLConverter.py.obsolete b/Tools/RunTimeTester/src/BaseXMLConverter.py.obsolete
deleted file mode 100755
index a5aacd8ed5ea1d552742dcb9c34debed314cda5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BaseXMLConverter.py.obsolete
+++ /dev/null
@@ -1,151 +0,0 @@
-"""
-A class to provide an aml file summary of an RTT run
-"""
-
-import xml.dom
-from XMLTools    import getText
-from xml.dom.ext import PrettyPrint
-from StringIO    import StringIO
-from Evaluate    import Evaluate
-
-import os.path, types, time
-
-#------------------------------------------------------------------------
-# Wrappers for Descriptor (and subclass) methods print method timings
-# -------------------------------------------------------------------------
-
-def minderContext(method):
-    """
-    Returns a method that wraps a the method passed in the argument.
-    The wrapper provides the context.
-    """
-
-    def contexed_method(self, *a, **k):
-
-        startTime = time.time()
-        result = method(self, *a, **k)
-        dTime = startTime - time.time()
-
-        
-        fn = '/local_disk/sherwood/RTT02/XMLtiming.dat'
-        if os.path.exists(fn):
-            of = open(fn, 'a')
-        else:
-            of = open(fn, 'w')
-
-        
-        line = 'XMLConverter: time for method %s: %d\n' % (method.__name__,
-                                                           dTime)
-        of.write(line)
-        of.close()
-        return result
-
-    return contexed_method
-
-class MetaContexter(type):
-    """
-    A meta class that loops through all methods of a class, excudes __init__
-    and wraps the rest with minderContext
-    """
-
-    def __new__(self, classname, bases, classdict):
-
-        for mn in classdict.keys():
-            if mn == '__init__': continue
-            m = classdict[mn]
-            if isinstance(m, types.FunctionType):
-                classdict[mn] = minderContext(m)
-
-        return type.__new__(self, classname, bases, classdict)
-
-class Contexter:
-    "A class to allow method wrapping through inheritance"
-
-    __metaclass__ = MetaContexter
-            
-
-from MethodTimer import MethodTimer  
-  
-class BaseXMLConverter(MethodTimer):
-    def __init__(self, elementCreator, textNodeCreator, idTagPath):
-        MethodTimer.__init__(self)
-        
-        self.elementCreator  = elementCreator
-        self.textNodeCreator = textNodeCreator
-        self.idTagPath       = idTagPath
-        self.callBack        = None
-
-    def findNode(self, newNode):
-        """
-        Example self.node = <package>
-                            <minder>
-                            <identifiedName>a</identifiedName>
-                            <identifiedName>b</identifiedName>
-                            </minder>
-                            </package>
-                newNode   = <minder>
-                            <identifiedName>b</identifiedname>
-                            </minder>
-
-        will match.
-        """
-        pathDirName  = os.path.dirname(self.idTagPath)
-        # add 'text()' here, not in argument to ensure pathDirName is
-        # calculated properly.
-        pathBaseName = os.path.join(os.path.basename(self.idTagPath),'text()')
-
-        # find the nodes witht the same structure as the new node
-        # (see assumptions stated in the doc string)
-        nodes = Evaluate(pathDirName, self.node)
-
-        matchText = Evaluate(pathBaseName, newNode)
-        matches = [n for n in nodes
-                   if Evaluate(pathBaseName, n) == matchText]
-
-        m = 'Matches: idTagPath- %s, matchText- %s, nmatch- %d' % (
-        self.idTagPath, matchText, len(matches))
-        self.logger.debug(m)
-        
-        if len(matches): return matches[0]
-        return None
-
-    # --------------------------------------------------------------------------
-    
-    def update(self, newNode):
-        matchingNode = self.findNode(newNode)
-        if matchingNode is None:
-            self.logger.debug('appending node %s to %s' % (newNode.nodeName,
-                                                           self.node.nodeName))
-            self.node.appendChild(newNode)
-        else:
-            self.node.removeChild(matchingNode)
-            self.node.appendChild(newNode)
-            self.logger.debug('replacing node %s for %s' % (newNode.nodeName,
-                                                            self.node.nodeName))
-        return self.node
-    
-    # --------------------------------------------------------------------------
-  
-    def addTextNode(self, parentNode, item):
-        try:
-            tn = self.textNodeCreator(str(item[1]).strip("'").strip())
-        except xml.dom.InvalidCharacterErr:
-            tn = self.textNodeCreator('invalid text')
-            
-        n  = self.elementCreator(str(item[0]).strip("'").strip())
-        
-        n.appendChild(tn)
-
-        parentNode.appendChild(n)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.node, temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-        
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/BaseXMLConverter2.py b/Tools/RunTimeTester/src/BaseXMLConverter2.py
deleted file mode 100755
index 95d86f5efb8a4c4c78509a99a784daf37ea6c7cb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BaseXMLConverter2.py
+++ /dev/null
@@ -1,49 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an aml file summary of an RTT run
-"""
-
-import xml.dom
-from XMLTools    import getText
-from xml.dom.ext import PrettyPrint
-from StringIO    import StringIO
-from Evaluate    import Evaluate
-
-import os.path, types, time
-
-
-from MethodTimer import MethodTimer  
-  
-class BaseXMLConverter(MethodTimer):
-    def __init__(self, elementCreator, textNodeCreator, logger):
-        self.elementCreator  = elementCreator
-        self.textNodeCreator = textNodeCreator
-        self.callBack        = None
-        self.node            = None
-        self.logger          = logger
-        MethodTimer.__init__(self)
-  
-    def addTextNode(self, parentNode, item):
-        try:
-            tn = self.textNodeCreator(str(item[1]).strip("'").strip())
-        except xml.dom.InvalidCharacterErr:
-            tn = self.textNodeCreator('invalid text')
-            
-        n  = self.elementCreator(str(item[0]).strip("'").strip())
-        
-        n.appendChild(tn)
-
-        parentNode.appendChild(n)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.node, temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-        
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/BatchJobMinder.py b/Tools/RunTimeTester/src/BatchJobMinder.py
deleted file mode 100755
index 3e55daae7cfecedf29684fecccd070723ad7dc60..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BatchJobMinder.py
+++ /dev/null
@@ -1,530 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-import os, time,string, sys
-
-from RunTimer                       import *
-from exc2string2                    import exc2string2
-from RetryTester                    import RetryTester
-from WorkerJobMinder                import WorkerJobMinder
-from ShellCommand                   import ShellCommand
-from RTTpath                        import RTTpath
-
-
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-            
-class BatchJobMinder(WorkerJobMinder):
-    
-    def __init__ (self, argBag, jDescriptor):
-        WorkerJobMinder.__init__(self, argBag, jDescriptor)
-
-        
-        self.logger.debug("initialising BatchJobMinder for job %s" % self.identifiedName)    
-
-        log  = str(RTTpath(self.runPath, self.log))
-        sp   = self.shareArea
-        rp   = self.runPath
-        # self.retryTester = RetryTester(log, elog, logger, sp, rp)
-        self.retryTester  = RetryTester(log, self.logger, sp, rp)
-        self.retryCount   =  0
-        self.retryMax     =  5
-        
-        self.pid          =  0
-        self.timedOut     = False
-        self.jobIsInBatch = False
-        self.logFileTimer = None
-        self.wallTimer    = None
-        self.batchError   = False
-
-        self.wallTimeTimeout = 72*60*60 # 72 hours
-        self.logFileTimeout  = 1*60*60 # 1
-
-        self.suggestedBatchWallTime = jDescriptor.suggestedBatchWallTime
-
-    # ------------------------------------------------------------------------
-
-    def handleBatchError(self):
-        self.batchError = True
-
-    # ------------------------------------------------------------------------
-
-    def stopJob(self):
-        pass
-
-    # -------------------------------------------------------------------------
-
-    def checkQueue(self):
-        #check if queue exists        
-        pass
-
-    # -------------------------------------------------------------------------
-
-    def jobIsRunning(self):
-        'Create the RunTimer if job is no longer in batch queues.'
-        
-
-        # self.jobIsInBatch is set true in fire()
-        queryOK = True
-        if not self.jobIsInBatch: return (queryOK, self.jobIsInBatch)
-
-        # if the job is flagged as being in the batch queue, check if this is still the case.
-        # flag failure of the interogation of the batch system.
-        queryOK, inBatch = self.isJobStillInBatch()
-        if not queryOK: return (queryOK, None)
-
-        
-        if inBatch: # job is still in batch
-            self.logger.debug('Job %s is running in batch' % str(self.pid))
-            return (queryOK, True)
-        else: # job no longer seen in batch system
-            m = 'Job %s is not visible in batch, setting a logFile timer for %d secs' % (self.pid,
-                                                                                         self.logFileTimeout)
-            self.logger.info(m)
-            self.logFileTimer = RunTimer(self.logFileTimeout, self.logger)
-            self.jobIsInBatch = False
-            return (queryOK, False)
-            
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsInQueue(self,reply): pass
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsAbsentFromQueue(self,reply): pass
-    # -------------------------------------------------------------------------
-
-    def sleepTime(self,attempt):
-        if attempt < 1: return 2
-        if attempt == 1: return 2
-        if attempt == 2: return 8
-        if attempt == 3: return 10
-
-        return 30
-
-    # -------------------------------------------------------------------------
-    
-    def isJobStillInBatch(self):
-        'Returns True is job still in batch, False otherwise.'
-        isFound = True
-        maxTry  = 5
-        attempt = -1
-        queryStat  = False
-        
-        bQuery = self.batchQuery()
-        while True:
-            attempt += 1
-                    
-            if attempt==maxTry:
-                # give up on this job
-                msg  = 'Abandoning attempt to query batch sys after %d attempts %s ' % (attempt,
-                                                                                        self.identifiedName)
-                self.logger.error(msg)
-                queryStat = False
-                isFound   = False
-                return (queryStat, isFound)
-
-
-            bReply = ShellCommand([bQuery], self.logger, 100, None, False).getReply()
-
-            if not self.batchQueryReplyFormatOK(bReply):
-
-                msg  = 'Attempt number: %d\nBad reply: command = %s; reply = %s; extra info =%s ' % (
-                    attempt,
-                    bQuery,
-                    bReply,
-                    self.getExtraInfoOnBatchQueryCommand())
-                self.logger.warning(msg)                    
-                stime = self.sleepTime(attempt)
-                m = 'sleep for %d seconds before next attempt to interrogate batch system with %s.' % (stime, bQuery)
-                self.logger.debug(m)
-                time.sleep(stime)
-                continue
-
-            if self.batchReplyProvesJobIsInQueue(bReply):
-                queryStat = True
-                isFound   = True
-                self.logger.debug('Job proved to be in queue')
-                return (queryStat,isFound)
-            elif self.batchReplyProvesJobIsAbsentFromQueue(bReply):
-                self.logger.debug('Job proved not to be in queue')
-                queryStat = True
-                isFound   = False
-                return (queryStat,isFound)
-            else:
-                msg = 'Cannot determine if job is in queue from reply: %s, will resissue query' % bReply
-                stime = self.sleepTime(attempt)
-                m = 'sleep for %d seconds before next attempt to interrogate batch system with %s.' % (stime, bQuery)
-                self.logger.debug(m)
-                time.sleep(stime)
-                continue
-    
-    # -------------------------------------------------------------------------
-
-    def getExtraInfoOnBatchQueryCommand(self): 
-        extraInfoCmd = ['echo "Echoing PATH variable"',
-                        'echo $PATH',
-                        'echo ""',
-                        'which %s' % self.batchQueryCommand()]
-        
-        try:
-            sc = ShellCommand(extraInfoCmd, self.logger)
-        except Exception, e:
-            reply = ['Exception was thrown trying to use ShellCommand to gather extra information on bjobs.']
-            reply.append(exc2string2())
-            reply.append(str(e))
-        else:
-            reply = sc.getReply()
-
-
-        replyString = ''
-        for line in reply:
-            replyString += line + '\n'
-        return replyString
-
-    # -------------------------------------------------------------------------
-    
-    def areLogFilesInRunDir(self, fn):        
-        files = os.listdir(self.runPath)
-
-        if fn in files:
-            m = "Found the log file %s in rundir %s" % (fn, self.runPath)            
-            self.logger.debug(m)
-
-            return True
-
-        m = "Log file %s NOT found in rundir %s" % (fn, self.runPath)            
-        self.logger.debug(m)
-
-        return False
-
-    # -------------------------------------------------------------------------
-
-    def retryTest(self):
-
-        if not self.logResult == 'error': return False
-        
-        if self.retryTester.shouldRetry():
-            m ='retryTest() %d %d' % (self.retryCount,
-                                      self.retryMax)
-            self.logger.debug(m)
-            if self.retryCount < self.retryMax:
-                self.retryCount += 1
-                return True
-            
-        return False
-
-    # -------------------------------------------------------------------------
-
-    def prepareForRetry(self):
-
-        logs = [str(RTTpath(self.runPath, self.log))]
-        
-        for log in logs:
-            if os.path.exists(log):
-                dest = log+'_'+str(self.errorStateCurRetry)
-                m = 'State retry: moving file %s to %s' % (log, dest)
-                self.logger.debug(m)
-                os.rename(log, dest)
-                self.makeKeepFileEntry(os.path.basename(dest),
-                                       "Log file leading to a retry")
-
-        self.logResult = 'unavailable'
-        self.logFileTimer  = None
-        self.wallTimer     = None
-
-    # -------------------------------------------------------------------------
-
-    def checkForWallTimeTimeOut(self): return self.checkForTimeOut(self.wallTimer)
-
-    # -------------------------------------------------------------------------
-
-    def checkForLogFileTimeOut(self): return self.checkForTimeOut(self.logFileTimer)
-
-    # -------------------------------------------------------------------------
-
-    def checkForTimeOut(self, timer):
-
-        if not timer: return False
-        
-        if not timer.timeOut():
-            self.timedOut = False
-            return False
-        else:
-            self.timedOut = True
-            return True
-            
-    # -------------------------------------------------------------------------
-
-    def fire(self, what):
-
-        if not self.checkQueue(): return False
-
-        self.logFileTimer = None # remove any time from a previous script
-        self.wallTimer    = RunTimer(self.wallTimeTimeout, self.logger)
-        self.jobIsInBatch = True
-        
-        WorkerJobMinder.fire(self, what)
-
-        self.setPid()
-        # time.sleep(10) # let the jobs reach the batch queues
-        self.logger.debug('Minder: pid = %s' % self.pid)
-
-    # -------------------------------------------------------------------------
-
-    def dataForMonitoring(self):
-        dict = WorkerJobMinder.dataForMonitoring(self)
-        dict['timedOut']   = self.timedOut
-        dict['batchError'] = self.batchError
-        return dict
-    # ------------------------------------------------------------------------
-    
-    def status(self):
-        
-        states  = {
-                   'timedOut':          self.timedOut,
-                   'batchError':        self.batchError
-                   }
-        for s in states.keys():
-            if states[s]:  return s
-
-        # if self.logResult == 'unavailable': return self.stateEngine.state.state
-        if self.logResult == 'unavailable': return self.currentState()
-        return self.logResult
-
-    # -------------------------------------------------------------------------
-
-    def __str__(self):
-        s  = '|-------------------------------------------|\n'
-        s += '|                                           |\n'
-        s += '|           Batch  Minder  dump           |\n'
-        s += '|                                           |\n'
-        s += '|-------------------------------------------|\n'
-        s += WorkerJobMinder.__str__(self)+'\n'
-        return s
-
-    # -------------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug(self.__str__())
-
-    # -------------------------------------------------------------------------
-
-    def fromQueued(self):
-        self.fire('runScript')
-
-        if not self.checkJobSubmissionWasOK(): return 'batchError'
-        return 'running'
-
-
-    # -------------------------------------------------------------------------
-
-    def fromRunning(self):
-        "Running is complete if the log files have returned"
-
-        id  = self.pid
-        name = self.identifiedName
-
-        if self.checkForWallTimeTimeOut():
-            self.diskspace('fromRunning, job has wall timed out')
-            self.logger.warning('BatchMinderStateEngine: wall time timeout %s %s' % (name, id) )
-            return 'timedOut'
-
-        # query if job is in batch queue
-        queryOK, inBatch = self.jobIsRunning()
-
-        # give up if the query failed
-        if not queryOK:
-            self.diskspace('fromRunning, job has batch error')
-            self.logger.debug('BatchMinderJob::fromRunning(), batch system failed to return intelligible response, abandon: %s %s' % (name, id) )
-            return 'batchError'
-
-        # if the job is still running, no state transition.
-        if inBatch:
-            self.logger.debug('Job is found to be in batch.')
-            try:
-                self.audit()
-            except:
-                self.logger.error('Exception during auditing: %s' % exc2string2())
-            return 'running'
-        
-        self.logger.debug('Job not seen in batch, checking for log file')
-
-        # Has the batch job log file come back to the run dir?
-        if not self.areLogFilesInRunDir(self.log):
-            self.logger.debug('Log file not found in run dir.')
-            if not self.checkForLogFileTimeOut(): return 'running'
-            self.logger.warning('Time out has occured, never found job log file.')
-            self.diskspace('Batch log absent, timing out')
-            return 'timedOut'
-
-        self.logger.debug('Found batch log file in run dir.')
-
-        self.runningActions() # must perform log checks before retry test.
-
-        if self.retryTest():
-            self.diskspace('Going to retry job')
-            self.logger.debug('Conditions for retry are met, retrying...')
-            return 'retry'
-
-        if self.doPostProcessing():
-            self.diskspace('Job done, going to postprocessing')
-            self.logger.debug('Job has postprocessing, will do that')
-            return 'postProcessingQueued'
-
-        self.logger.debug('No postprocessing for this job, cleaning up.')
-        self.diskspace('Job done, going to clean space')
-        self.cleanSpace()
-        self.diskspace('Job done, clean space is done.')
-        return 'done'
-
-    # -------------------------------------------------------------------------
-
-    def archive(self):
-        self.logger.debug('Going to archive...')
-        [a.archive(self.runPath) for a in self.archivers]   
-        self.logger.debug('Finished archiving...')
-
-        # Pop files that were archived from the keepfile dict
-        [self.keepFilesToVeto.extend(a.copiedFiles()) for a in self.archivers]
-
-    # -------------------------------------------------------------------------
-
-    def fromOperatorCloseDown(self):
-       self.stopJob()
-       self.setOperatorCloseDown()
-       return 'done'
-   
-    # -------------------------------------------------------------------------
-
-    def fromPostProcessingQueued(self):
-        self.fire('postProcessorScript')
-        return 'postProcessingRunning'
-
-    # -------------------------------------------------------------------------
-
-    def fromPostProcessingRunning(self):
-
-        if self.checkForWallTimeTimeOut():
-            self.logger.warning('Postprocessing walltime timeout')
-            self.diskspace('Wall timeout for postprocessing')
-            return 'timedOut'
-
-        queryOK, isRunning =  self.jobIsRunning()
-        if isRunning:  
-            return 'postProcessingRunning'
-
-        # is the post processir log file visible?
-        if not self.areLogFilesInRunDir('postProcessor.log'):
-            id  = self.pid
-            name = self.identifiedName
-            self.logger.debug('No post processing log file seen yet')
-            if not self.checkForLogFileTimeOut(): 
-                return 'postProcessingRunning'
-
-            self.logger.warning('Post processing log file never found, timing out.')
-            self.diskspace('Postprocessing timeout, batch log absent')
-            return 'timedOut'
-        
-        self.readTestResults()
-
-        self.diskspace('Postprocessing done, about to archive.')
-
-        # NOTE!!!!
-        # Archiving MUST happen before copyKeepFiles to prevent file duplication.        
-        self.archive() # Only for batch jobs, not linux interactive        
-        self.diskspace('Postprocessing done, archiving done.')
-        
-        self.copyKeepFiles() # grab any new keep files
-
-        self.diskspace('Postprocessing done, about to clean space.')        
-        self.cleanSpace()
-        self.diskspace('Postprocessing done, cleanup done')
-        return 'done'
-
-
-    # -------------------------------------------------------------------------
-
-    def fromRetry(self):
-
-        # the presence of the log files signals the end of the running
-        # phase - remove the files,and ensure they are copied to the
-        # results web pages.
-
-        self.prepareForRetry()
-        return 'queued'
-        
-
-    # -------------------------------------------------------------------------
-
-    def fromBatchError(self):
-        self.handleBatchError()
-        return 'done'
-
-    # -------------------------------------------------------------------------
-
-    def fromTimedOut(self):
-        'A marker state'
-
-        self.logger.info('Job timed out, remove from batch %s %s' % (self.pid, self.identifiedName)) 
-        self.stopJob()
-        return 'done'
-        
-    # -------------------------------------------------------------------------
-
-    def fromDone(self):
-        self.xmlConverter.makeSummaryReport()
-        self.setDone()
-        return 'done'
-
-    # -------------------------------------------------------------------------
-
-    def runningActions(self):
-        """Archiving MUST happen before copyKeepFiles to prevent file duplication."""
-
-        self.audit()
-        dirList = os.listdir(self.runPath)
-        self.logger.debug('Run dir contents at the end of running state %s' % str(dirList))
-        self.checkLog()
-
-        self.runChecks()
-
-        # Calculate disk space usage by the job, probably at a maximum just before archiving
-        self.diskspace('Job done, about to archive')
-
-        # NOTE!!!!
-        # Archiving MUST happen before copyKeepFiles to prevent file duplication.
-        self.archive() # Only for batch jobs, not linux interactive   
-        self.diskspace('Job done, archiving is done.')
-        self.copyKeepFiles()
-        self.diskspace('Job done, keepfiles copied.')
-
-    # -------------------------------------------------------------------------
-    # methods to be supplied by subclasses
-    # -------------------------------------------------------------------------
-
-    def batchQuery(self):
-        raise NotImplementedError()
-
-    # -------------------------------------------------------------------------
-
-    def batchQueryReplyFormatOK(self, bReply):
-        raise NotImplementedError()
-    # -------------------------------------------------------------------------
-
-    def batchQueryCmd(self):
-        raise NotImplementedError()
-
-    # -------------------------------------------------------------------------
-
-    def checkJobSubmissionWasOK(self):
-        raise NotImplementedError()
-
-    # -------------------------------------------------------------------------
-
-    def setPid(self):
-        raise NotImplementedError()
-
-    # -------------------------------------------------------------------------
-
diff --git a/Tools/RunTimeTester/src/BatchJobMinder.py.obsolete b/Tools/RunTimeTester/src/BatchJobMinder.py.obsolete
deleted file mode 100755
index 08baba3357d4428d764b50bfac94854ae785ae6e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BatchJobMinder.py.obsolete
+++ /dev/null
@@ -1,306 +0,0 @@
-
-import os, time,string, sys
-
-from RunTimer                       import *
-from exc2string2                    import exc2string2
-from RetryTester                    import RetryTester
-from WorkerMinder                   import WorkerJobMinder
-from ShellCommand                   import ShellCommand
-from RTTpath                        import RTTpath
-
-
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-            
-class BatchJobMinder(WorkerJobMinder):
-    
-    def __init__ (self, argBag, jDescriptor):
-        WorkerJobMinder.__init__(self, argBag, jDescriptor)
-
-        
-        self.logger.debug("initialising BatchJobMinder for job %s" % self.identifiedName)    
-
-        log  = str(RTTpath(self.runPath, self.log))
-        sp   = self.shareArea
-        rp   = self.runPath
-        # self.retryTester = RetryTester(log, elog, logger, sp, rp)
-        self.retryTester  = RetryTester(log, self.logger, sp, rp)
-        self.retryCount   =  0
-        self.retryMax     =  5
-        
-        self.pid          =  0
-        self.timedOut     = False
-        self.jobIsInBatch = False
-        self.logFileTimer = None
-        self.wallTimer    = None
-        self.batchError   = False
-
-        self.wallTimeTimeout = 72*60*60 # 72 hours
-        self.logFileTimeout  = 1*60*60 # 1
-
-        self.suggestedBatchWallTime = jDescriptor.suggestedBatchWallTime
-
-    # ------------------------------------------------------------------------
-
-    def handleBatchError(self):
-        self.batchError = True
-
-    # ------------------------------------------------------------------------
-
-    def stopJob(self):
-        pass
-
-    # -------------------------------------------------------------------------
-
-    def checkQueue(self):
-        #check if queue exists        
-        pass
-
-    # -------------------------------------------------------------------------
-
-    def jobIsRunning(self):
-        'Create the RunTimer if job is no longer in batch queues.'
-        
-
-        # self.jobIsInBatch is set true in fire()
-        queryOK = True
-        if not self.jobIsInBatch: return (queryOK, self.jobIsInBatch)
-
-        # if the job is flagged as being in the batch queue, check if this is still the case.
-        # flag failure of the interogation of the batch system.
-        queryOK, inBatch = self.isJobStillInBatch()
-        if not queryOK: return (queryOK, None)
-
-        
-        if inBatch: # job is still in batch
-            self.logger.debug('Job %s is running in batch' % str(self.pid))
-            return (queryOK, True)
-        else: # job no longer seen in batch system
-            m = 'Job %s is not visible in batch, setting a logFile timer for %d secs' % (self.pid,
-                                                                                         self.logFileTimeout)
-            self.logger.info(m)
-            self.logFileTimer = RunTimer(self.logFileTimeout, self.logger)
-            self.jobIsInBatch = False
-            return (queryOK, False)
-            
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsInQueue(self,reply): pass
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsAbsentFromQueue(self,reply): pass
-    # -------------------------------------------------------------------------
-
-    def sleepTime(self,attempt):
-        if attempt < 1: return 2
-        if attempt == 1: return 2
-        if attempt == 2: return 8
-        if attempt == 3: return 10
-
-        return 30
-
-    # -------------------------------------------------------------------------
-    
-    def isJobStillInBatch(self):
-        'Returns True is job still in batch, False otherwise.'
-        isFound = True
-        maxTry  = 5
-        attempt = -1
-        queryStat  = False
-        
-        bQuery = self.batchQuery()
-        while True:
-            attempt += 1
-                    
-            if attempt==maxTry:
-                # give up on this job
-                msg  = 'Abandoning attempt to query batch sys after %d attempts %s ' % (attempt,
-                                                                                        self.identifiedName)
-                self.logger.error(msg)
-                queryStat = False
-                isFound   = False
-                return (queryStat, isFound)
-
-
-            bReply = ShellCommand([bQuery], self.logger, 100, None, False).getReply()
-
-            if not self.batchQueryReplyFormatOK(bReply):
-
-                msg  = 'Attempt number: %d\nBad reply: command = %s; reply = %s; extra info =%s ' % (
-                    attempt,
-                    bQuery,
-                    bReply,
-                    self.getExtraInfoOnBatchQueryCommand())
-                self.logger.warning(msg)                    
-                stime = self.sleepTime(attempt)
-                m = 'sleep for %d seconds before next attempt to interrogate batch system with %s.' % (stime, bQuery)
-                self.logger.debug(m)
-                time.sleep(stime)
-                continue
-
-            if self.batchReplyProvesJobIsInQueue(bReply):
-                queryStat = True
-                isFound   = True
-                self.logger.debug('Job proved to be in queue')
-                return (queryStat,isFound)
-            elif self.batchReplyProvesJobIsAbsentFromQueue(bReply):
-                self.logger.debug('Job proved not to be in queue')
-                queryStat = True
-                isFound   = False
-                return (queryStat,isFound)
-            else:
-                msg = 'Cannot determine if job is in queue from reply: %s, will resissue query' % bReply
-                stime = self.sleepTime(attempt)
-                m = 'sleep for %d seconds before next attempt to interrogate batch system with %s.' % (stime, bQuery)
-                self.logger.debug(m)
-                time.sleep(stime)
-                continue
-    
-    # -------------------------------------------------------------------------
-
-    def getExtraInfoOnBatchQueryCommand(self): 
-        extraInfoCmd = ['echo "Echoing PATH variable"',
-                        'echo $PATH',
-                        'echo ""',
-                        'which %s' % self.batchQueryCommand()]
-        
-        try:
-            sc = ShellCommand(extraInfoCmd, self.logger)
-        except Exception, e:
-            reply = ['Exception was thrown trying to use ShellCommand to gather extra information on bjobs.']
-            reply.append(exc2string2())
-            reply.append(str(e))
-        else:
-            reply = sc.getReply()
-
-
-        replyString = ''
-        for line in reply:
-            replyString += line + '\n'
-        return replyString
-
-    # -------------------------------------------------------------------------
-    
-    def areLogFilesInRunDir(self):        
-        files = os.listdir(self.runPath)
-
-        if self.log in files:
-            m = "Found the log file %s in rundir %s"
-            m = m % (self.log, self.runPath)            
-            self.logger.debug(m)
-
-            return True
-
-        return False
-
-    # -------------------------------------------------------------------------
-
-    def retryTest(self):
-
-        if not self.logResult == 'error': return False
-        
-        if self.retryTester.shouldRetry():
-            m ='retryTest() %d %d' % (self.retryCount,
-                                      self.retryMax)
-            self.logger.debug(m)
-            if self.retryCount < self.retryMax:
-                self.retryCount += 1
-                return True
-            
-        return False
-
-    # -------------------------------------------------------------------------
-
-    def prepareForRetry(self):
-
-        logs = [str(RTTpath(self.runPath, self.log))]
-        
-        for log in logs:
-            if os.path.exists(log):
-                dest = log+'_'+str(self.errorStateCurRetry)
-                m = 'State retry: moving file %s to %s' % (log, dest)
-                self.logger.debug(m)
-                os.rename(log, dest)
-                self.makeKeepFileEntry(os.path.basename(dest),
-                                       "Log file leading to a retry")
-
-        self.logResult = 'unavailable'
-        self.logFileTimer  = None
-        self.wallTimer     = None
-
-    # -------------------------------------------------------------------------
-
-    def checkForWallTimeTimeOut(self): return self.checkForTimeOut(self.wallTimer)
-
-    # -------------------------------------------------------------------------
-
-    def checkForLogFileTimeOut(self): return self.checkForTimeOut(self.logFileTimer)
-
-    # -------------------------------------------------------------------------
-
-    def checkForTimeOut(self, timer):
-
-        if not timer: return False
-        
-        if not timer.timeOut():
-            self.timedOut = False
-            return False
-        else:
-            self.timedOut = True
-            return True
-            
-    # -------------------------------------------------------------------------
-
-    def fire(self, what):
-
-        if not self.checkQueue(): return False
-
-        self.logFileTimer = None # remove any time from a previous script
-        self.wallTimer    = RunTimer(self.wallTimeTimeout, self.logger)
-        self.jobIsInBatch = True
-        
-        WorkerJobMinder.fire(self, what)
-
-        self.setPid()
-        # time.sleep(10) # let the jobs reach the batch queues
-        self.logger.debug('Minder: pid = %s' % self.pid)
-
-    # -------------------------------------------------------------------------
-
-    def dataForMonitoring(self):
-        dict = WorkerJobMinder.dataForMonitoring(self)
-        dict['timedOut']   = self.timedOut
-        dict['batchError'] = self.batchError
-        return dict
-    # ------------------------------------------------------------------------
-    
-    def status(self):
-        
-        states  = {'operatorCloseDown': self.operatorCloseDown,
-                   'timedOut':          self.timedOut,
-                   'batchError':        self.batchError
-                   }
-        for s in states.keys():
-            if states[s]:  return s
-
-        if self.logResult == 'unavailable': return self.stateEngine.state.state
-        return self.logResult
-
-    # -------------------------------------------------------------------------
-
-    def __str__(self):
-        s  = '|-------------------------------------------|\n'
-        s += '|                                           |\n'
-        s += '|           Batch  Minder  dump           |\n'
-        s += '|                                           |\n'
-        s += '|-------------------------------------------|\n'
-        s += WorkerJobMinder.__str__(self)+'\n'
-        return s
-
-    # -------------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug(self.__str__())
-
diff --git a/Tools/RunTimeTester/src/BatchMinderStateEngine.py b/Tools/RunTimeTester/src/BatchMinderStateEngine.py
deleted file mode 100755
index 6b3b4cdcbf6a032c6fba5982783eac0f5d8f65c5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BatchMinderStateEngine.py
+++ /dev/null
@@ -1,28 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""State Engine for LinuxJobMinders: extends WorkerJobMinderStateEngine
-
-overrides: fromQueuedTest
-           fromRunningTest
-
-"""
-
-from MinderStateEngineBase  import MinderStateEngineBase
-
-class BatchMinderStateEngine(MinderStateEngineBase):
-    
-    def __init__(self, minder, state = 'queued'):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-        self.dispatchTable =  {
-            'queued':                minder.fromQueued,
-            'running':               minder.fromRunning,
-            'operatorCloseDown':     minder.fromOperatorCloseDown,
-            'postProcessingQueued':  minder.fromPostProcessingQueued,
-            'postProcessingRunning': minder.fromPostProcessingRunning,
-            'operatorCloseDown':     minder.fromOperatorCloseDown,
-            'retry':                 minder.fromRetry,
-            'batchError':            minder.fromBatchError,
-            'timedOut':              minder.fromTimedOut,
-            'done':                  minder.fromDone
-            }
-
diff --git a/Tools/RunTimeTester/src/BatchMinderStateEngine.py.obsolete b/Tools/RunTimeTester/src/BatchMinderStateEngine.py.obsolete
deleted file mode 100755
index 791a3f8ad5c2952e7cddb95952011ebb721141b5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BatchMinderStateEngine.py.obsolete
+++ /dev/null
@@ -1,190 +0,0 @@
-"""State Engine for PBSJobMinders: extends WorkerJobMinderStateEngine
-
-overrides: fromQueued
-           fromQueuedTest
-           fromRunning
-
-removes transitions
-
-
-adds state transitions
-
-queued   -> pbsError
-running  -> retry
-retry    -> queued
-pbsError -> done
-"""
-
-from WorkerMinderStateEngine  import WorkerMinderStateEngine
-from exc2string2              import exc2string2
-
-import os
-
-class BatchMinderStateEngine(WorkerMinderStateEngine):
-    'Overrides for Batch'
-    
-    def __init__(self, minder, state='queued'):
-        WorkerMinderStateEngine.__init__(self, minder, state)
-
-        newDispatcher = {
-            'retry':      self.fromRetry,
-            'batchError': self.fromBatchError,
-            'timedOut':   self.fromTimedOut
-            }
-        self.dispatcher.update(newDispatcher)
-
-        
-    # -----------------------------------------------
-
-    def functionsMinderMustHave(self):
-        return ['fire', 'checkJobSubmissionWasOK', 'checkForWallTimeTimeOut', 'jobIsRunning',
-                'checkForLogFileTimeOut', 'retryTest', 'doPostProcessing', 'cleanSpace',
-                'audit', 'prepareForRetry', 'stopJob', 'handleBatchError', 'areLogFilesInRunDir']
-    # -----------------------------------------------
-
-    def fromQueuedTest(self): assert(False) # base method refused
-            
-    # -----------------------------------------------
-    
-    def fromQueued(self):
-
-
-        self.minder.fire('runScript')
-                    
-        if not self.minder.checkJobSubmissionWasOK():
-            self.state.set('batchError')
-            return
-
-
-        self.state.set('running')
-        # othewise stay in 'queued'
-
-    # -----------------------------------------------
-
-    def fromRunning(self):
-        "Running is complete if the log files have returned"
-
-        id  = self.minder.pid
-        name = self.minder.identifiedName
-
-        if self.minder.checkForWallTimeTimeOut():
-            self.logger.warning('BatchMinderStateEngine: wall time timeout %s %s' % (name, id) )
-            self.state.set('timedOut')
-            return
-
-
-
-        # query if job is in batch queue
-        queryOK, inBatch = self.minder.jobIsRunning()
-        # give up if the query failed
-        self.logger.debug('BatchMinderStateEngine: check if job is running job %s %s' % (name, id))
-        if not queryOK:
-            self.logger.debug('BatchMinderStateEngine: batch query failed job %s %s' % (name, id) )
-            self.state.set('batchError')
-            return
-
-        # if the job is still running, no state transition.
-        if inBatch:
-            self.logger.debug('BatchMinderStateEngine: job is running %s %s' % (name, id) )
-            self.whileRunningActions()
-            return
-        
-        self.logger.debug('BatchMinderStateEngine: job not seen in batch %s %s' % (name, id) )
-        # are the log files visible?
-        if not self.minder.areLogFilesInRunDir():
-            self.logger.debug('BatchMinderStateEngine: no log file seen %s %s' % (name, id) )
-            if self.minder.checkForLogFileTimeOut():
-                self.logger.warning('BatchMinderStateEngine: no log file seen, timed out %s %s' % (name, id) )
-                self.state.set('timedOut')
-                return
-            # waiting for log files to return: no state change
-            return
-
-        self.logger.debug('BatchMinderStateEngine: log file seen %s %s' % (name, id) )
-
-        self.fromRunningActions() # must perform log checks before retry test.
-
-        self.logger.debug('BatchMinderStateEngine: actions complete %s %s' % (name, id) )
-
-        if self.minder.retryTest():
-            self.logger.debug('BatchMinderStateEngine: setting retry %s %s' % (name, id) )
-            self.state.set('retry')
-            return
-
-        self.logger.debug('BatchMinderStateEngine: starting postprocessing %s %s' % (name, id) )
-
-        if self.minder.doPostProcessing():
-            self.state.set('postProcessingQueued')
-            return
-
-        self.logger.debug('BatchMinderStateEngine: pp complete, setting done %s %s' % (name, id) )
-
-        self.minder.cleanSpace()
-        
-        self.state.set('done')
-        
-    # -------------------------------------------------------------------------
-
-    def whileRunningActions(self):
-        try:
-            self.minder.audit()
-        except:
-            self.logger.error('exception during auditing: %s' % exc2string2())
-
-    # -------------------------------------------------------------------------
-
-    def fromRunningActions(self):
-        self.minder.audit()
-        dirList = os.listdir(self.minder.runPath)
-        self.logger.debug('Run dir contents at the end of running state %s' % str(dirList))
-        WorkerMinderStateEngine.fromRunningActions(self)
-        
-        
-    # -------------------------------------------------------------------------
-    
-    def fromRetry(self):
-
-        # the presence of the log files signals the end of the running
-        # phase - remove the files,and ensure they are copied to the
-        # results web pages.
-
-        self.minder.prepareForRetry()
-        self.state.set('queued')
-        
-    # -----------------------------------------------
-        
-    def fromTimedOut(self):
-        'A marker state'
-
-        #if self.minder.doPostProcessing():
-        #    self.state.set('postProcessingQueued')
-        #    return
-
-        id  = self.minder.pid
-        name = self.minder.identifiedName
-        self.logger.info('Job timed out, remove from batch %s %s' % (id, name)) 
-        self.minder.stopJob()
-        self.state.set('done')
-        
-        
-    # -----------------------------------------------
-        
-    def fromBatchError(self):
-        self.minder.handleBatchError()
-        self.state.set('done')
-        
-    def fromOperatorCloseDown(self):
-        self.minder.stopJob()
-        WorkerMinderStateEngine.fromOperatorCloseDown(self)
-
-    def fromPostProcessingRunning(self):
-        
-        id  = self.minder.pid
-        name = self.minder.identifiedName
-
-        if self.minder.checkForWallTimeTimeOut():
-            self.logger.warning('BatchMinderStateEngine: wall time timeout %s %s' % (name, id) )
-            self.state.set('timedOut')
-            return
-
-        WorkerMinderStateEngine.fromPostProcessingRunning(self)
diff --git a/Tools/RunTimeTester/src/BigFileIO.py b/Tools/RunTimeTester/src/BigFileIO.py
deleted file mode 100644
index 27ddf6d8f04133d52bf7971a9f75fc00e68a6500..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BigFileIO.py
+++ /dev/null
@@ -1,131 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-from os.path import join, isdir, exists
-from Tools2 import getFileOrDirDateStamp
-from Tools import emptyDirMkDir, mkdirIfNew
-
-# Here's how it works:
-#    We have a base volume location /some/base/path
-#    Below which we have some volumes, say 2, called X and Y
-#    For a dev run, rel_2, SLC4 opt, AFS build, AtlasProduction...
-#    Procedure:
-#       1. Path creation
-#            For this example, the path, relative to base volume+X or Y, is:
-#               nightly/dev/build/i686-slc4-gcc34-opt/offline
-#            If this relative path exists:    
-#            a) below neither X or Y ==> pick either X or Y, create path below it
-#            b) below one of X or Y  ==> create path below the other
-#            c) below both X and Y   ==> compare time stamps on the time_stamp file
-#                                        that exists in relative path dir and pick
-#                                        the older of the two.
-# 
-
-bigFileIO = None
-
-def createBigFileIO(site, bigFilesArea, workDirs, isNightly):
-    dict = {'CERNslc4':  CERNBigFileIO}
-    big = dict.get(site, BigFileIO)
-    global bigFileIO
-    if not bigFileIO:
-        bigFileIO = big(bigFilesArea, workDirs, isNightly)
-    return bigFileIO
-
-class BigFileIO:
-    """Responsible for creating and yielding the path to the volume
-    where big files will be dumped."""
-
-    def __init__(self, bigFilesArea, workDirs, isNightly ):
-
-        self.bigFilesArea = bigFilesArea
-        self.workDirs     = workDirs
-        self.isNightly    = isNightly
-        
-        self.setRelativeBaseDumpLocation()
-        self.chooseDumpLocation()
-        if hasattr(self, 'pathToDump'):
-            self.createDumpLocation()
-            self.deleteDumpLocationContents()
-            self.createTimeStamp()
-
-                    
-    def createDumpLocation(self):
-        mkdirIfNew(self.pathToDump)
-
-    def deleteDumpLocationContents(self):
-        emptyDirMkDir(self.pathToDump)
-
-    def createTimeStamp(self):
-        stampFile = join(self.pathToDump, 'time_stamp')
-        if not exists(stampFile):
-            os.system('touch %s' % stampFile)
-        
-    def cmpTimeStamps(self, a, b):
-        aStamp = getFileOrDirDateStamp(a)
-        bStamp = getFileOrDirDateStamp(b)
-        cmp = 0
-        if aStamp < bStamp: cmp = -1
-        if aStamp > bStamp: cmp = 1
-        return cmp
-        
-    def chooseDumpLocation(self):
-        if not self.relLocation: return
-        area = self.bigFilesArea
-        tgts = [join(area, d, self.relLocation) for d in os.listdir(area) if isdir(join(area, d))]
-        
-        notExists = [t for t in tgts if not exists(t)]
-        if notExists:
-            self.pathToDump = notExists[0]
-        else:
-            tgts = [join(t, 'time_stamp') for t in tgts]
-            notExists = [t for t in tgts if not exists(t)]
-            if notExists:
-                self.pathToDump = os.path.dirname(notExists[0])
-            else:
-                tgts.sort(self.cmpTimeStamps)
-                self.pathToDump = os.path.dirname(tgts[0])
-            
-    def setRelativeBaseDumpLocation(self):
-        if not self.bigFilesArea or not exists(self.bigFilesArea):
-            self.relLocation = None
-            return
-
-        self.relLocation = self.workDirs['packages'].split(self.workDirs['base'])[1]
-        if self.relLocation.startswith('/'): self.relLocation = self.relLocation[1:]
-
-        # replace rel_X with 'nightly'
-        if self.isNightly:
-            self.relLocation = 'nightly/' + self.relLocation[self.relLocation.find('/')+1:]
-
-    def getJobDumpLocation(self, minder):
-        relPath = minder.runPath.split(minder.workDirs['packages'])[1]
-        if relPath.startswith('/'): relPath = relPath[1:]
-
-        try:
-            descDumpLoc = join(self.pathToDump, relPath)
-        except:
-            descDumpLoc = None
-        
-        if descDumpLoc and not exists(descDumpLoc):
-            mkdirIfNew(descDumpLoc)
-
-        return descDumpLoc
-
-class CERNBigFileIO(BigFileIO):
-    def __init__(self, bigFilesArea, workDirs, isNightly):
-        BigFileIO.__init__(self, bigFilesArea, workDirs, isNightly)
-        
-    # Do not make it for CERN as volume hangs
-    # below the path, not above it
-    def createDumpLocation(self):
-        if not exists(self.pathToDump):
-            # this will stop big file moving
-            self.pathToDump = None
-
-    def deleteDumpLocationContents(self):
-        if not self.pathToDump or not exists(self.pathToDump): return
-        BigFileIO.deleteDumpLocationContents(self)
-        
-    def createTimeStamp(self):
-        if not self.pathToDump: return
-        BigFileIO.createTimeStamp(self)
diff --git a/Tools/RunTimeTester/src/BlackLister.py.obsolete b/Tools/RunTimeTester/src/BlackLister.py.obsolete
deleted file mode 100755
index ffa8054274bab1782ec7e4efe087341f0c4ab750..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/BlackLister.py.obsolete
+++ /dev/null
@@ -1,42 +0,0 @@
-class BlackLister:
-    def __init__(self,testRuns):
-        # first entry=smallest offender, last entry=worst offender
-        self.blackList = ['AthenaCaloRecEx','AthenaCaloDigEx','AthenaCaloSimEx','RuleChecker','AthenaMissingET']
-        self.tr = testRuns
-
-    def punish(self):
-        punished = []
-        for group in self.blackList:
-            for tr in self.tr:
-                if group in tr.groupsInTestRun():
-                    punished.append(tr)
-                    break
-
-        finalList = [tr for tr in self.tr if tr not in punished]
-        finalList.extend(punished)
-        return finalList
-                
-#-------- TESTS ---------------
-
-if __name__ == '__main__':
-    class Blah:
-        def __init__(self,groups,name):
-            self.groups = groups
-            self.name = name
-            
-        def groupsInTestRun(self):
-            return self.groups
-        
-                
-    testRuns = [Blah([u'AthenaAtlfast'],'Atlfast'),
-                Blah([u'AthenaMissingET'],'MissingET'),
-                Blah([u'AthenaG4AtlasApps'],'G4AtlasApps'),
-                Blah([u'AthenaInDetRTT'],'InDetRTT'),
-                Blah([u'AthenaCaloRecEx'],'CaloRec'),
-                Blah(['RuleChecker', 'KitValidation', 'RuleCheckerWatcher'],'RC'),
-                ]
-    
-    print [b.name for b in testRuns]
-    testRuns = BlackLister(testRuns).punish()
-    print [b.name for b in testRuns]
-    
diff --git a/Tools/RunTimeTester/src/Borg.py.obsolete b/Tools/RunTimeTester/src/Borg.py.obsolete
deleted file mode 100755
index 3ba4eb9db9195926cb259d976347c752fb11d0b3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Borg.py.obsolete
+++ /dev/null
@@ -1,9 +0,0 @@
-"""See Python Cookbook.
-Singleton implemention: All instances change step together
-"""
-
-class Borg:
-    shared_state = {}
-    def __init__(self):
-        self.__dict__ = self.shared_state
-
diff --git a/Tools/RunTimeTester/src/CMTConfigurer.py b/Tools/RunTimeTester/src/CMTConfigurer.py
deleted file mode 100755
index ef2b54fab6d67fb3dcdc30cdabe11c7e3093e7ca..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CMTConfigurer.py
+++ /dev/null
@@ -1,149 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools   import rmdirMkdir
-from RTTpath import RTTpath
-import os, string
-
-# -------------------------------------------------------------------------
-
-from MethodTracer        import MetaTracer
-
-# __metaclass__ = MetaTracer
-    
-"""
-CMTConfigurer. Configures the CMT directories. Makes them first, if necessary
-"""
-
-def massageCmd(cmd):
-    # remove excess white space, make separator = ';' 
-    tokens = cmd.split()
-    cmd = string.join(tokens)
-    tokens = cmd.split('; ')
-    cmd = string.join(tokens,';')
-    return cmd
-
-def pathUpToReleaseArea(branch, isNightly, distArea, hardCoded, logger):    
-    return os.path.join(str(distArea), hardCoded.nightlyDirIfAny(branch,isNightly))
-
-def getCMTversionFromReleaseFile(release, branch, isNightly, topProject, distArea, hardCoded, logger):
-    pathToFile   = RTTpath(pathUpToReleaseArea(branch,isNightly, distArea, hardCoded, logger),
-                           topProject,
-                           release,
-                           topProject+'Release/cmt/setup.sh')
-    searchString = '/afs/cern.ch/sw/contrib/CMT/'
-    version = 'v1r18p20060606' # default
-    logger.debug('Looking for CMT version, setup.sh file: %s' % str(pathToFile))
-
-    if pathToFile.exists():
-        try:
-            logger.debug('Looking for CMT version, setup.sh file %s found' % str(pathToFile))
-            
-            ifile = open(str(pathToFile),'r')
-            lines = ifile.readlines()
-            ifile.close
-            
-            for l in lines:
-                if l.find(searchString)!=-1:                        
-                        version = l[l.rfind('/')+1:l.rfind(';')]
-                        logger.debug('Using release CMT version: %s' % version)
-        except Exception, e:
-            logger.warning('Problem grepping setup.sh file for CMT version')
-            logger.warning('Using default CMT version: %s' % version)
-    else:
-        logger.warning('%s does not exist.' % str(pathToFile))
-        logger.warning('Using default CMT version: %s' % version)
-
-    return version
-
-class CMTConfigurerBase:
-
-    def __init__(self, argDict, logger):
-        logger.debug('%s.__init__ starts'% self.__class__.__name__)
-
-        # must haves...
-        self.release    = argDict.pop('release')
-        self.branch     = argDict.pop('branch')
-        self.topProject = argDict.pop('topProject')
-        self.distArea   = argDict.pop('distArea')
-        self.hardCoded  = argDict.pop('hardCoded')
-
-        self.__dict__.update(argDict)
-        self.logger                 = logger
-        self.cmtVersion             = self.getCMTversion()
-        # paramsForCMT.cmtVersion     = self.cmtVersion
-        
-    def getCMTversion(self):
-        isNightly  = True
-        if self.release[0] in string.digits: isNightly = False
-        return getCMTversionFromReleaseFile(self.release,
-                                            self.branch,
-                                            isNightly,
-                                            self.topProject,
-                                            self.distArea,
-                                            self.hardCoded,
-                                            self.logger)
-
-    def makeHomeCMTDir(self): pass # override for sub classes
-
-
-class CMTConfigurerBuild(CMTConfigurerBase):
-    def __init__(self, argDict, logger):
-
-        cmtProjDir      = argDict.pop('cmtProjDir')
-        self.cmtHomeDir = argDict.pop('cmtHomeDir')
-        CMTConfigurerBase.__init__(self, argDict, logger)
-
-        ### NOTE: ATLAS_DIST_AREA is not the dist area as we know it i.e. the base in
-        ### which you find releases, rather it is the path to the AtlasLogin package.
-        ### It is a constant, hence hardcoded.
-        self.hReq  = 'set CMTSITE CERN\n'
-        self.hReq += 'set SITEROOT /afs/cern.ch\n'
-        self.hReq += 'macro ATLAS_DIST_AREA $(SITEROOT)/atlas/software/dist\n'
-        self.hReq += 'macro ATLAS_TEST_AREA %s\n' %  cmtProjDir
-        self.hReq += 'use AtlasLogin AtlasLogin-* $(ATLAS_DIST_AREA)\n'
-
-    def makeHomeCMTDir(self):
-        dir = str(self.cmtHomeDir)
-        rmdirMkdir(dir)
-        ofile = open(os.path.join(dir,'requirements'), 'w')
-        ofile.write(self.hReq)
-        ofile.close()        
-    
-    def configure(self):
-        msg = 'Configure CMT for release %s\n' % self.release
-        self.logger.debug(msg)
-
-        r  = """cd %s;
-        source /afs/cern.ch/sw/contrib/CMT/%s/mgr/setup.sh>& /dev/null;
-        cmt config >& /dev/null;
-        """ % (self.cmtHomeDir,
-               self.cmtVersion)
-
-        r = massageCmd(r)
-        self.logger.debug('configure command %s' % r)
-        return r
-    
-# ===============================================================
-
-class CMTConfigurerKit(CMTConfigurerBase):
-    def __init__(self, argDict, logger):
-        CMTConfigurerBase.__init__(self, argDict, logger)
-
-    def configure(self):
-        msg = 'Setting up CMT for kit, release %s\n' % self.release
-        self.logger.debug(msg)
-        setup  = str(self.distArea)
-        setup += '/cmtsite/setup.sh '
-        return setup
-
-# ===============================================================
-
-class CMTConfigurerLocal(CMTConfigurerBase):
-    'User does all, nothing to be done.'    
-    def __init__(self, argDict, logger):
-        self.userHomeCmtDir = argDict['userHomeCmtDir']
-        CMTConfigurerBase.__init__(self,  argDict, logger)       
-        
-    def configure(self): pass
-        
-    def getCMTversion(self): return self.userHomeCmtDir
diff --git a/Tools/RunTimeTester/src/CMTProjectDirCalculator.py b/Tools/RunTimeTester/src/CMTProjectDirCalculator.py
deleted file mode 100644
index ae86075978c45fb87edf13ef827c91fadd25de13..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CMTProjectDirCalculator.py
+++ /dev/null
@@ -1,101 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from CMTtools     import decodeCMTCONFIG
-from exc2string2  import exc2string2
-from ShellCommand import ShellCommand
-import os.path
-class CMTProjectDirCalculator:
-    def __init__(self,
-                 topProject,
-                 originalBranch,
-                 release,
-                 otherProject,
-                 isNightly,
-                 cmtConfig,
-                 cmtHomeDir,
-                 logger
-                 ):
-        
-        self.topProject = topProject
-        self.originalBranch   = originalBranch
-        self.release          = release
-        self.otherProject     = otherProject
-        self.isNightly        = isNightly
-        self.cmtConfig        = cmtConfig
-        self.cmtHomeDir       = cmtHomeDir
-        self.logger           = logger
-         
-        self.projectCMTDirs = {}
-
-        self.findProjectPaths()
-        
-    def findProjectPaths(self):
-        self.projectCMTDirs = {self.topProject:{'Release':'', 'RunTime':''}}
-
-        for thing in ['Release','RunTime']:
-            if self.isNightly:
-                topProjectPath = os.path.join('/afs/cern.ch/atlas/software/builds/nightlies/',
-                                              self.originalBranch,
-                                              self.topProject,
-                                              self.release,
-                                              self.topProject+thing,
-                                              'cmt')
-            else:
-                topProjectPath = os.path.join('/afs/cern.ch/atlas/software/builds/',
-                                              self.topProject,
-                                              self.release,
-                                              self.topProject+thing,
-                                              'cmt')
-                
-            self.projectCMTDirs[self.topProject][thing] = topProjectPath
-                    
-
-        cmtInfo = decodeCMTCONFIG(self.cmtConfig)
-                
-        cmd = ['cd %s' % self.cmtHomeDir,
-               'source setup.sh -tag=%s,%s,%s,%s,%s,%s,%s' % (self.topProject,
-                                                              self.otherProject,
-                                                              cmtInfo['build'],
-                                                              self.originalBranch,
-                                                              cmtInfo['compiler'],
-                                                              cmtInfo['nbits'],
-                                                              self.release),
-               'cd %s;source setup.sh;cmt show projects' % self.projectCMTDirs[self.topProject]['Release']
-               ]
-
-        try:
-            sc = ShellCommand(cmd, self.logger, 100, None, True)
-        except Exception, e:
-            m = 'Unable to obtain project dirs. cmd:\n %s\nException:%s\nTraceback:%s' % (cmd, str(e),exc2string2())
-            self.logger.critical(m)
-            return
-
-        self.logger.debug('printing directories start')
-        reply = sc.getReply()
-            
-        for l in reply:
-            
-            if l.find('No such')>-1:
-                self.logger.error('error getting project directories. cmd:\n %s\n%s\n%s' % (cmd,str(reply),l))
-                break
-            
-            l.strip()
-            tokens = l.split()
-            if len(tokens) < 4: continue
-            project = tokens[0]
-            directory = tokens[3][:-1]
-
-            self.logger.debug(project)
-            if not (project.startswith('Atlas') or project.startswith('DetCommon')): continue
-
-            self.projectCMTDirs[project] = {'Release':'', 'RunTime':''}
-            for thing in self.projectCMTDirs[project].keys():
-                dir = os.path.join(directory, project+thing, 'cmt')
-                if os.path.exists(dir): self.projectCMTDirs[project][thing] =  dir
-
-        self.logger.debug('printing directories stop')
-        self.logger.debug('Project dirs:')
-        for proj, dict in self.projectCMTDirs.items():
-            self.logger.debug('  %-20s %s' % (proj, dict['Release']))
-            self.logger.debug('  %-20s %s' % (proj, dict['RunTime']))
-
diff --git a/Tools/RunTimeTester/src/CMTtools.py b/Tools/RunTimeTester/src/CMTtools.py
deleted file mode 100755
index 12903ecfdaaec21442c28eafccf197c319919fb9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CMTtools.py
+++ /dev/null
@@ -1,10 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-def decodeCMTCONFIG(cfg):
-    tokens = cfg.split('-')
-    nbits = 32
-    if cfg.startswith('x86_64'): nbits = 64
-    assert (len(tokens) == 4)
-    dict = {'chip':tokens[0], 'os':tokens[1], 'compiler':tokens[2], 'build': tokens[3], 'nbits': nbits}
-    return dict
diff --git a/Tools/RunTimeTester/src/CMTtools.py.obsolete b/Tools/RunTimeTester/src/CMTtools.py.obsolete
deleted file mode 100755
index 498abebc7b1010a68509ff6dc77319051df4a8db..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CMTtools.py.obsolete
+++ /dev/null
@@ -1,86 +0,0 @@
-#! /usr/local/bin/python
-
-from requirementsWriter import requirementsWriter
-from requirementsWriter import requirementsWriter_Project
-
-import os
-from releaseToNum      import releaseToNum
-from Tools             import setProtections
-import shellCommand 
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-smallLogger = logging.getLogger('rtt.standalone')
-# -------------------------------------------------------------------------
-import Reporter
-globalReporter = Reporter.getGlobalReporter()
-# ------------------------------------------------------------------------
-def writeTopCMTRequirementsFile(paths, packagesToUse):
-    'write and setup a top level cmt file.'
-    
-    
-    os.chdir(paths.workDirs['workBase/'])
-    m  = 'writeTopCMTRequirementsFile cwd = '+str(os.getcwd())
-    logger.debug(m)
-    open("requirements","w").write(requirementsWriter(
-        paths.releasePath,\
-        paths.runType,
-        paths.workDirs['workBase/'],
-        packagesToUse
-        ))
-
-    nightly,relNum = releaseToNum(paths.release)
-
-    cmds = []
-    if nightly or relNum >=870:
-        cmd = 'source '
-        cmd += '/afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh'
-        cmds.append(cmd)
-        #cmd += '/afs/cern.ch/sw/contrib/CMT/v1r16p20040901/mgr/setup.sh;'
-    else:
-        cmd = 'source '
-        cmd += '/afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh'
-        cmds.append(cmd)
-        
-    cmds.append('cmt config')
-    # os.system(cmd)
-
-    shellCommand.shellCommand(cmds)
-
-    setProtections(os.getcwd(),0775)
-
-# ------------------------------------------------------------------------
-
-def writeTopCMTRequirementsFile_Project(paths, packagesToUse):
-    'write and setup a top level cmt file.'
-    
-    
-    os.chdir(paths.workDirs['workBase/'])
-    m  = 'writeTopCMTRequirementsFile_Project cwd = '+str(os.getcwd())
-    logger.debug(m)
-    ofile = open("requirements","w")
-    ofile.write(requirementsWriter_Project(
-        paths.workDirs['workBase/'],
-        packagesToUse)
-                )
-
-    nightly,relNum = releaseToNum(paths.release)
-
-    cmds = []
-    if nightly or relNum >=870:
-        cmd = 'source '
-        # cmd += '/afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh'
-        cmd += '/afs/cern.ch/sw/contrib/CMT/v1r18p20051108/mgr/setup.sh'
-        cmds.append(cmd)
-        #cmd += '/afs/cern.ch/sw/contrib/CMT/v1r16p20040901/mgr/setup.sh;'
-    else:
-        cmd = 'source '
-        cmd += '/afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh'
-        cmds.append(cmd)
-        
-    cmds.append('cmt config')
-    # os.system(cmd)
-
-    shellCommand.shellCommand(cmds)
-
-    setProtections(os.getcwd(),0775)
diff --git a/Tools/RunTimeTester/src/CVSPackage.py b/Tools/RunTimeTester/src/CVSPackage.py
deleted file mode 100755
index 88d6d39a7b89f9457167b55043caf2fbae33a550..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CVSPackage.py
+++ /dev/null
@@ -1,119 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection import formatCollection
-from RTTpath          import RTTpath
-from exc2string       import exc2string
-from os.path          import join, exists
-import sys, xml.dom.minidom
-from PackageXMLFile   import PackageXMLFile
-import time
-
-
-"""
- Examples:
- ---------
- package.name = JiveXML
- package.pathToXMLfile = /some/base/path/JiveXML_TestConfiguration.xml
- package.fullName = graphics/JiveXML
- 
- Example of dictionary of parameters rececovered from cmt
- {'cmtPath':
-   '/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_0',
-
- 'name': 'AthExHelloWorld',
- 'local': False,
- 'errorMessage': '',
- 'version': 'AthExHelloWorld-00-01-15',
-  'pathToXMLfile' : '/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/
-      rel_0/Control/AthenaExamples/AthExHelloWorld/test/
-      AthExHelloWorld_TestConfiguration.xml',
- 'offset': 'Control/AthenaExamples',
- 'root':
-      '/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/
-       rel_0/Control/AthenaExamples/AthExHelloWorld',
-"""
-
-class CVSPackage:
-    def __init__(self, packageName, pathToTestConfig, packageOffset, packageVersion,
-                 packageCmtPath, packageRoot, packageSharePath, errorMsg, local=False):
-        
-        self.name              = packageName
-        self.pathToTestConfig  = pathToTestConfig
-        self.errorMessage      = errorMsg
-        self.offset            = packageOffset
-        self.version           = packageVersion
-        self.cmtPath           = packageCmtPath
-        self.root              = packageRoot
-        self.packageSharePath  = packageSharePath
-        self.local             = local
-
-        # packages are made for all Test conf files, but not all test
-        # conf files ask for the RTT.
-        self.wantsRTT         = True
-
-        # Some packages are needed even though its jobs are not run.
-        self.runJobs = True
-
-        self.packageXMLfile = PackageXMLFile(self.pathToTestConfig)
-        
-        if not self.packageXMLfile.exists():
-            self.wantsRTT = False
-            return
-
-        # Do not set wantsRTT to False, if XML file is invalid
-        if self.packageXMLfile.isValid:
-            rtt = self.packageXMLfile['rtt']
-            rttjobs = self.packageXMLfile['rtt/jobList']
-            if not rtt or not rttjobs:
-                self.wantsRTT = False
-        else:
-            self.errorMessage = self.packageXMLfile.errMsg
-            
-    # ---------------------------------------------------
-            
-    def getPathToXMLfile(self):
-        return self.pathToTestConfig
-
-    # ---------------------------------------------------
-    
-    def getFullName(self):
-        return self.root
-
-    # ---------------------------------------------------
-
-    def getOffset(self):
-        return self.offset
-
-    # ---------------------------------------------------
-
-    def getVersion(self):
-        return self.version
-
-    # ---------------------------------------------------
-    
-    def isLocal(self):
-        return self.local
-
-    # for project builds, CMTpath is the directory in which
-    # the cmt directory is found
-    def getPackageCmtPath(self):
-        return str(RTTpath(self.cmtPath))
-
-    def setErrorMessage(self,msg):
-        self.errorMessage += msg+'\n'
-
-    def noError(self):
-        return len(self.errorMessage) == 0
-        
-    def setNoRun(self): self.runJobs = False
-
-    def runMyJobs(self): return self.runJobs
-
-    # def getRttCmtHomePath(self): return self.rttCmtHomePath
-    # def getRttCmtProjPath(self): return self.rttCmtProjPath
-
-    def __str__(self):
-        return 'CVSPackage:\n'+formatCollection(self.__dict__)+'\n'
-
-    def __del__(self):
-        if self.packageXMLfile: del self.packageXMLfile
diff --git a/Tools/RunTimeTester/src/ChainFileCopier.py b/Tools/RunTimeTester/src/ChainFileCopier.py
deleted file mode 100644
index 15a811b98bc75fb2967dd4247283d19e199d239a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ChainFileCopier.py
+++ /dev/null
@@ -1,78 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#from Tracer import Tracer
-#class ChainFileCopier(Tracer):
-import os
-import shutil
-class ChainFileCopier:
-    def __init__(self, chainFiles, runPath, chainFileArea, logger):
-        self.logger      = logger
-        self.copyTable   = [(os.path.join(runPath,f), os.path.join(chainFileArea,f)) for f in chainFiles]
-        self.copiedFiles = {} 
-
-    def copyFiles(self):
-        """
-        Checks if file is valid (for now, valid = exists). If so, copy to
-        chain store area.
-        """
-
-        msg  = '\n%s::copyFiles: files to copy begin\n:' % self.__class__.__name__
-        for src, dest in self.copyTable:
-            msg += '   %s %s\n' % (src, dest)
-        msg  += '%s::copyFiles: files to copy end' % self.__class__.__name__
-        self.logger.info(msg)
-            
-
-        # src is the original chain file out name (no directory)
-        # dest is where the file is now. 
-
-        for src, dest in self.copyTable:
-            
-            # this test could be performed a sepcialised
-            # FileVerifier object for different test criteria
-            
-            if not os.path.exists(src):
-                self.logger.warning('Chain file out not present, skipping: %s %s' % (src, dest)) 
-                continue
-
-            # src might refer to a file in a subdirectory
-            dirName = os.path.dirname(dest)
-            if dirName:
-                try:
-                    os.makedirs(dirName)
-                except:
-                    if not os.path.exists(dirName):
-                        self.logger.error('Cannot make subdirectory in Chain store: %s' % dirName)
-                        continue
-            try:
-                shutil.copyfile(src, dest)
-            except:
-                self.logger.error('Error copying chain file %s to %s' % (src, dest))
-            else:
-                self.logger.info('Copied chain file file %s to %s' % (src, dest))
-                self.copiedFiles[src] = dest
-
-            msg = 'Copied chain files\n:'
-            for item in self.copiedFiles.items():
-                msg += '   %s %s' % item
-            self.logger.info(msg)
-
-    def fileLocation(self, f):
-        "Return new location if file has been copied"
-        
-        if f in self.copiedFiles:
-            return self.copiedFiles[f]
-        return None
-
-        
-    def __str__(self):
-
-        s = '\n'
-        s+='ChainFileCopier  start\n'
-        s+='   Chain Files to copy:\n '
-        for f in self.copyTable: s+='      %s %s\n' % f
-        s+='   \n Copied chain files:\n '
-        for f in self.copiedFiles.items(): s+='      %s %s\n' % f
-        s+='ChainJobMinder  end\n'
-        
-        return s
diff --git a/Tools/RunTimeTester/src/ChainJobDescriptor.py b/Tools/RunTimeTester/src/ChainJobDescriptor.py
deleted file mode 100644
index d05d0c9cc04ec324aae719da5c04510b4dbf170c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ChainJobDescriptor.py
+++ /dev/null
@@ -1,21 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Descriptor import Descriptor
-class ChainJobDescriptor(Descriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-        Descriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-        self.descriptor = None
-        self.outChainFiles = [] # filenames (posibly with subdirectories)
-        self.name = 'ChainJob'
-        self.jobGroup = 'ContainerGroup' # jobGroup is used by MinderFactory
-
-        
-    def setOutFiles(self, fileList):
-        self.outChainFiles = fileList
-
-    def addDescriptor(self, desc):
-        self.descriptor = desc
-        #self.jobGroup   = None
-        #self.name = '%s_%s_' % (self.name, self.descriptor.identifiedName)
-        #self.fixName()
-        #self.log = ''
diff --git a/Tools/RunTimeTester/src/ChainMinderStateEngine.py b/Tools/RunTimeTester/src/ChainMinderStateEngine.py
deleted file mode 100644
index f5b6f0eda0265dca0c4e76e4de67b072f60e98b0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ChainMinderStateEngine.py
+++ /dev/null
@@ -1,18 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from MinderStateEngineBase import MinderStateEngineBase
-
-class ChainMinderStateEngine(MinderStateEngineBase):
-    def __init__(self, minder, state):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-
-
-        
-        self.dispatchTable = {
-            'queued'           : minder.fromQueued,
-            'running'          : minder.fromRunning,
-            'finishing'        : minder.fromFinishing,
-            'operatorCloseDown': minder.fromOperatorCloseDown,
-            'done'             : minder.fromDone
-            }
-
diff --git a/Tools/RunTimeTester/src/ChainStoreMaker.py b/Tools/RunTimeTester/src/ChainStoreMaker.py
deleted file mode 100644
index 5213a3d0524bff90486353decb373ebdedc35118..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ChainStoreMaker.py
+++ /dev/null
@@ -1,22 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from createDirectoryDictionary import createDirectoryDictionary
-
-import os
-
-class ChainStoreMaker:
-    def __init__(self, argBag, enableMake):
-        
-        the_path         = createDirectoryDictionary(argBag, withoutReleaseName=True)['packages']
-        self.csPath      = os.path.join(argBag.base, the_path, argBag.package)
-    
-        self.enableMake  = enableMake
-
-    def chainStorePath(self, chainName):
-        path = os.path.join(self.csPath, chainName)
-        if self.enableMake:
-            if not os.path.exists(path):
-                os.makedirs(path)
-        return path
-        
-
diff --git a/Tools/RunTimeTester/src/Checkers.py b/Tools/RunTimeTester/src/Checkers.py
deleted file mode 100755
index b389915aa908fb97e06eda33c4640b26c923d0a0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Checkers.py
+++ /dev/null
@@ -1,175 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Checkers that run on  logfiles returned after processing.
-
-simpleLogChecker - A Log checker that examines log files for the success message
-                   printed by the  athena Applicationmanager.
-		   Gets the logname from a JobInfo object.
-
-PS
-"""
-from popen2      import Popen4 
-import os,shutil, string, logging
-from JobGroupKits import *
-from os.path      import join
-from Defs         import *
-from Factory_SiteSpecifics   import SiteSpecificsFactory
-
-# -----------------------------------------------------------------------------
-logger = logging.getLogger('rtt')
-# -----------------------------------------------------------------------------
-
-success =  0
-failure = -1
-
-class Checker:
-	def __init__(self):
-		self.logger = logger
-
-	def setLogger(self, newLogger): self.logger=newLogger
-
-	def __call__(self, minder):
-		return self.check(minder)
-	
-	def check(self):
-		pass
-
-#--------------------------------------------------------------------------------
-
-class SimpleLogChecker(Checker):
-	def __init__(self,searchStringList,vetoStringList=[], mustFindAllSearch=True):
-
-		Checker.__init__(self)
-		self.mustFindList      = searchStringList
-		self.mustNotFindList   = vetoStringList
-		self.mustFindAllSearch = mustFindAllSearch
-		self.nSuccesses        = 0
-		self.chainLength       = 0
-		
-	def equivalent(self, other):
-		if len(self.__dict__.keys()) != len(other.__dict__.keys()):
-			return False
-		for k in self.__dict__.keys():
-			if k not in other.__dict__.keys(): return False
-
-		for e in self.mustFindList:
-			if e not in other.mustFindList: return False
-			
-		for e in self.mustNotFindList:
-			if e not in other.mustNotFindList: return False
-
-		return True
-			
-	def check(self,jMinder):
-
-		# The "raw" logname is jMinder.jDescriptor.log
-		# It is copied to the results directory
-		# with name jMinder.keepFiles[jMinder.jDescriptor.log]
-		logName = join(jMinder.runPath, jMinder.log) #
-		
-		self.logger.debug("%s: looking for: %s" %(self.__class__.__name__, logName))
-		
-		if not os.path.exists(logName):
-			m = "%s: No logfile found: %s" % (self.__class__.__name__, logName)
-			self.logger.warning(m)
-			return RTTdefs.error
-
-		self.logger.debug("%s: found logfile: %s" % (self.__class__.__name__, logName))
-                    
-		logf=open(logName, "r")
-		logContents=logf.read()
-		logf.close()
-
-		self.logger.debug("%s: read in logfile %s length %d" % (self.__class__.__name__, logName, len(logContents)))
-
-                if len(logContents) ==0:
-			return RTTdefs.error
-
-		splices = self.spliceLogFile(logContents)
-
-		self.logger.debug("%s: have sliced  logfile %s into %d " % (self.__class__.__name__, logName, len(splices)))
-
-		self.chainLength = len(splices)
-
-		tests = [(self.checkSearchString(s), self.checkVetoString(s)) for s in splices]
-		self.logger.debug('results: %s', str(tests))
-		self.nSuccesses = tests.count( (True, True) )
-		self.logger.debug('length, successes: %d %d' % (self.chainLength, self.nSuccesses))
-
-		if self.nSuccesses == self.chainLength:
-			return RTTdefs.success
-		else:
-			return RTTdefs.error
-		
-
-	def spliceLogFile(self, text):
-		# other log checkers elaborate this method
-		return [text]
-	
-	def checkSearchString(self, text):
-		"determine if all the substrings in list self.searchStrings"
-		found = []
-		for s in self.mustFindList:
-			if s not in text:
-				if self.mustFindAllSearch:
-					self.logger.debug('%s: required string %s not found' % (self.__class__.__name__, s) )
-					return False
-			else:
-				found.append(s)
-		if not found:
-			self.logger.debug('Did not match on any search strings!')
-			return False
-		return True
-
-	
-	def checkVetoString(self, text):
-
-		for s in self.mustNotFindList:
-			if s in text:
-				self.logger.debug('%s: veto string %s found' % (self.__class__.__name__, s) )
-				return False
-		return True
-	
-#--------------------------------------------------------------------------------
-
-class ChainedJobsLogChecker(SimpleLogChecker):
-
-	def __init__(self, searchStringList, vetoStringList=[]):
-		SimpleLogChecker.__init__(self, searchStringList, vetoStringList)
-		
-	def spliceLogFile(self, text):
-
-		splitText = 'Starting job'
-
-		njobs = text.count(splitText)
-		
-		if not njobs:
-			return [text]
-		
-		splices = text.split(splitText)
-
-		self.logger.debug('%s: spliced into %d return %d' % (self.__class__.__name__, len(splices),
-								     len(splices[1:]) ) )
-		return splices[1:]
-#--------------------------------------------------------------------------------
-
-class SimpleFileDiffer(Checker):
-	def __init__(self):
-		Checker.__init__(self)
-		
-	def check(self,jMinder):
-		m = "SimpleFileDiffer is now obsolete. Please use the DCube framework instead."
-		self.logger.error(m)
-		return RTTdefs.error
-
-#-------------------------------------------------------------------------------
-
-class HistogramDiffer(Checker):
-	def __init__(self,comparisonType):
-		Checker.__init__(self)
-	def check(self, jMinder):
-		m = "HistogramDiffer is now obsolete. Please use the DCube framework instead."
-		self.logger.error(m)
-		return RTTdefs.error
-	
diff --git a/Tools/RunTimeTester/src/CmtLines.py b/Tools/RunTimeTester/src/CmtLines.py
deleted file mode 100755
index df1b5743d2a5e0a2e7b604294f3a0161b398bbb1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/CmtLines.py
+++ /dev/null
@@ -1,145 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, os.path
-from RTTpath        import RTTpath
-from CMTtools       import decodeCMTCONFIG
-
-from RTTSException import RTTCodingError
-
-class CmtLinesBase:
-    def __init__(self, logger, argBag, useRunTime=False):
-
-        cmtConfig = argBag.pop('cmtConfig')
-        self.__dict__.update(argBag.__dict__)
-
-        self.useRunTime = useRunTime # True=AtlasXXXRunTime dir, False=AtlasXXXRelease dir
-
-        cmtInfo       = decodeCMTCONFIG(cmtConfig)
-        self.build    = cmtInfo['build']
-        self.nbits    = cmtInfo['nbits']
-        self.compiler = cmtInfo['compiler']
-        
-    def cmtLinesScript(self):
-        """
-        Returns lines of script needed to set up cmt if this is being
-        done from scratch(eg for nightly running)
-        Returns an empty string otherwise
-        (for running from checked-out code, where it is
-        assumed the user has performed the cmt operations.
-        """
-
-        # is the package in the local package list declared in the top level
-        # RTT config file? Uncaught exception thrown if there is a problem 
-
-        self.checkConditions()
-        #lines = self.lines()
-        #lines.append('source setup.sh') # needed to setup athena.py
-        #return lines
-        return self.lines()
-
-    def checkConditions(self):
-        """
-        check we have made the project build cmt requirements files
-        for now I am going to hardwire a lot of the code. This is less than
-        ideal as we are assuming what is done in other parts of the code
-        (UserStuffRetriever)
-        Need to make an USR in order to refind the cmt requirements
-        files...
-        """
-        hExists = os.path.exists(self.cmtHomeDir)
-        msg     = ''
-        
-        if not hExists:
-            msg += 'cmtLinesProject no file:  self.cmtHomeDir  = %s\n' % self.cmtHomeDir
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-                
-# -------------------------------------------------------------------------
-
-class CmtLinesBuild(CmtLinesBase):
-    def __init__(self, logger, argBag, useRunTime=False):
-        CmtLinesBase.__init__(self, logger, argBag, useRunTime)
-
-    def lines(self):
-        topProjectPath = str(self.projectCMTDirs[self.topProject][{True:'RunTime',False:'Release'}[self.useRunTime]])
-        lines = [
-            'unset PYTHONPATH',
-            'cd %s' % self.cmtHomeDir,
-            'source setup.sh -tag=%s' % self.tags()
-            ]
-        
-        return lines
-
-    def hardCodedFlags(self):
-        flags = {('14.1.0.Y', 'AtlasHLT'): 'noTest,noPath,setup,bugfix',
-                 ('14.0.10.Y', 'AtlasHLT'): 'noTest,noPath,setup,bugfix'
-                 }
-
-        return flags.get((self.originalBranch, self.otherProject), 'noTest,setup')
-
-    def tags(self):
-        tags = '%s,%s,%s,%s,%s,%s,%s,%s'  % (self.topProject,
-                                             self.otherProject,
-                                             self.release,
-                                             self.branch,
-                                             self.build,
-                                             self.nbits,
-                                             self.compiler,
-                                             self.hardCodedFlags())
-        return tags
-
-# -------------------------------------------------------------------------
-
-class CmtLinesBuildHLT(CmtLinesBuild):
-    def __init__(self, logger, argBag, useRunTime=False):
-        CmtLinesBuild.__init__(self, logger, argBag, useRunTime)
-
-    def hardCodedFlags(self):
-        flags = {'14.0.10.Y': 'noTest,noPath,setup,bugfix'}
-        return flags.get(self.originalBranch, 'noTest,setup')
-
-    def tags(self):
-        tags = '%s,%s,%s,%s,%s,%s,%s,%s' % (
-            self.topProject,
-            self.release,
-            self.otherProject,
-            self.branch,
-            self.build,
-            self.nbits,
-            self.compiler,
-            self.hardCodedFlags())
-        
-        return tags
-        
-# -------------------------------------------------------------------------
-        
-class CmtLinesLocal(CmtLinesBase):
-    def __init__(self, logger, argBag, useRunTime=False):
-        CmtLinesBase.__init__(self, logger, argBag, useRunTime)
-        self.testProjectCmtDir = argBag.testProjectCmtDir # test project dir
-    def checkConditions(self):
-        """ do nothing: the existence of the cmt dirs is checked in
-        the argBag.__init__"""
-        pass
-
-    def lines(self):
-        lines = [
-            'unset PYTHONPATH',
-            'cd %s' % self.userHomeCmtDir,
-            'source setup.sh -tag=%s' % self.tags(),
-            'cd %s' % str(self.testProjectCmtDir),
-            # 'cmt broadcast'
-            ]
-
-        return lines
-
-    def tags(self):
-        tags = '%s,%s,%s,%s,%s,%s,%s,oneTest,setup' % (self.topProject,
-                                                       self.otherProject,
-                                                       self.release,
-                                                       self.branch,
-                                                       self.build,
-                                                       self.nbits,
-                                                       self.compiler)
-        return tags
-
diff --git a/Tools/RunTimeTester/src/Coloriser.py.obsolete b/Tools/RunTimeTester/src/Coloriser.py.obsolete
deleted file mode 100755
index 81f876909884456f0ae483d480e516258627c86f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Coloriser.py.obsolete
+++ /dev/null
@@ -1,58 +0,0 @@
-"""
-Coloriser returns string of ccolored htm table elements
-coloriserMN M- number of colors
-            N - number of html table entries
-            
-Coloriser21 2:two colors (green/red. User gives green string, all others red
-            1:always returns 1 table entry
-            
-Coloriser32 3:three colors (green/red/brown. User gives green string, and red strings
-              all others brown
-            2: always return with 2 entries (padd if neccessary)
-"""
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class Coloriser32:
-    def __init__(self, greenString, redString):
-        self.greenString = greenString
-        self.redString   = redString
-
-    def colorise(self, string):
-        strings = string.split()
-        if len(strings)==0:
-            strings.append("&nbsp")
-        if len(strings)==1:
-            strings.append("n/a")
-
-        totString = ""
-        for s in strings[0:2]:
-            if s == self.greenString:
-                totString = totString+"<td align=\"center\"><font color=\"green\">"+s+"</font></td>"
-            elif s == self.redString:
-                totString = totString+"<td align=\"center\"><font color=\"red\">"+s+"</font></td>"
-            else:
-                totString = totString+"<td align=\"center\"><font color=\"brown\">"+s+"</font></td>"
-
-        return totString
-        
-class Coloriser21:
-    def __init__(self, greenString):
-        self.greenString = greenString
-
-    def colorise(self, string):
-        strings = string.split()
-        if len(strings)==0: strings.append("&nbsp")
-
-        
-        totString = ""
-        for s in strings[0:1]:
-            if s == self.greenString:
-                totString = totString+"<font color=\"green\">"+s+"</font>"
-            else:
-                totString = totString+"<font color=\"red\">"+s+"</font>"
-
-        return totString
-        
diff --git a/Tools/RunTimeTester/src/Commander.py b/Tools/RunTimeTester/src/Commander.py
deleted file mode 100755
index 95f6404939de2fda46f59ef35a27e268d9902867..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Commander.py
+++ /dev/null
@@ -1,171 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from os.path import join
-import CMTtools
-
-# -------------------------------------------------------------------------
-
-class Commander:    
-
-    """
-    Commander - a base class that issues commands to the operating system.
-    Derived classe overide the commands to allow for differences due to
-    different mode values.
-    
-    Commands:  athena
-"""
-    def __init__(self):
-        pass
-    
-    def workNodes(self):
-        return ['lxb7513',
-                'lxb7516',
-                'lxb7517',
-                'lxbrb2201',
-                'lxbrb2202',
-                'lxbrb2203',
-                'lxbrb2204',
-                'lxbrb2205',
-                'lxbrb2206',
-                'lxbrb2207',
-                'lxbrb2208',
-                'lxbrb2209',
-                'lxbrb2210',
-                'lxbrb2211',
-                'lxbrb2212',
-                'lxbrb2213',
-                'lxbrb2214',
-                'lxb7511',
-                'lxb7512',
-                'lxb7514',
-                'lxb7515',
-                'lxb7518',
-                'lxb7519',
-                'lxb7534',
-                'lxb7535',
-                'lxb7536',
-                'lxb7537',
-                'lxb7538',
-                'lxb7539',
-                'lxb8761',
-                'lxb8762',
-                'lxb8763',
-                'lxb8764',
-                'lxb8765',
-                'lxb8766',
-                'lxb8768',
-                'lxb8769',
-                'lxb8770',
-                'lxb8773',
-                'lxb8774',
-                'lxb8775',
-                'lxb8848',
-                'lxb9019',
-                'lxbra2707',
-                'lxbra2708',
-                'lxbrg3809',
-                'lxbrg4004',
-                'lxbrg4005',
-                'lxbrg4010',
-                'lxbrg4201',
-                'lxbrg4204',
-                'lxbrg4205',
-                'lxbrg4206',
-                'lxbrg4209',
-                'lxbrg4210',
-                'lxbrg4401',
-                'lxbrg4406',
-                'lxbrg4407',
-                'lxb7520',
-                'lxb7526',
-                'lxb7527',
-                'lxb7533',
-                'lxb7566',
-                'lxb8767',
-                'lxb8771',
-                'lxb8772',
-                'lxbra2709',
-                'lxbrg4001',
-                'lxbrg4404',
-                'lxbrg4405']
-
-
-class LinuxInteractiveCommander(Commander):
-    """ Commands for running interactively (foreground) on Linux"""
-
-    def __init__(self):
-        Commander.__init__(self)
-    
-    def submitCommand(self, runPath, log):        
-        script  = join(runPath, "runScript.sh")
-        logName = join(runPath, log)
-
-        #!!! do not put in the background!!! joblog is present before job ends
-        return "rm %s; nice %s>& %s" % (logName, script, logName)
-    
-    def postProcessorCommand(self, runPath):
-        script   = join(runPath, 'postProcessorScript.sh')
-        logName  = join(runPath, 'postProcessor.log')
-
-        cmd =  "rm %s; source %s>& %s" % (logName, script, logName)
-
-        return cmd
-        
-    
-
-class BatchCommander(Commander):
-    def __init__(self):
-        Commander.__init__(self)
-
-    def submitCommand(self, runPath, identifiedName, log, elog, queue, walltimelimit=''):
-        log    = join(runPath, log)
-        elog   = join(runPath, elog)
-        script = join(runPath, "runScript.sh")
-
-        return self.command(script, identifiedName, log, elog, queue, walltimelimit)
-
-    def postProcessorCommand(self, runPath, identifiedName, queue):
-        log    = join(runPath, 'postProcessor.log')
-        elog   = join(runPath, 'postProcessor.elog')
-        script = join(runPath, "postProcessorScript.sh")
-
-        return self.command(script, identifiedName, log, elog, queue)
-
-class LSFBatchCommander(BatchCommander):
-    """ Commands for running jobs on the CERN LSF queues"""
-    
-    def __init__(self):
-        BatchCommander.__init__(self)
-
-    def command(self, script, identifiedName, log, elog, queue, walltimelimit=''):
-        if walltimelimit:
-            try:
-                int(walltimelimit)
-                walltimelimit='-W %s' % walltimelimit
-            except:
-                walltimelimit=''
-        nodeSelectionOption = '-m "' + reduce(lambda x,y:x+' '+y, self.workNodes()) + '"'
-        cmd = 'LSB_STDOUT_DIRECT=Y;/usr/bin/bsub %s -o %s %s -q %s -R "atlasrtt" %s -J "%s"' % (nodeSelectionOption, 
-                                                                                                log,
-                                                                                                walltimelimit,
-                                                                                                queue,
-                                                                                                script,
-                                                                                                identifiedName)
-        return cmd
-
-
-            
-class DummyCommander(Commander):
-
-    def __init__(self):
-        Commander.__init__(self)
-    
-    def command(self, identifiedName):
-        return  'print Dummy submit'+ identifiedName
-
-    def submitCommand(self, identifiedName): return self.command(identifiedName)
-
-    def postProcessorCommand(self, identifiedName):  return self.command(identifiedName)
-
-
-
diff --git a/Tools/RunTimeTester/src/ConfigFileDumper.py.obsolete b/Tools/RunTimeTester/src/ConfigFileDumper.py.obsolete
deleted file mode 100755
index 81ddb6b00aabe6f3bd5ac0441d33d8a9674861c3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ConfigFileDumper.py.obsolete
+++ /dev/null
@@ -1,98 +0,0 @@
-from Paths                     import NoPackagePaths
-from RunTimer                 import RunTimer
-from RTTConfigParser          import RTTConfiguration
-from GetUserConfigData        import GetUserConfigData
-from Evaluate                 import Evaluate
-from makeRTTDOM               import makeRTTDOM
-from ParametersForCMT         import ParametersForCMT
-from Factory_CMTConfigurer    import CMTConfigurerFactory
-from ShellCommand             import ShellCommand
-
-
-import os, logging, xml.dom.minidom
-from LoggerSetup import setupLoggers
-
-
-# logFileDir = '/local_disk/sherwood/RTT02'
-logFileDir = '/afs/cern.ch/atlas/project/RTT/data/peter/logs/JobIdentifier'
-logFileDir = setupLoggers(logFileDir)
-logger = logging.getLogger('rtt')
-
-confFile = '/afs/cern.ch/atlas/project/RTT/data/peter/cfg/example.xml'
-rttConfig = RTTConfiguration(confFile)
-RTTSrcDir = os.getcwd()
-RTTLibDir = os.path.join(RTTSrcDir,'../share')
-print rttConfig
-
-summaryDocument    = makeRTTDOM('RTTSummary')
-parametersForCMT   = ParametersForCMT(rttConfig, logger)
-
-
-
-class ConfigFileDumper:
-    def __init__(self):
-
-        factory = CMTConfigurerFactory(logger)
-        configurer = factory.create(parametersForCMT)
-        configurer.makeHomeCMTDir()
-        cmd  = configurer.configure()
-        ShellCommand([cmd], logger)
-
-        
-        paths = NoPackagePaths(RTTSrcDir,
-                               RTTLibDir,                               
-                               rttConfig,
-                               RunTimer(1, logger),
-                               logFileDir,
-                               summaryDocument.createElement,
-                               summaryDocument.createTextNode,
-                               parametersForCMT,
-                               logger)
-        
-
-        getUserConfigData  = GetUserConfigData(rttConfig, paths)
-        self.packages           = getUserConfigData.userConfigData()
-        
-        msg = 'CFD: packages: %s' % str([p.name for p in self.packages])
-        logger.debug(msg)
-
-        self.paths = paths
-
-    def dumpConfig(self):
-        
-        for p in self.packages:
-            print '\n========== %s ==============\n\n' % p.name
-            print open(p.pathToTestConfig,'r').read()
-
-
-
-def parseXML(path):
-    try:
-        return  xml.dom.minidom.parse(path)
-    except:
-        print 'parse unsuccessful for ', path
-        return None
-
-
-class PetersConfigFileDumper(ConfigFileDumper):
-    def __init__(self):
-        ConfigFileDumper.__init__(self)
-
-        fn = '%s_%s' % (rttConfig.get('cmtConfig'), rttConfig.get('release'))
-        of = open(fn, 'w')
-        for p in self.packages:
-            msg = ('package %s:  hash: %i' %(p.name,hash(open(p.pathToTestConfig,'r').read())))
-            of.write(msg+'\n')
-            logger.debug(msg)
-
-        for p in self.packages:
-            of.write(p.pathToTestConfig+'\n')
-
-        of.close()
-            
-
-if __name__ == '__main__':
-
-
-    pd = PetersConfigFileDumper()
-    pd.dumpConfig()
diff --git a/Tools/RunTimeTester/src/ConnectionTester.py.obsolete b/Tools/RunTimeTester/src/ConnectionTester.py.obsolete
deleted file mode 100755
index 51858af4e41bc62f2db3ad1ce493db740f3f3d53..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ConnectionTester.py.obsolete
+++ /dev/null
@@ -1,63 +0,0 @@
-if __name__ == '__main__':
-    from TimedDataHist  import TimeHist
-    from shellCommand   import shellCommand
-    from exc2string     import exc2string
-    import TestSuite2
-    import time, sys, os
-
-    hist = TimeHist()
-    home = os.getcwd()
-    longTest = True
-    #
-    # cmd = []
-    # cmd.append(
-    #    "source /afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh;cmt config'>&/dev/null"
-    #    )
-    # cmd.append('cmt -quiet show macro_names _TestConfiguration')
-
-    print TestSuite2.init()
-    
-    while (True):
-        start = time.time()
-        packages = []
-        try:
-            usr = TestSuite2.makeUserStuffRetriever()
-            print 'paths', usr.paths
-            print 'substring', usr.substring
-            print 'accept', usr.accept
-            print 'veto', usr.veto
-            if longTest:
-                files = usr.userStuff()
-            else:
-                usr.findMacros()
-                # usr.extractPackages()
-                # print 'UserStuffRetriever found',files
-                delta = time.time()-start
-        except Exception, e:
-            print exc2string(sys.exc_info())
-            delta = -1.
-            files = []
-           
-
-        print 'packages:'
-        for p in files: print p.__dict__,'\n\n'
-    
-        hist.newPoint(delta)
-
-    
-        fn = os.path.join(home,'ConnectionTimeData.txt')
-        # print 'plots to ',fn
-        try:
-            of = open(fn, 'w')
-            of.write(hist.valuesPlot())
-            of.write(hist.timesPlot())
-            of.close
-        except Exception, e:
-            print exc2string(sys.exc_info())
-        
-        time.sleep(30)
-        print hist.valuesPlot()
-        print hist.timesPlot()
-
-
-
diff --git a/Tools/RunTimeTester/src/ConnectionTester2.py b/Tools/RunTimeTester/src/ConnectionTester2.py
deleted file mode 100755
index be315ec924024a6de8fe03d017fa8122526ffe5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ConnectionTester2.py
+++ /dev/null
@@ -1,65 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-if __name__ == '__main__':
-    from TimedDataHist  import TimeHist
-    from shellCommand   import shellCommand
-    from exc2string     import exc2string
-    import TestSuite2
-    import time, sys, os
-
-    hist = TimeHist()
-    home = os.getcwd()
-    longTest = True
-    #
-    # cmd = []
-    # cmd.append(
-    #    "source /afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh;cmt config'>&/dev/null"
-    #    )
-    # cmd.append('cmt -quiet show macro_names _TestConfiguration')
-
-    print TestSuite2.init()
-    
-    while (True):
-        start = time.time()
-        packages = []
-        try:
-            usr = TestSuite2.makeUserStuffRetriever()
-            print 'paths', usr.paths
-            print 'substring', usr.substring
-            print 'accept', usr.accept
-            print 'veto', usr.veto
-            if longTest:
-                files = usr.userStuff()
-            else:
-                usr.findMacros()
-                # usr.extractPackages()
-                # print 'UserStuffRetriever found',files
-                delta = time.time()-start
-        except Exception, e:
-            print exc2string(sys.exc_info())
-            delta = -1.
-            files = []
-           
-
-        print 'packages:'
-        for p in files: print p.__dict__,'\n\n'
-    
-        hist.newPoint(delta)
-
-    
-        fn = os.path.join(home,'ConnectionTimeData.txt')
-        # print 'plots to ',fn
-        try:
-            of = open(fn, 'w')
-            of.write(hist.valuesPlot())
-            of.write(hist.timesPlot())
-            of.close
-        except Exception, e:
-            print exc2string(sys.exc_info())
-        
-        time.sleep(30)
-        print hist.valuesPlot()
-        print hist.timesPlot()
-
-
-
diff --git a/Tools/RunTimeTester/src/ContainerDescriptor.py b/Tools/RunTimeTester/src/ContainerDescriptor.py
deleted file mode 100644
index f5683b3ae49d312c7c18c7891bb2202ad778c853..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ContainerDescriptor.py
+++ /dev/null
@@ -1,16 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Descriptor import Descriptor
-class ContainerDescriptor(Descriptor):
-      def __init__(self, paths, jobSerialNumber, logger):
-            Descriptor.__init__(self, paths, jobSerialNumber, logger, {})
-            self.descs = []
-            self.jobGroup = 'ContainerGroup' # jobGroup is used by MinderFactory
-            self.name = 'Container'
-            self.identifiedName = self.name+str(self.jobSerialNumber)
-            
-      def addDescriptor(self, desc): self.descs.append(desc)
-            
-      def descriptors(self): return self.descs
- 
-      def deferredMinderUpdate(self): pass
diff --git a/Tools/RunTimeTester/src/ContainerMinder.py b/Tools/RunTimeTester/src/ContainerMinder.py
deleted file mode 100644
index ef8c3886a892aeb882d9b63ec72a02950c911708..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ContainerMinder.py
+++ /dev/null
@@ -1,351 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from MinderBase                 import MinderBase
-from simpleDataForMonitoring    import simpleDataForMonitoring
-from RTTSException              import RTTCodingError
-from exc2string2                import exc2string2
-from Tools                      import nowSecs
-
-import os.path
-# note there are more imports at the end of the module
-# placed there to avoid circular import problems.
-
-"""
-Recursive minder creation: if te cointainerDescriptor contains container
-descriptors, a containerMinder will be constructed tat n turn will build its
-contained minders. These need to be connected to the outer minders by providing
-setSummaryCallBack and poll metods.
-"""
-#from Tracer import Tracer
-#class ContainerMinder(MinderBase, Tracer):
-class ContainerMinder(MinderBase):
-    def __init__(self, argBag, containerDescriptor):
-
-
-        MinderBase.__init__(self, argBag.logDir,
-                            argBag.package,
-                            containerDescriptor.identifiedName,
-                            containerDescriptor.jobSerialNumber,
-                            argBag.elementCreator,
-                            argBag.textNodeCreator
-                            )
-
-
-        self.summaryCallBack = None
-        self.finishedTests   = [] # added purely while I (PS: Who is I) was testing the querier code, needs to be thought out
-        try:                      # maybe there arn't any such files
-            self.outChainFiles   = argBag.outChainFiles
-        except:
-            self.outChainFiles   = []
-        self.jobGroup        = containerDescriptor.jobGroup
-        self.name            = containerDescriptor.name
-        self.errorMessages   = []
-        self.resPath         = containerDescriptor.resPath
-
-        self.minders = [argBag.minderFactory.create(argBag, d) for d in containerDescriptor.descriptors()]
-        self.noOfMinders = len(self.minders)
-
-    def cullMinders(self):
-        doneMinders = [m for m in self.minders if m.done]
-        for m in doneMinders: self.minders.remove(m)
-
-
-    def dump(self): self.logger.debug(self.__str__())
-
-    def isSuccess(self): return True
-    
-
-    def setSummaryCallBack(self, newCallBack):
-        'Chain the xml converters'
-        
-        self.xmlConverter.setSummaryCallBack(newCallBack)
-        [m.setSummaryCallBack(self.xmlConverter) for m in self.minders]
-        #[m.setSummaryCallBack(newCallBack) for m in self.minders]
-        self.logger.debug('Set summary call back in self.minders %s'% self.xmlConverter.__class__.__name__)
-
-    def dataForMonitoring(self):
-        return simpleDataForMonitoring(self)
-
-    def fromQueued(self): return 'running'
-    
-    def fromRunning(self):
-        self.fromRunningActions()
-        if self.fromRunningTest():
-            self.logger.debug('ContainerMinder::fromRunning setting done for %s:%s' % (self.__class__.__name__,
-                                                                                       self.identifiedName))
-            return 'done'
-        return 'running'
-    
-    def fromRunningActions(self):
-        self.logger.debug('ContainerMinder::fromRunningActions: no of active Minders %d' % self.noOfActiveMinders())
-        self.cullMinders()
-        
-    def fromRunningTest(self):
-        self.logger.debug('ContainerMinderStateEngine::fromRunningTest: no of active Minders %d' % self.noOfActiveMinders())
-        return self.noOfActiveMinders() == 0
-        
-   
-        
-class SequentialMinder(ContainerMinder):
-    def __init__(self, argBag, containerDescriptor):
-        ContainerMinder.__init__(self, argBag, containerDescriptor)
-        self.currentMinder = None
-        self.setCurrentMinder()
-        
-    def setCurrentMinder(self):
-        self.currentMinder = None
-        if self.minders: 
-            self.currentMinder = self.minders.pop(0)
-
-    def cullMinders(self):
-        if not self.currentMinder: return
-            
-        # current minder is set and running - leave as is
-        if not self.currentMinder.isDone(): return
-
-        # current minder is set and done. Set new current and return 
-        if self.minders:
-            self.currentMinder = self.minders.pop(0)
-        else:
-            self.currentMinder = None
-        return
-                           
-    def poll(self):
-        if self.currentMinder:
-            try:
-                self.currentMinder.poll()
-            except Exception, e:
-                msg = 'Minder: %s: Contained minder %s threw an exception during poll\nTraceBack\n%s' % (
-                    self.identifiedName,
-                    self.currentMinder.identifiedName,
-                    exc2string2())
-                
-                self.logger.error(msg)
-                raise RTTCodingError(msg)
-                
-        ContainerMinder.poll(self)
-        # m = '%s End %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        
-    def noOfActiveMinders(self):
-        if not self.currentMinder: return len(self.minders)
-        return len(self.minders)+1
-
-
-    def setSummaryCallBack(self, newCallBack):
-        'Chain the xml converters'
-        if self.currentMinder:
-            ContainerMinder.setSummaryCallBack(self, newCallBack)
-            self.currentMinder.setSummaryCallBack(self.xmlConverter)
-            self.logger.debug('Set summary call back in self.currentMinder %s'% self.xmlConverter.__class__.__name__)
-
-
-
-    def __str__(self):
-        s = '\n'
-        s+='SequentialMinder start\n'
-        
-        s+= '   Minder name: %s\n' % self.identifiedName
-        s+= '   %d minders in list\n' % len(self.minders)
-        s+= '   Dump of Current Minder: \n %s\n' % (str(self.currentMinder))
-
-        s+='SequentialMinder end\n'
-        
-        return s
-    
-        
-class ParallelMinder(ContainerMinder):
-    def __init__(self, argBag, containerDescriptor):
-        ContainerMinder.__init__(self, argBag, containerDescriptor)
-
-    def poll(self):
-        # m = '%s Start %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        # print m
-
-        try:
-            [m.poll() for m in self.minders]
-        except Exception, e:
-            msg = 'Minder %s: a contained minder threw an exception during poll\nTraceBack"\n%s' % (self.identifiedName,
-                                                                                                    exc2string2())
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-        
-        ContainerMinder.poll(self)
-        # m = '%s End %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        # print m
-
-    def noOfActiveMinders(self):
-        return len(self.minders)
-
-    def cullMinders(self):
-        self.minders = [m for m in self.minders if not m.isDone()]
-
-    def __str__(self):
-        s='\n'
-        s+='ParallelMinder start\n'
-
-        
-        s+= '   Minder name: %s\n' % self.identifiedName
-        s+= '   Minder list:\n %s\n' % str(self.minders)
-
-        s+='ParallelMinder end\n'
-
-        return s
-     
-#class ChainJobMinder(Tracer):
-from ChainFileCopier import ChainFileCopier
-
-class ChainJobMinder(MinderBase):
-    def __init__(self, argBag, chainJobDescriptor):
-        MinderBase.__init__(self,
-                            argBag.logDir,
-                            argBag.package,
-                            chainJobDescriptor.identifiedName,
-                            chainJobDescriptor.jobSerialNumber,
-                            argBag.elementCreator,
-                            argBag.textNodeCreator
-                            )
-        paths = chainJobDescriptor.descriptor.paths
-        # minder instantiation deferred to first poll - bacause before it is unknown of the chain
-        # files are present in the store.
-        self.minder          = None # minder instantiation deferred to first poll - bacause before 
-        self.fileCopier      = None # instantiated when minder is instantiated
-        self.innerDescriptor = chainJobDescriptor.descriptor
-        self.argsForMinder   = argBag
-        self.name            = chainJobDescriptor.name
-        self.jobGroup        = chainJobDescriptor.jobGroup
-        self.errorMessages   = []
-        self.resPath         = chainJobDescriptor.resPath
-        self.chainStoreArea  = chainJobDescriptor.chainStoreArea
-        self.noOfMinders     = 0
-        # setup links to be placed in the chain store area.
-        try:                      # maybe there arn't any
-            self.outChainFiles   = argBag.outChainFiles
-        except:
-            self.outChainFiles   = []
-
-    def makeMinder(self):
-        factory         = self.argsForMinder.minderFactory
-        # route the logged messages to the files for this minder (descriptor gets the TestRun logger)
-        self.innerDescriptor.logger = self.logger
-        # decide which input files to use depending on what is available in the chain store
-        self.innerDescriptor.deferredMinderUpdate()
-        
-        self.minder     = factory.create(self.argsForMinder, self.innerDescriptor)
-        # grab the inner minders cleanSpace method, so it an be called by this outer Minder
-        # after the out minder has finished with the work directory. Done in this manner,
-        # the cleanSpace method still accesses the inner minders attributes.
-        # self.cleanSpace    = self.minder.cleanSpace
-        self.fileCopier = ChainFileCopier( self.outChainFiles,
-                                           self.minder.runPath,
-                                           self.chainStoreArea,
-                                           self.logger)
-        def makeFn(obj):
-            def fn():
-                obj.logger.info('%s dummy fn: no action is beinbg taken', obj.name)
-
-            return fn
-
-        # set the inner minder cleanup function to a do nothing function so it will
-        # not delete files in work directory
-        # self.minder.cleanSpace = makeFn(self.minder)
-
-        # copy the list of chainFiles to the inner minder to prevent those chainFiles
-        # which are also keep files to be "kept"
-        self.minder.neverCopyAndNeverDelete.extend([os.path.join(self.minder.runPath, ocf) for ocf in self.outChainFiles])
-        
-        self.noOfMinders = 1
-        self.logger.debug('%s:%s created inner minder %s:%s' % (self.__class__.__name__,
-                                                                self.identifiedName,
-                                                                self.minder.__class__.__name__,
-                                                                self.minder.identifiedName))
-    def poll(self):
-        # m = '%s Start %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        # print m
-
-        # minders instantiated on first poll
-        if not self.minder:
-            try:
-                self.makeMinder()
-            except:
-                msg = '%s error making inner minder Traceback:\n%s' % (self.__class__.__name__,  exc2string2())
-                raise RTTCodingError(msg)
-
-            self.minder.setSummaryCallBack(self.xmlConverter)
-                
-            
-        try:
-            self.minder.poll()
-        except Exception, e:
-            msg = '%s %s: inner minder  %s threw an exception during poll\nTraceBack"\n%s' % (self.__class__.__name__,
-                                                                                              self.identifiedName,
-                                                                                              self.minder.identifiedName,
-                                                                                              exc2string2())
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-
-        MinderBase.poll(self)
-
-    def noOfActiveMinders(self):
-        if self.minder:
-            if self.minder.isDone():
-                return 0
-            return 1
-        
-        return 0
-
-    def cleanSpace(self):
-        for oc in self.outChainFiles:
-            try:
-                os.remove(os.path.join(self.minder.runPath, oc))
-            except:
-                self.logger.error('Unable to delete chain file (%s) from run directory.' % oc)
-                self.logger.error(exc2string2())
-            else:
-                self.logger.debug('Successfully deleted chain file (%s) from run directory.' % oc)
-                
-    def setSummaryCallBack(self, newCallBack):
-        'Chain the xml converters'
-        
-        self.xmlConverter.setSummaryCallBack(newCallBack)
-        ## self.minder.setSummaryCallBack(newCallBack)
-        self.logger.debug('Set summary call back in self.xmlConverter %s'% str(newCallBack.__class__.__name__))
-        #self.logger.debug('Set summary call back in self.minder %s'% str(newCallBack))
-        
-    def __str__(self):
-        s = '\n'
-        s+='ChainJobMinder  start\n'
-        s+='   Contained Minder: %s:%s\n '%(self.minder.__class__.__name__,
-                                            self.minder.identifiedName)
-        s+='   Chain Files:\n '
-        s+='       '+str(self.fileCopier)
-        s+='ChainJobMinder  end\n'
-        return s
-
-    def dataForMonitoring(self):
-        return simpleDataForMonitoring(self)
-
-    def fromQueued(self): return 'running'
-    
-    def fromRunning(self):
-        self.fromRunningActions()
-        if self.fromRunningTest():
-            self.logger.debug('ContainerMinder::fromRunning setting finishing for %s:%s' % (self.__class__.__name__,
-                                                                                            self.identifiedName))
-            return 'finishing'
-        return 'running'
-
-    def fromFinishing(self):
-        self.cleanSpace()
-        return 'done'
-    
-    def fromRunningActions(self):
-        self.logger.debug('ContainerMinder::fromRunningActions: no of active Minders %d' % self.noOfActiveMinders())
-        #self.cullMinders()
-        if (self.minder.isSuccess() and self.minder.isDone()): self.fileCopier.copyFiles()
-
-    def fromRunningTest(self):
-        self.logger.debug('ContainerMinderStateEngine::fromRunningTest: no of active Minders %d' % self.noOfActiveMinders())
-        return self.noOfActiveMinders() == 0
-
-        
-   
diff --git a/Tools/RunTimeTester/src/ContainerMinderStateEngine.py b/Tools/RunTimeTester/src/ContainerMinderStateEngine.py
deleted file mode 100644
index 1f4bbb006fcff89621c57e73ad7fd9aca9897f5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ContainerMinderStateEngine.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from MinderStateEngineBase import MinderStateEngineBase
-
-class ContainerMinderStateEngine(MinderStateEngineBase):
-    def __init__(self, minder, state):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-
-
-        
-        self.dispatchTable = {
-            'queued'           : minder.fromQueued,
-            'running'          : minder.fromRunning,
-            'operatorCloseDown': minder.fromOperatorCloseDown,
-            'done'             : minder.fromDone
-            }
-
diff --git a/Tools/RunTimeTester/src/ContainerXMLConverter.py b/Tools/RunTimeTester/src/ContainerXMLConverter.py
deleted file mode 100644
index 3d9a1aa52f61598c217f09549f9e4c4a10006e39..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ContainerXMLConverter.py
+++ /dev/null
@@ -1,124 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an xml file summary of an RTT run
-"""
-import sys, os, fnmatch, os.path
-from StringIO      import StringIO
-from makeRTTDOM    import makeRTTDOM
-# from xml.dom.ext   import PrettyPrint
-from Defs          import RTTdefs
-from exc2string2   import exc2string2
-from MethodTimer   import MethodTimer
-from sortValues    import sortValues
-
-class ContainerXMLConverter(MethodTimer):
-    
-    def __init__(self, minder):
-
-        MethodTimer.__init__(self)
-        
-        self.minder          = minder
-        self.logger          = minder.logger
-        self.elementCreator  = minder.elementCreator
-        self.textNodeCreator = minder.textNodeCreator
-        self.callBack        = None
-        self.minderXMLNode   = None
-        self.containedMinderNodeDict = {}
-        
-    def unlink(self):
-        del self.minder
-        if self.callBack: del self.callBack
-        
-    def update(self, dict, node):
-
-        # if node:
-        #    print '---------ContainerXMLNode converter for %s received node:' % self.minder.identifiedName
-        #    temp = StringIO() 
-        #    PrettyPrint(node, temp)
-        #    print temp.getvalue()
-        # else:
-        #    print '---------ContainerXMLNode converter for  received None as node'
-        
-        if node:
-            oldNode = self.containedMinderNodeDict.pop(dict['name'], None)
-            # if oldNode: oldNode.unlink() # not sure if this will work - is this node in the deocument?
-
-            self.containedMinderNodeDict[dict['name']] = node
-
-    def makeSummaryReport(self):
-        if not self.callBack:
-            self.logger.error('No object for Minder to call back to, abandoning update')
-            return
-                                      
-        minder = self.minder # for nested methods
-        # nested scopes are used to pass parameters to these nested functions
-        def findErrorMessages():
-
-            element = self.elementCreator('errors')
-            
-            text = reduce(lambda x, y: x+y+'\n',
-                          minder.errorMessages,
-                          '')
-            if text:
-                tn = self.textNodeCreator(text)
-                element.appendChild(tn)
-
-            return element
-
-        def makeTextNode((tagName, text)):
-            tn = self.textNodeCreator(str(text).strip())
-            el = self.elementCreator(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(document))
-        
-        mElement = self.elementCreator('%s' % self.minder.__class__.__name__)
-        textInfo = {
-            'identifiedName':    self.minder.identifiedName,
-            'jobID':             self.minder.jobSerialNumber,
-            'state':             self.minder.stateEngine.state.state,
-            'status':            self.minder.status(),
-            'jobGroup':          self.minder.jobGroup,
-            'stateHistory':      str(self.minder.stateEngine.state.history),
-            'nContainedMinders': str(self.minder.noOfMinders),
-            'nActiveMinders':    str(self.minder.noOfActiveMinders())
-            }
-
-
-        
-        for key,val in textInfo.items():
-            if val.__class__.__name__ in ['list', 'tuple']: # 1-to-many tag/value mapping
-                [mElement.appendChild(makeTextNode((key,v))) for v in val]
-            else: #1-to-1 tag/value mapping 
-                mElement.appendChild(makeTextNode((key,val)))
-
-        # keep fixed ordering of nodes. Needed by RTT self tests
-        sortedNodes = sortValues(self.containedMinderNodeDict)
-        #[mElement.appendChild(n) for n in self.containedMinderNodeDict.values()]
-        [mElement.appendChild(n) for n in sortedNodes]
-        self.outputToFile(mElement)
-
-        m = 'Created, minder node for %s' % self.minder.identifiedName
-        dict = self.minder.dataForMonitoring()
-        self.callBack.update(dict, mElement)
-        self.logger.debug(m)
-        
-
-    def outputToFile(self, minderNode):
-        space = ' '
-        f = os.path.join(self.minder.resPath, 'rttjobinfo.xml')
-        h = None
-        try:
-            h = open(f, 'w')
-            h.write(minderNode.toprettyxml(3*space, '\n'))
-            h.close()
-        except:
-            self.logger.error('Unable to output minder info XML file. Traceback: \n%s' % exc2string2())
-        if h: h.close()    
-
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('received a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/DBFileRetriever.py.obsolete b/Tools/RunTimeTester/src/DBFileRetriever.py.obsolete
deleted file mode 100755
index dce03016c50fa200d1afeb6e621d965e38c17c42..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/DBFileRetriever.py.obsolete
+++ /dev/null
@@ -1,80 +0,0 @@
-##Retrieves reference file from RTT database
-
-import anydbm, logging,os
-from dataFiles2String import dataFiles2String
-from os.path import join
-
-logger=logging.getLogger('rtt')
-
-class DBFileRetriever:
-    def __init__(self, jDescriptor):
-        self.jDescriptor = jDescriptor
-
-        
-    def writeRefFile(self, file):
-        "copy file from the data base to the run directory"
-
-        dbHandle    = anydbm.open(self.jDescriptor.paths.refDB, 'c')
-        
-        
-        key = [unicode(self.jDescriptor.refRelease),
-                   unicode(self.jDescriptor.paths.build),
-                   unicode(self.jDescriptor.paths.branch),
-                   unicode(self.jDescriptor.paths.runType),
-                   unicode(self.jDescriptor.name),
-                   dataFiles2String(self.jDescriptor),
-                   unicode(file)]
-            
-
-            
-        refFileName = join(self.jDescriptor.runPath,'refFile_'+file)
-        
-            
-        try:
-            refFile = dbHandle[key.__repr__()]
-            logger.debug("Found reference file in DB")
-            logger.debug("DataBase at: "+self.jDescriptor.paths.refDB)    
-            logger.debug("key = "+key.__repr__())
-            open(refFileName, 'w').write(refFile)
-            
-        except KeyError:
-            logger.warning("NO reference file found in DB")
-            logger.warning("DataBase at: "+self.jDescriptor.paths.refDB)    
-            logger.warning("key = "+key.__repr__())
-            
-        dbHandle.close()
-        
-        return refFileName
-
-
-    def storeFileInDB(self, file):
-        
-        key = [unicode(self.jDescriptor.paths.release),
-                   unicode(self.jDescriptor.paths.build),
-                   unicode(self.jDescriptor.paths.branch),
-                   unicode(self.jDescriptor.paths.runType),
-                   unicode(self.jDescriptor.name),
-                   dataFiles2String(self.jDescriptor),
-                   unicode(file)]
-        try:
-            dbHandle = anydbm.open(self.jDescriptor.paths.refDB, 'c')
-            tFile    = join(self.jDescriptor.runPath,file)
-        
-            rFile    = open(tFile,'r+b').read()
-            rFile    = str(rFile)
-            keyString= key.__repr__()
-            
-            logger.debug("storing file :"+str(file))
-            logger.debug("Key :"+key.__repr__())
-            
-            dbHandle[keyString]=rFile
-            
-            #dbHandle[key.__repr__()]=file
-            #dbHandle.close()
-        except:
-            logger.error('Error in storing file in DB: Key ='+key.__repr__())
-        
-        dbHandle.close()
-
-
-    
diff --git a/Tools/RunTimeTester/src/DBIntegrityChecker.py.obsolete b/Tools/RunTimeTester/src/DBIntegrityChecker.py.obsolete
deleted file mode 100755
index dbd176e92a947d83e64e6c642c314daefb067675..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/DBIntegrityChecker.py.obsolete
+++ /dev/null
@@ -1,203 +0,0 @@
-"""
-A class to hold checks for the data base (was corrupted - and fixed - 4/05 -
-so we know we need to check it
-"""
-
-import anydbm, os
-
-from formatCollection import formatCollection
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class DBTestBase:
-    def __init__(self, dbFileName):
-        self.dbFileName = dbFileName
-        self.passed     = False
-        self.msg        = ''
-    
-class ReadAllKeysTest(DBTestBase):
-    def __init__(self, dbFileName):
-        DBTestBase.__init__(self, dbFileName)
-        self.testName = 'readAllKeysTest'
-        
-    def test(self):
-        "Check all entries can be read"
-
-        tempDBfn = os.path.join(os.path.dirname(self.dbFileName), 'temp.db')
-        if os.path.exists(tempDBfn): os.remove(tempDBfn)
-        try:
-            tempDB   = anydbm.open(tempDBfn, 'n')
-        except:
-            self.msg = 'failed to open',tempDBfn
-            
-        testDB   = anydbm.open(self.dbFileName, 'r')
-
-        try:
-            for key in testDB: tempDB[key] = testDB[key]
-        except:
-            self.msg = 'failed to read',tempDBfn
-
-
-        self.passed = True
-        tempDB.close()
-        testDB.close()
-        os.remove('temp.db')
-
-class SingleKeyRWTest(DBTestBase):
-    def __init__(self, dbFileName):
-        DBTestBase.__init__(self, dbFileName)        
-        self.testName = 'singleKeyRWTest'
-        
-    def test(self):
-        """
-        check we can write and read to the database
-        check to values to ensure we get at least one new one.
-        """
-        
-        testDB   = anydbm.open(self.dbFileName, 'w')
-        testKey  = 'TestKeyTestKeyTestKey'
-        testVals = ['', 'TestValTestValTestVal']
-
-        try:
-            for tv in testVals:
-                testDB[testKey] = tv
-                rv = testDB[testKey]
-                if not rv == tv:
-                    self.msg = 'wrote %s but read %s' % (tv, rv)            
-                    return
-        except:
-            return
-
-        del testDB[testKey]
-        
-        self.passed = True
-        testDB.close()    
-
-
-class BrinicksTest(DBTestBase):
-    def __init__(self, dbFileName):
-        DBTestBase.__init__(self, dbFileName)        
-        self.testName = 'BrinicksTest'
-        
-    def test(self):
-        """
-        check we can write and read to the database
-        check to values to ensure we get at least one new one.
-        """
-        
-        handle   = anydbm.open(self.dbFileName, 'w')
-
-        badKeys,keysWithBadValues = self.corruptDBKeys(handle)
-        if badKeys:
-            m = 'The monitoring DB was found to contain %d corrupt keys. These will be removed.' % len(badKeys)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in badKeys]
-        elif keysWithBadValues:
-            m = 'The monitoring DB was found to contain %d keys with corrupt values. These will be removed if not recoverable.' % len(keysWithBadValues)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in keysWithBadValues]
-        else:
-            m = 'Moni DB OK: no corrupt keys to remove.'
-            self.passed = True
-            logger.info(m)
-            
-        handle.close()
-        # -----------------------------------------------------------------
-
-    def recoverOrDeleteDBEntry(self,handle,key,badValue = True):
-        """For now, we just plain delete the corrupt entry.
-        An improvement could be to grab the file version, create a dom and
-        stick in in the DB.badValue = True if the value if corrupt,
-        False if it is the key itself that is bad.
-        """
-        try:
-            del handle[key]
-            logger.warning('Removed monitoring DB key %s' % str(key))
-        except Exception, e:
-            logger.error('Unable to remove corrupt DB key %s' % str(key))
-            logger.error(exc2string2())
-                        
-    # ------------------------------------------------------------------------
-
-    def corruptDBKeys(self,handle):
-        keys    = handle.keys()
-        keysThatAreCorrupt    = []
-        keysWithCorruptValues = []
-        
-        for key in keys:
-            isPrintableKey  = len(
-                [c for c in key if c not in string.printable])==0
-            if not isPrintableKey:
-                keysThatAreCorrupt.append(key)
-            else:
-                try:
-                    parseString(handle[key])
-                except:
-                    keysWithCorruptValues.append(key)
-                    
-        return (keysThatAreCorrupt,keysWithCorruptValues)
-
-    # ------------------------------------------------------------------------
-
-
-
-class DBIntegrityChecker:
-    def __init__(self, dbFileName):
-
-        if not os.path.exists(dbFileName):
-            logger.info('Database does not exist %s ' % dbFileName)
-            self.tests = []
-            return
-        
-        self.tests = [
-            ReadAllKeysTest(dbFileName),
-            SingleKeyRWTest(dbFileName)
-            # BrinicksTest(dbFileName)
-            ]
-
-    def check(self):
-        
-        def runTest(tester):
-            try:
-                #m = 'running test %s' % tester.__class__.__name__
-                #print m
-                #logger.debug(m)
-                t.test()
-            except:
-                pass
-
-        if not self.tests:
-            m = """No DBIntegrityChecker tests, probably becqause
-            moniDb did not exist at instantiation"""
-            logger.warning(m)
-        [runTest(t) for t in self.tests]
-
-        #m = 'DBIC finished running tests'
-        #print m
-        #logger.debug(m)
-
-        def react(t):
-            if not t.passed:  
-                msg  = 'Database not intact, failed test: %s\n' % t.testName
-                msg += 'error: %s' % t.msg
-                logger.error(msg)
-
-        [react(t) for t in self.tests]
-            
-        if False not in [t.passed for t in self.tests]:
-            msg = 'DBintegrity with %d subtests  passed' % len(self.tests)
-            logger.debug(msg)
-
-if __name__ == '__main__':
-
-    db = '/unix/atlas/RTT/refFile.db.corrupt'
-    assert( not DBIntegrityChecker(db).OK)
-
-    db = '/unix/atlas/RTT/refFile.db'
-    assert( DBIntegrityChecker(db).OK)
-
-    print 'DBIntegrityChecker: all OK'
-        
diff --git a/Tools/RunTimeTester/src/DDDescriptor.py b/Tools/RunTimeTester/src/DDDescriptor.py
deleted file mode 100644
index d874c7d93b084615574cd3f92b6c475b9d0c546b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/DDDescriptor.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-class DDDescriptor: pass
diff --git a/Tools/RunTimeTester/src/DataSetCatalog.py b/Tools/RunTimeTester/src/DataSetCatalog.py
deleted file mode 100755
index 0b3830c0ded987ac279183a6f7404cb1531799d5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/DataSetCatalog.py
+++ /dev/null
@@ -1,91 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection import formatCollection
-import os
-import xml.dom.minidom
-
-# -------------------------------------------------------------------------
-
-class DataSetCatalog:
-    """A data structure which handles the logical and phyical
-    data set name look up"""
-    
-    def __init__(self, pathToCatalog, logger):
-        self.logger = logger
-        self.catalog = {}
-        if os.path.exists(pathToCatalog):
-            self.logger.debug('opening RTT data catalog')
-            self.parseCatalogFile(pathToCatalog)
-            # for line in open(paths.dataSetCatalog):
-            #    tokens = line.split()
-            #    self.catalog[tokens[0].strip()] =  tokens[1].strip()
-            #self.logger.error(formatCollection(self.catalog))
-        else:
-            msg='Could not find RTT data catalog,'
-            msg='logical data set names will not be resolved'
-            self.logger.warning(msg)
-
-    def inverse_get(self,physicalDsn):
-        for logical,physical in self.catalog.items():
-            if physical.strip()==physicalDsn.strip():
-                return logical
-        return None
-    
-    def has_key(self, key):
-        return self.catalog.has_key(key)
-        
-    def get(self, logicalDsn):
-        return self.catalog.get(logicalDsn, None)
-
-    def parseCatalogFile(self, catalog):
-        dom = xml.dom.minidom.parse(catalog)
-        self.getEntries(dom)
-
-    def getEntries(self, dom):
-        dsns    = dom.getElementsByTagName('dsn')
-        for dsn in dsns: self.handleDSN(dsn)
-
-    def handleDSN(self, dsn):
-        physicals = dsn.getElementsByTagName('physical')
-        logicals  = dsn.getElementsByTagName('logical')
-        if len(physicals) !=1 or len(logicals) != 1:
-            self.logger.error('bad tag in dataset catalog')
-            return
-        
-        key = self.getText(logicals[0].childNodes)
-        val = self.getText(physicals[0].childNodes)
-        self.catalog[key] = val
-    
- 
-    # ---------------------------------------------------------------
-    # Extract text from tag
-    # ---------------------------------------------------------------
-    
-    def getText(self,nodelist):
-        rc = ""
-        for node in nodelist:
-            if node.nodeType == node.TEXT_NODE:
-                rc = rc + node.data.strip()
-        return rc
-
-
-if __name__ == '__main__':
-
-    import os
-    catFile = '/local_disk/sherwood/RTT/rttDataSetCatalogPCUCL02.xml'
-    def applyAssert(file):
-        print file
-        if file.find('/castor/') != -1:
-            return 1
-        else:
-            assert os.path.exists(file)
-            return 0
-    class Thing:
-        def __init__(self, file):
-            self.dataSetCatalog = file
-            
-    catalog = DataSetCatalog(Thing(catFile)).catalog
-    print 'No of entries = ',len(catalog.keys())
-    isCastor = [applyAssert(item) for item in catalog.values()]
-    print reduce(lambda x, y: x+y, isCastor),' physical castor files in catalog'
-    print 'all non castor physical files present '
diff --git a/Tools/RunTimeTester/src/Defs.py b/Tools/RunTimeTester/src/Defs.py
deleted file mode 100755
index 4b62ed9153b6230a48879ebb28c676cba2e1184f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Defs.py
+++ /dev/null
@@ -1,31 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-class RTTdefs:
-    success =  0
-    error   = -1
-
-    def isSuccess(cls, code):
-        if code == 0: return True
-        return False
-    isSuccess = classmethod(isSuccess)
-    
-
-    def isError(cls, code):
-        return not RTTdefs.isSuccess(code)
-    isError = classmethod(isError)
-
-    def status(cls, code):
-        if RTTdefs.isError(code): return 'error'
-        return 'success'
-
-    status = classmethod(status)
-
-
-if __name__ == '__main__':
-
-    print RTTdefs.status(5)
-    print RTTdefs.status(0)
-    print RTTdefs.status(-1)
-    print RTTdefs.isSuccess(0)
-   
diff --git a/Tools/RunTimeTester/src/Descriptor.py b/Tools/RunTimeTester/src/Descriptor.py
deleted file mode 100755
index ff764990f0d3349c0e002ce43bb011dcfbc62737..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Descriptor.py
+++ /dev/null
@@ -1,220 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""
-Top class for the Descriptor heirarchy.
-Contains stuff needed to make run directories.
-
-"""
-import os
-import copy
-
-from Paths               import Paths
-from os.path             import join
-from RTTpath             import RTTpath
-from RTTSException       import RTTCodingError
-
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-class Descriptor:    
-    """Objects of this class hold information read from package UnifiedCOnfigurationFiles.
-    They are used to instantiate JobMinders at which point the information is
-    transferred to the minder"""
-    
-    # job serial number is now provided by the descriptor factory
-    # jobSerialNumber=0  # class variable (static in C++)
-
-    def __init__(self, paths, logger, jobSerialNumber, argDict = {}):
-        # BaseDescriptor.__init__(self)
-        
-        self.logger          = logger
-        self.paths           = paths
-        self.jobSerialNumber = jobSerialNumber
-        # self.jobSerialNumber=Descriptor.jobSerialNumber
-        # Descriptor.jobSerialNumber=Descriptor.jobSerialNumber+1
-
-        
-        # Default values for name and identified name (which has the job serial
-        # number stuck on the end of name). These should be overridden with more useful
-        # values in subclasses.
-        self.name              = "Job"
-        # self.release           = paths.release
-        self.jobDisplayName    = ''
-        
-        self.identifiedName    = self.name+str(self.jobSerialNumber)
-        self.runPath           = '' # set later in this method
-        self.resPath           = '' # set later in this method
-        self.suggestedQueue    = ''
-        self.jobGroup          = ''
-        self.confederation     = None
-        self.log               = ''
-        self.elog              = ''
-        self.suggestedBatchWallTime = ''
-
-        self.rttPilotJob       = False # a flag to signal job is a pilot job
-        self.rttATNJob         = False # job comes from ATN config file
-        self.numberedOnly      = False # when to run the job
-        self.limitedNightlies  = False # when to run the job
-
-        self.jobDocString      = ''
-        self.jobDocURL         = ''
-        
-        self.keepFilePatterns  = []   # fill from Jobs config file
-        self.auxFilePatterns   = []   # fill from Jobs config file
-        self.pScripts          = []   # fill from Jobs config file
-        self.actions           = []   # filled by JobsXMLReader: list of UserTestDescriptor instances
-        self.tests             = []   # filled by JobsXMLReader: list of UserTestDescriptor instances
-
-        # name: need to come up with a name that be used to mkdir a directory
-        # when added concatenated with a base path
-        self.identifiedName   = self.name+str(self.jobSerialNumber)
-        self.log              = self.identifiedName+"_log"
-        self.hashString       = 'None'
-        self.trendId          = None # CoolDozer id
-        self.keepFiles        = {}
-        self.missingDatasets  = []
-        
-        self.__dict__.update(argDict)
-        
-        # record of errors during coinstruction/configuration
-        self.constructionError = False
-        self.errorMessages     = []
-
-        # a list of strings used to calculate run a nd results directory names
-        self.fixName()
-
-        self.setRunPath()
-        # self.setHashString() # PS: the hash is in the dist used to update self.dict
-
-
-
-    # --------------------------------------------------------------
-    def hasData(self): return False
-    # --------------------------------------------------------------
-    
-    def setRunPath(self, pathExtensions=[]):
-        # Default setting of runpath, delegated to the PathsForDescriptors
-        # object. The paths object as well as the present descriptor
-        # is updated. Some descriptors act when
-        # they have the runpath, and so will overide this function
-
-        # --------------------------------------------------------------
-    
-        self.runPath    = RTTpath(self.paths.workDirs['packages'])
-        self.resPath    = RTTpath(self.paths.resultsDirs['packages'])
-
-        [self.runPath.postpend(item[1]) for item in pathExtensions]
-        [self.resPath.postpend(item[1]) for item in pathExtensions]
-
-        # this code has been written (too) quickly
-        # we will store the final locations of resPath and runPath in Descriptor,
-        # but not store the intermediate locations. These could have been put into the paths
-        # object but:
-        # - which one - not all clients of paths objects use a PathsForDescriptor object, so updating
-        #   this might not be very useful
-        # - to limit curly-queues: Descriptor is built from paths and then updates paths.. makes
-        #   debugging , dumping etc difficult.
-        
-        self.runPath = str(self.runPath)
-        self.resPath = str(self.resPath)
-
-    # --------------------------------------------------------------
-
-    def setName(self,name):
-        self.name = name
-        
-    # --------------------------------------------------------------
-
-    def setGroup(self, jobGroup):
-        self.jobGroup = jobGroup
-
-    # --------------------------------------------------------------
-
-    def fixName(self):
-        self.identifiedName   = self.name+str(self.jobSerialNumber)
-        self.log              = self.identifiedName+"_log"
-        # self.elog             = self.identifiedName+"_elog"
-
-    # --------------------------------------------------------------
-
-    def setupRunDir(self):
-        """
-        Allow descriptor specific actions for setting up the run
-        directory. Specialisations implemented in derived classes
-        """
-
-        pass
-    # --------------------------------------------------------------
-    
-    def __str__(self):
-        """converts to a string"""
-        s = '--------- Descriptor ---------\n'
-        s += 'serial no.:            %s\n' % self.jobSerialNumber
-        s += 'name:                  %s\n' % self.name
-        s += 'identifiedName:        %s\n' % self.identifiedName
-        s += 'log name:              %s\n' % self.log
-        s += 'job group:             %s\n' % self.jobGroup
-        s += 'runPath:               %s\n' % self.runPath
-        s += 'resPath:               %s\n' % self.resPath
-        s += 'suggested batch queue: %s\n' % self.suggestedQueue
-        s += 'confederation:         %s\n' % self.confederation
-        s += self.paths.__str__()
-        
-        return s
-
-    # --------------------------------------------------------------
-    
-    def dump(self):
-        """Dumps out jobname"""
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info('|                                           |')
-        self.logger.info('|            Descriptor  dump               |')
-        self.logger.info('|                                           |')
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info(self.__str__())
-
-    def dataForXMLNode(self): return {'hashString':      self.hashString} # derived classes extend
-
-    def deferredMinderUpdate(self): raise NotImplementedError()
-
-    def deferredDataSetUpdate(self):
-        self.datasets = {}
-
-        # chainStoreArea must be set before this method is called.
-        if not hasattr(self, 'chainStoreArea'):
-            m = 'Descriptor %s is trying to use the chain store without its path being set' % self.identifiedName
-            raise RTTCodingError(m)
-
-        # container minders do not have chainStoreFiles
-        if not hasattr(self, 'chainStoreFiles'): return
-
-        # self.chainbstore files is a list of tuples like (dataSetInstance, dataDict of fallback files)
-        #for item in self.__dict__.items(): print '%20s:    %s', item
-            
-        for datasetInstance, fallbackFileDict in self.chainStoreFiles:
-            physicalDatasetName = os.path.join(self.chainStoreArea, datasetInstance.logical)
-            self.logger.error('Descriptor.deferredDataSetUpdate => physicalDatasetName: %s' % physicalDatasetName)
-            datasetInstance.physical = physicalDatasetName
-            # if the physical file exists (ie is in chain store, use the datasetInstance,
-            # otherwise make note of the fact it is missing
-            datasetInstance.setStats() # update the mod time, size for web page displaying
-            if datasetInstance.exists_():
-                self.logger.error('Descriptor.deferredDataSetUpdate => File exists!')
-                self.datasets.setdefault(datasetInstance.type, []).append(datasetInstance)
-            else:
-                self.logger.error('Descriptor.deferredDataSetUpdate => File NOT exists!')
-                self.missingDatasets.append(datasetInstance)
-                for dst in fallbackFileDict.keys():
-                    self.datasets.setdefault(dst, []).extend(fallbackFileDict[dst])
-
-    
-        
-        m = '%s::deferredDatasetUpdate: chainStoreFiles:\n %s\n\nfinale datasets:\n %s' % (self.__class__.__name__,
-                                                                                           str(self.chainStoreFiles),
-                                                                                           str(self.datasets))
-        self.logger.debug(m)
-
-    
-    
diff --git a/Tools/RunTimeTester/src/DirectoryMaker.py.obsolete b/Tools/RunTimeTester/src/DirectoryMaker.py.obsolete
deleted file mode 100755
index c59cf1f32cb6b17287d1444dc3fa67f86ba10b6a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/DirectoryMaker.py.obsolete
+++ /dev/null
@@ -1,584 +0,0 @@
-from Tools        import mkdirIfNew, rmdirMkdir, changePerm
-import Tools2
-
-#===== RTT MODULES =====
-from requirementsWriter      import requirementsWriter
-from JobGroupKits            import JobGroupKits
-# from scriptWriterFactory     import scriptWriterFactory
-from Factory_ScriptWriter    import ScriptWriterFactory
-from PathCompleter           import PathCompleter
-from releaseToNum            import releaseToNum
-from dataFiles2String        import dataFiles2String
-from getInheritanceHeirarchy import getInheritanceHeirarchy
-from formatCollection        import formatCollection
-from exc2string              import exc2string
-#===== PYTHON MODULES ==
-import os, shutil, shelve, sys
-from os.path              import join, splitext
-
-"""
-A class to make Work, Results and Run directories.
-
-User             calls: makeDirs
-makeDirs         calls: makeBaseDirs - sets up non minder dependent parts of
-                                work and results directory tree
-
-User             calls: makeJobGroupDirs - extracts results and working
-                                           directory names (minder dependent)
-                                           
-
-makeJobGroupDirs calls: createJobGroupDir - for each of [resDir, workDir]
-                                            to actually create the directory
-
-                 calls: setupRunDir
-"""
-
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-import Reporter
-globalReporter = Reporter.getGlobalReporter()
-# -------------------------------------------------------------------------
-
-class DirectoryMaker:
-    def __init__(self, paths, logger):
-        self.logger = logger
-        self.paths = paths
-        self.makeDirs()
-
-    def setLogger(self, logger): self.logger = logger
-    
-    def makeDirs(self):
-        try:
-            # create base directories for work directory and results
-            # directory trees
-            self.makeBaseDirs() 
-
-            # to permit signal in Tester to know if directories have been made
-            self.status =  True
-        except Exception, e:
-            msg = 'Error making directories:\n%s' % str(e)
-            self.logger.error(exc2string(sys.exc_info()))
-            
-            globalReporter.couldNotMakeDirectoriesError(e)
-    
-    def makeBaseDirs(self):
-        """
-        Function for creating empty results directories.
-        
-        paths.resPath:      Directory to hold results
-        paths.workBuildRel  Last directory in the work heirarchy which
-        is not job specific.
-        """
-
-        mkdirIfNew(self.paths.resultsDirs['resBase/build/branch/runType'])
-        rmdirMkdir(self.paths.resultsDirs['resBase/build/branch/runType/release'])
-        mkdirIfNew(self.paths.workDirs['workBase/build/branch/runType'])
-        rmdirMkdir(self.paths.workDirs['workBase/build/branch/runType/release'])
-
-    def createJobGroupDir(self, jMinder, pathToDirectory):
-        """
-        create run and results directory trees. This function is sometimes
-        overridden by a null function in derived classes.
-        """
-
-        try:
-            rmdirMkdir(pathToDirectory)
-            self.logger.debug('making '+pathToDirectory)
-        except Exception, e:
-            errmsg = exc2string(sys.exc_info())
-            self.logger.error(errmsg)
-            jMinder.jDescriptor.reporter.createJobGroupDirError(errmsg)
-            jMinder.jDescriptor.reporter.writeReport()
-
-            descList.remove(desc)
-
-
-    def makeJobGroupDirs(self, jMinder):
-
-        path = jMinder.jDescriptor.runPath
-        self.createJobGroupDir(jMinder,path)
-        self.logger.debug('Making run directory %s', path)
-        self.setupRunDir(jMinder, path)
-
-        path = jMinder.jDescriptor.resPath
-        self.logger.debug('Making results directory %s', path)
-        self.createJobGroupDir(jMinder,path)
-        
-    def setupRunDir(self, jMinder, runPath):
-
-        
-        """
-        Creates run directory (obtained from the descriptor)
-        
-        Uses JobGroupKits to establish what is needed to be copied to it.
-        Uses ScriptWriter to write a run script to it.
-        """
-        self.logger.info("entering setupRunDir")
-        jDescriptor = jMinder.jDescriptor
-        self.logger.debug("setupRunDir processing: "+jDescriptor.identifiedName)
-
-        os.chdir(runPath)
-        
-        self.logger.debug("writing requirements to  "+os.getcwd())
-        paths      = jDescriptor.paths
-        distArea   = paths.hardCoded.distArea()
-        relRelease = paths.hardCoded.relativeRelease(paths.branch,
-                                                     paths.release)
-        runType    = paths.runType
-        open("requirements","w").write(
-            requirementsWriter(distArea, relRelease, runPath, runType))
-        #
-        # If the following two lines were done in separate os.system calls, the
-        # second would fail as it would not know about the env variables of the
-        # first.
-        #
-        # os.system(
-        #    "source /afs/cern.ch/sw/contrib/CMT/v1r16p20040701/mgr/setup.sh;cmt config")os.system(
-        # os.system(
-        #   "source /afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;cmt config")
-
-        nightly,relNum = releaseToNum(jDescriptor.paths.release)
-        if nightly or relNum >=870:
-            #            os.system("source /afs/cern.ch/sw/contrib/CMT/v1r16p20040901/mgr/setup.sh;cmt config")
-            os.system("source /afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh;cmt config")
-            
-        else:
-            os.system(
-                "source /afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;cmt config")
-        
-
-            self.logger.debug("   asking ScriptWriter to write run script")
-    
-        #        runScriptWriter = scriptWriterFactory(jDescriptor)
-        runScriptWriter = ScriptWriterFactory(self.logger).create(jDescriptor)
-        jDescriptor.dump()
-        
-        scr = runScriptWriter.runScript()
-    
-
-        #######################################################################
-        # handle aux files (= files needed in the run dir to run the job)
-        #######################################################################
-
-        self.logger.debug("copying auxilliary files")  
-
-        # copy the files
-        for file in jMinder.auxFiles:
-            shutil.copy(file, runPath)
-            fileName = os.path.basename(file)
-            changePerm(join(runPath,fileName),0777)
-            
-        #######################################################################
-        # handle ref files
-        #######################################################################
-
-        self.logger.debug("copying reference files")  
-        self.logger.debug(jDescriptor.paths.refDB)
-        self.retrieveRefFiles(jMinder)
-
-        #######################################################################
-        # The following block copies files from  the release to the new run directory.
-        # The files tp copy are listed in a file in the RTT share area (!?!).
-        #######################################################################
-
-        # aux = open(join(jDescriptor.paths.sharePath, "auxFiles.txt"), "r")
-        
-        # pc = PathCompleter()
-        # for line in aux.readlines():
-        #     tokens  = line.split()
-        #     package = tokens[0]
-        #     file    = tokens[1]
-        #     fn      = pc.fileFromReleasedFile(package, file, jDescriptor.paths)
-        #    if os.path.exists(fn):
-        #        shutil.copy(fn, runPath)
-            
-        #        fileName = os.path.basename(fn)
-        #        os.chmod(join(runPath,fileName),0777)
-        
-        #        self.logger.debug("copied "+fn+" to "+runPath)
-        #    else:
-        #        self.logger.debug(fn + "was not copied to runPath") #remove if statement later
-        
-        #######################################################################
-        # handle aux files desguised as string in the kit
-        #######################################################################
-
-        kit = JobGroupKits().kit(jDescriptor.jobGroup)
-
-        for  item in kit.getItem('stringsToFiles'):
-            filename = item[0]
-            string = item[1]
-            file = open(join(runPath, filename), 'w')
-            file.write(string)
-            file.close()
-
-
-        #
-        # Now perform tasks specific to the different descriptors
-        #
-
-        baseClasses = getInheritanceHeirarchy(jDescriptor)
-        if 'AthenaJobDescriptor' in baseClasses:
-            self.athenaSetUp(jDescriptor)
-
-        if 'WatcherJobDescriptor' in baseClasses:
-            self.watcherSetUp(jMinder)
-            
-        if 'WorkerJobDescriptor' in baseClasses:
-            self.workerSetUp(jMinder)
-
-        if 'ReleaseJobDescriptor' in baseClasses:
-            self.releaseSetUp(jDescriptor)
-            
-        self.logger.info("returning from setupRunDir()")
-
-
-    def releaseSetUp(self, jDescriptor):
-        of = open('srcFileInfo.txt','w')
-        of.write(str(jDescriptor.nSrcFiles)+'\n')
-        of.write(formatCollection(jDescriptor.srcFileList)+'\n')
-        of.close()
-
-    def writeOptionToRunDir(self,option, jDescriptor):
-        ext         = option[-3:]
-        if ext     != '.py':
-            option  = option+'.py'
-        dest = join(jDescriptor.runPath, option)      
-        src  = join(jDescriptor.paths.installArea,
-                    'jobOptions',
-                    jDescriptor.package,
-                    option)  
-        try:
-            shutil.copyfile(src, dest)
-        except Exception, e:
-            self.logger.error(str(e))
-
-    def athenaSetUp(self, jDescriptor):
-        # dataLink = createDataLink(runPath,jDescriptor)
-        # Create the Zebra.P link only of there is DC1 data for the job
-        if jDescriptor.hasDataSet:
-            if jDescriptor.isDC1data:
-                msg ="creating soft link to data set: "+jDescriptor.DC1dataSet
-                self.logger.debug(msg)
-                dataLink = 'ZEBRA.P'
-                if os.path.exists(dataLink): os.remove(dataLink)
-                os.symlink(jDescriptor.DC1dataSet, dataLink)
-                # elif jDescriptor.isBSdata:
-                # msg ="creating soft link to data set: "+jDescriptor.BSdataSet
-                # self.logger.debug(msg)
-                # dataLink = 'RawEvent.re'
-                # if os.path.exists(dataLink): os.remove(dataLink)
-                # os.symlink(jDescriptor.BSdataSet, dataLink)
-
-        if len(jDescriptor.furtherJobOptions)>0:
-            self.logger.debug('further options '+str(jDescriptor.furtherJobOptions))
-            self.logger.debug('Writing futher options to runDir')
-            for option in jDescriptor.furtherJobOptions:              
-                self.writeOptionToRunDir(option, jDescriptor)
-
-    def watcherSetUp(self, jMinder):
-        workerRunPaths = [minder.jDescriptor.runPath for minder in jMinder.jobsToWatch]
-        histsFile = open(join(jMinder.jDescriptor.runPath,'histLocation.txt'),'w')
-        [histsFile.write(path+'\n') for path in workerRunPaths]
-        histsFile.close()
-
-    def workerSetUp(self, jMinder):
-        
-        self.logger.debug("writing requirements to  "+os.getcwd())
-        paths      = jMinder.jDescriptor.paths
-        distArea   = paths.hardCoded.distArea()
-        relRelease = paths.hardCoded.relativeRelease(paths.branch,
-                                                     paths.release)
-        runPath    = jMinder.jDescriptor.runPath
-        runType    = paths.runType
-
-        open("requirements","w").write(requirementsWriter(distArea,
-                                                          relRelease,
-                                                          runPath,
-                                                          runType))
-        #
-        # If the following two lines were done in separate os.system calls, the
-        # second would fail as it would not know about the env variables of the
-        # first.
-        #
-        os.system(
-           "source /afs/cern.ch/sw/contrib/CMT/v1r16p20040701/mgr/setup.sh;cmt config")
-        # os.system(
-        #   "source /afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;cmt config")
-
-        release =  jMinder.jDescriptor.paths.release
-        nightly,relNum = releaseToNum(release)
-        if nightly or relNum >=870:
-            cmd  = 'source '
-            cmd += '/afs/cern.ch/sw/contrib/CMT/v1r18p20050501/mgr/setup.sh;'
-            cmd += 'cmt config'
-            os.system(cmd)
-
-                
-        else:
-            cmd  = 'source '
-            cmd += '/afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;'
-            cmd += 'cmt config'
-            os.system(cmd)
-        #
-
-        self.logger.debug("asking ScriptWriter to write run script")
-        # method has been truncated?
-            
-
-    # --------------------------------------------------------------------------
-
-    def createDataLink(self, runPath, jDescriptor):
-        """ return the appropriate run directory name for the soft link to the data,
-        depending on whether the latter is DC1 or DC2"""
-    
-        if jDescriptor.isDC1data:
-            link = "ZEBRA.P"
-            self.logger.debug('createDataLink: ' + jDescriptor.DC1dataSet + " is DC1 data")
-            #        elif jDescriptor.isBSdata:
-            #link ="RawEvent.re"
-            #self.logger.debug('createDataLink: ' + jDescriptor.DC1dataSet + " is BS data")
-        else:
-            self.logger.debug('createDataLink: ' + jDescriptor.DC1dataSet + " is DC2 data")
-            return link
-
-    # --------------------------------------------------------------------------
-    
-    def writeRefFiles(self, dbHandle, minder):
-        "copy file from the data base to the run directory"
-
-        
-        # kit = jobGroupKitFactory(jDescriptor.jobGroup)
-
-        files = []
-        files.extend(minder.asciiRefFiles)
-        files.extend(minder.binRefFiles)
-        
-        # files = jDescriptor.asciiRefFiles
-        # files.extend(kit.asciiRefFiles)
-        # files.extend(jDescriptor.binRefFiles)
-        # files.extend(kit.binRefFiles)
-
-        jDescriptor = minder.jDescriptor
-        
-        for file in files:
-
-            
-            # By default the makeDBkey method constucts the key using the same branch
-            # value as jDescriptor.paths.release.
-            #
-            # ----- Consequence 1:
-            #          If you run a nightly (branch = N.0.X or N.X.0) and want to regress
-            #          against a numbered release (branch = Numbered) you're in trouble.
-            #          When the key was stored it would have been stored with 'Numbered'
-            #          and now you're trying to access with branch = N.0.X/N.X.0. So it
-            #          must be set to 'Numbered'
-            #
-            # ----- Consequence 2:
-            #          If you're running a Numbered release and want to regress against a nightly
-            #          release, you can't. Why?  Well, what branch do you pass: N.0.X or N.X.0? 
-            
-            theBranch = jDescriptor.paths.branch
-            if Tools2.isValidReleaseNumber(jDescriptor.refRelease):
-                theBranch = 'Numbered'
-                
-            key = minder.makeDBKey(jDescriptor.refRelease, theBranch, file)
-
-            #            key = minder.makeDBKey(jDescriptor.refRelease, file)
-            refFileName = join(jDescriptor.runPath,'refFile_'+file)
-        
-            
-            try:
-                refFile = dbHandle[key.__repr__()]
-                self.logger.debug("Found reference file in DB")
-                self.logger.debug("DataBase at: "+jDescriptor.paths.refDB)    
-                self.logger.debug("key = "+key.__repr__())
-                open(refFileName, 'w').write(refFile)
-            except KeyError:
-                self.logger.warning("NO reference file found in DB")
-                self.logger.warning("DataBase at: "+jDescriptor.paths.refDB)    
-                self.logger.warning("key = "+key.__repr__())
-                reporter = jDescriptor.reporter
-                reporter.failedDBKeyFindError(jDescriptor.paths.refDB,key.__repr__())
-        
-    # --------------------------------------------------------------------------
-
-    def retrieveRefFiles(self, minder):
-        descriptor = minder.jDescriptor
-        try:            
-            dbHandle = shelve.open(descriptor.paths.refDB, 'r')
-            self.writeRefFiles(dbHandle, minder)
-            m  = 'Succesful retrieval from reference'
-            m += ' database %s' % descriptor.paths.refDB
-            self.logger.debug(m)
-
-        except Exception, e:
-            m = 'Error opening reference database %s' % descriptor.paths.refDB
-            self.logger.error(m)
-            self.logger.error("Error :"+str(e))
-            self.logger.error(exc2string(sys.exc_info()))
-        #except shelve.error:
-        #    reporter = jDescriptor.reporter
-        #    reporter.failedDBConnectError(jDescriptor.paths.refDB)
-        #    self.logger.debug("Could not open reference file dbm "+jDescriptor.paths.refDB)
-        
-# --------------------------------------------------------------------------
-"""A class to make Results directory, but not Run or Work directories.
-   Run directory will be set up however!
-   testMode : NoNewRunDirs"""
-
-class NoWorkDirCreation(DirectoryMaker):
-    def __init__(self, paths, logger):
-        DirectoryMaker.__init__(self, paths, logger)
-
-    def makeWorkDirs(self):
-        self.logger.info('Not creating Work directores')
-
-    def createRunDir(self, jMinder, runPath):
-        self.logger.info('Not creating run directory '+runPath)
-
-# --------------------------------------------------------------------------
-"""A class to make Results directory, but not Run or Work directories
-   testMode: NoNewDirs"""
-
-class DummyDirectoryMaker(DirectoryMaker):
-    def __init__(self, paths, logger):
-        DirectoryMaker.__init__(self, paths, logger)
-
-    def makeWorkDirs(self):
-        self.logger.info('Not creating Work directores')
-
-    def createRunDir(self, jMinder, runPath):
-        self.logger.info('Not creating run directory '+runPath)
-
-
-    def makeJobGroupDirs(self, jMinder):
-        self.logger.info('Not setting up the run directory')
-
-        dir = self.paths.resultsDirs['resBase/build/branch/runType/release']
-        msg = 'Not creating the results directory %s' % dir
-        self.logger.info(msg)
-
-
-# --------------------------------------------------------------------------
-"A class that only uses full setupRunDir, rest is dummy"
-
-class SpecificJobWorkDirMaker(DirectoryMaker):
-    def __init__(self, paths, logger):
-        DirectoryMaker.__init__(self, paths, logger)
-
-    def makeWorkDirs(self):
-        self.logger.info('Not creating Work directores')
-
-    def makeResultsDirs(self):
-        self.logger.info('Not creating the results directory '+self.paths.resultsDirs['resBase/build/branch/runType/release'])
-        
-    def rmAllJobResultsFiles(self,minder):
-        startsWith = minder.jDescriptor.identifiedName
-        foundOne = False
-        jobSerialNum = 1
-        for file in os.listdir(minder.jDescriptor.paths.resultsDirs['resBase/build/branch/runType/release']):
-            serial = getJobSerialNumber(file)
-            if serial > jobSerialNum:
-                jobSerialNum = serial
-                
-            if file.find(startsWith)!=-1 or file.find(startsWith[:-1]+'.py'):
-                os.remove(join(minder.jDescriptor.paths.resultsDirs['resBase/build/branch/runType/release'],file))
-                foundOne = True
-
-        # there were no such files in the results area
-        # i.e. maybe serial number wrong due to different jobs conf file
-        if not foundOne and getJobSerialNumber(startsWith) <= jobSerialNum:
-            globalReporter.differentXMLFileThanBeforeError()
-            sys.exit()
-
-class DirectoryMakerProject(DirectoryMaker):
-    def __init__(self, paths, logger):
-        DirectoryMaker.__init__(self, paths, logger)
-
-    def setupRunDir(self, jMinder, runPath):
-        
-        """
-        Creates run directory (obtained from the descriptor)
-        
-        Uses JobGroupKits to establish what is needed to be copied to it.
-        Uses ScriptWriter to write a run script to it.
-
-        This version does not write a requirements file into the run
-        directory.
-        """
-        self.logger.info("DirectroyMakerProject: entering setupRunDir")
-        jDescriptor = jMinder.jDescriptor
-        self.logger.debug("setupRunDir processing: "+jDescriptor.identifiedName)
-
-        os.chdir(runPath)
-
-        self.logger.debug("SetupRunDir: calling ScriptWriter")
-    
-        runScriptWriter = ScriptWriterFactory(self.logger).create(jDescriptor)
-        
-        scr = runScriptWriter.runScript()
-    
-
-        #######################################################################
-        # handle aux files (= files needed in the run dir to run the job)
-        #######################################################################
-
-        self.logger.debug("copying auxilliary files")  
-
-        # copy the files
-        for file in jMinder.auxFiles:
-            shutil.copy(file, runPath)
-            fileName = os.path.basename(file)
-            changePerm(join(runPath,fileName),0777)
-            
-        #######################################################################
-        # handle ref files
-        #######################################################################
-
-        self.logger.debug("copying reference files")  
-        self.logger.debug(jDescriptor.paths.refDB)
-        self.retrieveRefFiles(jMinder)
-
-        #######################################################################
-        # handle aux files desguised as string in the kit
-        #######################################################################
-
-        kit = JobGroupKits().kit(jDescriptor.jobGroup)
-
-        for  item in kit.getItem('stringsToFiles'):
-            filename = item[0]
-            string = item[1]
-            file = open(join(runPath, filename), 'w')
-            file.write(string)
-            file.close()
-
-
-        #
-        # Now perform tasks specific to the different descriptors
-        #
-
-        baseClasses = getInheritanceHeirarchy(jDescriptor)
-        if 'AthenaJobDescriptor' in baseClasses:
-            self.athenaSetUp(jDescriptor)
-
-        if 'WatcherJobDescriptor' in baseClasses:
-            self.watcherSetUp(jMinder)
-            
-        if 'WorkerJobDescriptor' in baseClasses:
-            # self.workerSetUp(jMinder)
-            pass
-
-        if 'ReleaseJobDescriptor' in baseClasses:
-            self.releaseSetUp(jDescriptor)
-            
-        self.logger.info("returning from setupRunDir()")
-
-
-
-# --------------------------------------------------------------------------
-
-if __name__ == "__main__":
-    pass
diff --git a/Tools/RunTimeTester/src/ErrorDescriptor.py b/Tools/RunTimeTester/src/ErrorDescriptor.py
deleted file mode 100755
index cafb1243cc91ef7ef03190d0f0585f5879b1c1e1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ErrorDescriptor.py
+++ /dev/null
@@ -1,62 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""This is not part of the Descriptor heirarchy (despite the name)."""
-
-class ErrorDescriptor:
-    """When an error is encountered during Minder construction, an ErrorJobMNinder is created in MinderFactory.
-    ErrorDescriptor is a step in the process - it collects all the info needed to make the ErrorJobMinder"""
-    def __init__(self,desc):
-        'copy stuff from the dead descriptor needed for reporting and error reporting'
-        
-        # make the attributes visible to PyLint
-        self.name              = None
-        self.identifiedName    = None
-        self.errorMessages     = None
-        self.constructionError = None
-        self.rttPilotJob       = None
-        self.rttATNJob         = None
-        self.runPath           = None
-        self.resPath           = None
-        self.paths             = None
-        self.confederation     = None
-        self.jobSerialNumber   = None
-        self.log               = None
-        self.jobGroup          = None
-        self.displayClass      = None
-        self.displayProcess    = None
-        self.displayComponent  = None
-        self.hashString        = None
-        self.jobDisplayName    = None
-        self.jobDocString      = None
-        self.jobDocURL         = None
-        self.elog              = None
-        self.logger            = None
-
-        desckeys = desc.__dict__.keys()
-        for k in self.__dict__.keys():
-            if k in desckeys:
-                self.__dict__[k] = desc.__dict__[k] 
-        
-
-    def __str__(self):
-        'Provide a string description'
-
-        s = ''
-        for item in self.__dict__.items():
-            s+= '%30s              %s' % item
-        return s
-
-    def dataForXMLNode(self):
-        'Nothing to send up to RTTummaryLog (info will be obtined from ErrorJobMinder'
-        return {}
-    
-    # --------------------------------------------------------------
-    
-    def dump(self):
-        'Write tring description to the log'
-
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info('|Error job descriptor  dump                 |')
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info(self.__str__())
-        self.logger.info('')
diff --git a/Tools/RunTimeTester/src/ErrorJobMinder.py b/Tools/RunTimeTester/src/ErrorJobMinder.py
deleted file mode 100755
index 9083197bbddf440102756beba1224d9f4122489b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ErrorJobMinder.py
+++ /dev/null
@@ -1,84 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from MinderBase       import MinderBase
-from ErrorMinderToXML import ErrorMinderToXML
-# =============================================================================
-class ErrorJobMinder(MinderBase):
-    def __init__(self, argBag, jDescriptor):
-
-        MinderBase.__init__(self, argBag.logDir,
-                            argBag.package,
-                            jDescriptor.identifiedName,
-                            jDescriptor.jobSerialNumber,
-                            argBag.elementCreator,
-                            argBag.textNodeCreator
-                            )
-
-        self.elementCreator   = argBag.elementCreator
-        self.textNodeCreator  = argBag.textNodeCreator
-        self.xmlConverter     = ErrorMinderToXML(self)
-        self.errorMessages    = jDescriptor.errorMessages
-        self.name             = jDescriptor.name
-        self.jobSerialNumber  = jDescriptor.jobSerialNumber
-        self.jobGroup         = 'ErrorGroup'
-        self.resPath          = jDescriptor.resPath
-        self.runPath          = jDescriptor.runPath
-        self.neverCopyAndNeverDelete = []
-        
-        self.done             = False
-        self.logger.info('ErrorMinder is now in charge of job %s\nError messages: %s' % (self.identifiedName,
-                                                                                         self.errorMessages))
-        
-    def unlink(self):
-        'break circular references to allow garbage collection'
-
-        self.stateEngine.unlink()
-        del self.stateEngine
-        self.xmlConverter.unlink()
-        del self.xmlConverter
-        MinderBase.unlink(self)
-
-
-        
-    def setDone(self):
-        self.logger.info('ErrorJobMinder: setDone: error messages: %s' % str(self.errorMessages))
-        self.done = True
-        m = 'done set in Error Minder %s' % self.identifiedName
-        self.logger.info(m)
-        self.xmlConverter.update()
-
-
-    def status(self): return 'error'
-
-    def dataForMonitoring(self):
-        dict = MinderBase.dataForMonitoring(self)
-        dict['jobSuccess'] = False
-        dict['jobFailure'] = True
-        dict['done']       = True
-        return dict
-
-    def isSuccess(self): return False
-    def isDone(self):    return self.done
-
-    # --------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|            ErrorJobMinder  dump           |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('\n'+self.__str__())
-        
-    def fromQueued(self):
-        self.setDone()
-        return 'done'
-    
-    def fromOperatorCloseDown(self):
-        self.setDone()
-        return 'done'
-
-    def fromDone(self):
-        self.setDone()
-        return 'done'
-        
diff --git a/Tools/RunTimeTester/src/ErrorMinderStateEngine.py b/Tools/RunTimeTester/src/ErrorMinderStateEngine.py
deleted file mode 100755
index 08c3819210f2ece9822cb9f9e7c1eee066b4cd2a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ErrorMinderStateEngine.py
+++ /dev/null
@@ -1,22 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-'State enegine for Error Minders'
-
-from MinderStateEngineBase import MinderStateEngineBase
-
-
-class ErrorMinderStateEngine(MinderStateEngineBase):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-    
-    def __init__(self, minder, state='queued'):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-        self.dispatchTable = {
-            'queued':            minder.fromQueued,
-            'operatorCloseDown': minder.fromOperatorCloseDown,
-            'done':              minder.fromDone
-            }
-        
diff --git a/Tools/RunTimeTester/src/ErrorMinderStateEngine.py.obsolete b/Tools/RunTimeTester/src/ErrorMinderStateEngine.py.obsolete
deleted file mode 100755
index 970932492c4f960ddb57e2343f65dc91c2f43086..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ErrorMinderStateEngine.py.obsolete
+++ /dev/null
@@ -1,29 +0,0 @@
-'State enegine for Error Minders'
-
-from MinderStateEngine import MinderStateEngine
-
-
-class ErrorMinderStateEngine(MinderStateEngine):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-    
-    def __init__(self, minder, state='queued'):
-        MinderStateEngine.__init__(self, minder, state)
-        
-    # -----------------------------------------------
-
-    def functionsMinderMustHave(self): return []
-
-    # -----------------------------------------------
-
-    def fromQueued(self):self.state.set('done')
-
-    # -----------------------------------------------
-
-    def fromRunning(self):
-        if self.operatorCloseDownTest(): self.state.set('operatorCloseDown') 
-        self.state.set('done')
-
diff --git a/Tools/RunTimeTester/src/ErrorMinderToXML.py b/Tools/RunTimeTester/src/ErrorMinderToXML.py
deleted file mode 100755
index 1029ce12edd91b82c30dd5cbb53de55f0945c5e5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ErrorMinderToXML.py
+++ /dev/null
@@ -1,129 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an aml file summary of an RTT run
-"""
-import sys, os, fnmatch
-from StringIO      import StringIO
-from makeRTTDOM    import makeRTTDOM
-from xml.dom.ext   import PrettyPrint
-from Defs          import RTTdefs
-
-from MethodTimer import MethodTimer
-
-class ErrorMinderToXML(MethodTimer):
-    
-    def __init__(self, minder):
-
-        MethodTimer.__init__(self)
-        
-        self.minder          = minder
-        self.logger          = minder.logger
-        self.elementCreator  = minder.elementCreator
-        self.textNodeCreator = minder.textNodeCreator
-        self.callBack        = None
-
-    def unlink(self):
-        'break circular references to allow garbage collection'
-        del self.minder
-    def update(self):
-        self.logger.debug('%s: XMLconversion starts' % self.__class__.__name__)
-        minder = self.minder # for nested methods
-        # nested scopes are used to pass parameters to these nested functions
-        def findErrorMessages():
-
-            element = self.elementCreator('errors')
-            
-            text = reduce(lambda x, y: x+y+'\n',
-                          minder.errorMessages,
-                          '')
-            if text:
-                tn = self.textNodeCreator(text)
-                element.appendChild(tn)
-
-            return element
-
-        def findDataSets():
-            return self.elementCreator('datasets')
-
-        def findNtestSuccess():
-            return self.elementCreator('sUserTests')
-
-        def makeKeepFileNode():
-            return self.elementCreator('deliveredkeepfiles')
-            
-        def makeTextNode((tagName, text)):
-            tn = self.textNodeCreator(str(text).strip())
-            el = self.elementCreator(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(document))
-        
-        mElement = self.elementCreator('minder')
-        textInfo = [
-            ('identifiedName',  self.minder.identifiedName),
-            ('jobName',         self.minder.name),
-            ('jobID',           self.minder.jobSerialNumber),
-            ('state',           self.minder.stateEngine.state.state),
-            ('status',          self.minder.status()),
-            ('jobGroup',        self.minder.jobGroup),
-            ('resultsPath',     self.minder.resPath),
-            ('stateHistory',    str(self.minder.stateEngine.state.history))
-            ]
-
-        [mElement.appendChild(makeTextNode(ti)) for ti in textInfo]
-        
-        mElement.appendChild(findDataSets())
-        mElement.appendChild(findNtestSuccess() )
-        mElement.appendChild(makeKeepFileNode() )
-        mElement.appendChild(findErrorMessages() )
-          
-        
-        def makeFlagNode((tagName, flag)):
-            if flag: return self.elementCreator(str(tagName))
-            return None
-
-        flagInfo = [
-            ('rttPilotJob',   self.minder.rttPilotJob),
-            ('rttATNJob',     self.minder.rttATNJob)
-            ]
-
-        for f in flagInfo:
-            node = makeFlagNode(f)
-            if node: mElement.appendChild(node)
-
-        if self.minder.isDone():
-            mElement.appendChild(self.elementCreator('done'))
-        
-        m = 'Created, Error minder node: %s' % str(mElement)
-        self.logger.debug(m)
-        dict = self.minder.dataForMonitoring()
-        if self.callBack:
-            self.callBack.update(dict, mElement)
-            self.logger.debug('%s: call back performed' % self.__class__.__name__)
-
-        self.logger.debug('%s: XMLconversion ends' % self.__class__.__name__)
-        self.minderXMLNode =  mElement
-
-    def makeSummaryReport(self):
-        if not self.callBack:
-            self.logger.error('No object for Minder to call back to, abandoning update')
-            return 
-        self.update()
-        assert (self.minderXMLNode != None)
-        dict = self.minder.dataForMonitoring()
-        self.callBack.update(dict, self.minderXMLNode)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.update(), temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-
-
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/Evaluate.py b/Tools/RunTimeTester/src/Evaluate.py
deleted file mode 100755
index 751e571fd9c64208b017c05ea2fcc226c0809f32..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Evaluate.py
+++ /dev/null
@@ -1,216 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Replace xpath version Evaluate because I (PS)cant get xpath to work at cern
-for the moment (14/6/05).
-"""
-
-from xml.dom.minidom import parse
-from exc2string2     import exc2string2
-from RTTSException   import RTTCodingError
-
-import os.path
-import re
-import string
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-# This is the main function of this module, call this one
-def Evaluate(path, element):
-    # This function does everything except the washing up.
-    # It can be used to grab just about any elements or attributes you wish
-    # by an appropriately formed path parameter string.
-    # 
-    # ---- Elements:
-    # 'a/b/c' ==> return a list of all matching elements c
-    # 'a/b/c/text()' ==> return a list of the text content of all matching elements c
-    # 'a/b[0]/c' ==> return a list of all c tags below the first b tag
-    # '//b' ==> return a list of all b elements wherever they are situated
-    # 
-    # ---- Attributes
-    # 'a/b["hello"]' ==> return a list of value of the 'hello' attribute in all matching b elements
-
-    def isIndexSelector(sel):
-        sel = sel[1:-1]
-        try:
-            int(sel)
-            return True
-        except:
-            return False
-    def isAttrSelector(sel):
-        sel = sel[1:-1]
-        return sel[0]==sel[-1] and sel[0] in ['"', "'"]
-    def hasSelectors(path): # has at least one element like a[0] or a["attr"]?
-        return len(re.findall('\[.*?\]', path)) > 0
-    def getSelectors(path, onlyAttrs=False):
-        sels = re.findall('\[.*?\]', path)
-        if not onlyAttrs: return sels
-        return [s for s in sels if not isIndexSelector(s)] # return items which have non-integers in them
-    def wantsText(path):
-        return path.endswith('/text()')
-    def illegalPath(path):
-        # Illegal characters in path?
-        legal = ['.',',','-','_','"',"'",'(',')','/', '[', ']']
-        punctuation = string.punctuation
-        for b in legal:
-            punctuation = punctuation.replace(b, '')
-
-        ok = True not in [p in string.whitespace for p in path] # no whitespace chars
-        ok = ok and (False not in [p in string.printable for p in path]) # must only contain printable chars
-        ok = ok and (True not in [p in punctuation for p in path]) # must not contain punctuation chars
-        ok = ok and (path.count('text()')==path.count('(') and path.count('text()')==path.count(')'))
-        if not ok: return (True, "Illegal characters found in %s" % path)
-
-        # Badly constructed path?
-        bad = path.count('text()') > 1 or ('text()' in path and not path.endswith('text()'))
-        if bad: return (True, "%s: Illegally placed 'text()'." % path)
-
-        # Bad use of selectors in path?
-        selectors = getSelectors(path, True) # get only attribute selectors i.e. a['hello'], rather than, say, a[9]
-        if selectors:
-            bad = len(selectors) > 1
-            bad = bad or (selectors[0] not in path.split('/')[-1]) # bad = selector not at end of path
-            bad = bad or (False in [a[0]=='/' for a in path.split(']')[1:] if a]) # bad = selector not followed by '/' or end of path
-            bad = bad or (not isAttrSelector(selectors[0]) and not isIndexSelector(selectors[0]))
-            if bad: return (True, "%s: Illegally placed attribute selector(s)." % path)
-            
-        # All ok
-        return (False, "")
-    def next(path):
-        for thing in path.split('/'): yield thing
-    def applySelector(els, sel):
-        isAttr  = isAttrSelector(sel)
-        isIndex = isIndexSelector(sel)
-        if isAttr:
-            sel = sel[2:-2] # remove [ and ], as well as quotes
-            attrs = [el.getAttribute(sel) for el in els]
-            if len(attrs)==1: return str(attrs[0])
-            if len(attrs)>1: return attrs
-            return None
-        if isIndex:
-            sel = sel[1:-1] # remove [ and ], as well as quotes
-            return els[int(sel)] # may throw an out of bounds exception; list return required to allow selector chaining
-        
-    # -----------------------------------------------------------------------
-    # ---- Method proper starts here, the above are nested helper functions
-    # -----------------------------------------------------------------------
-    isIllegal, message = illegalPath(path)
-    if isIllegal:
-        m = 'Evaluate: IllegalXPathExpression: %s\nTraceback:\n' % (message, exc2string2())
-        raise RTTCodingError(m)
-        
-
-    # Does not have any selectors like a[0] or a['attr']
-    if not hasSelectors(path):
-        if not wantsText(path):
-            return _Evaluate(path, element)
-
-        # has a text() at the end
-        pathMinusText = path[:path.rfind('/')]
-        els = _Evaluate(pathMinusText, element)
-        if not els: return ''
-        blah = [_Evaluate('text()', el) for el in els]
-        if len(blah)==1: return blah[0]
-        return blah
-
-    # OK, it does have selectors    
-    parents = [element]
-    result = None
-    for tok in next(path):
-        sels = getSelectors(tok)
-        tok = tok.split('[')[0].strip() # remove any selectors from the string
-
-        newparents = []
-        for parent in parents:
-            children = _Evaluate(tok, parent)
-            for sel in sels:
-                children = applySelector(children, sel)
-                if type(children)==type('') or type(children)==type(u''): return str(children) # got a final answer, just return it
-            if type(children)==type('') or type(children)==type(u''): return children # got a final answer, just return it
-
-            try:
-                newparents.extend(children)
-            except:
-                newparents.append(children)
-
-        parents = newparents
-
-
-    if wantsText(path):
-        blah = [_Evaluate('text()', p) for p in parents]
-        if len(blah)==1: return blah[0]
-        return blah
-
-    if len(parents)==1 and path.endswith(']'):
-        return parents[0]
-    return parents
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-def _Evaluate(path, element):
-    paths = path.split('/')
-    curGenEls = [element]
-    for p in paths:
-        nextGenEls = []
-        if p == 'text()':
-            texts = []
-            [texts.extend(getText(e)) for e in curGenEls]
-            text = reduce(lambda x,y:x+y, texts,'')
-            text = text.strip()
-            return text
-        else:
-            [nextGenEls.extend(e.childNodes) for e in curGenEls]
-            nextGenEls = [n for n in nextGenEls if n.nodeName == p]
-            
-        curGenEls = nextGenEls
-    return curGenEls
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-def getText(element):
-    texts = [node.data for node in element.childNodes if node.nodeType == node.TEXT_NODE]
-    text = reduce(lambda x,y:x+y, texts, '')
-    return text.strip()
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-if __name__=="__main__":
-
-    import getopt, sys, xml.dom.minidom
-    from findRTThome import findRTThome
-
-    from setupLoggerForDebug import setupLoggerForDebug
-    setupLoggerForDebug()
-
-    try:
-        opts = getopt.getopt(sys.argv[1:],'c:')
-    except getopt.GetoptError:
-        print 'bad options'
-        sys.exit(2)
-
-    cfg = None
-    for opt in opts[0]:
-        print opt
-        if opt[0] == '-c': cfg = opt[1]
-
-
-    (base, cfg, src) = findRTThome(cfg)
-
-    if not os.path.exists:
-        print 'could not find cfg file',cfg
-
-    dom = xml.dom.minidom.parse(cfg)
-
-    pathsToEvaluate = ['rttconfig', 'mode', 'mode/text()']
-    for p in pathsToEvaluate:
-        print 'path to find ',p
-        print Evaluate(p, dom.documentElement)
-        
diff --git a/Tools/RunTimeTester/src/Factory_BaseClass.py b/Tools/RunTimeTester/src/Factory_BaseClass.py
deleted file mode 100755
index 4a48247fb96c33b7728b0da7c9c82f3be53a14a3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_BaseClass.py
+++ /dev/null
@@ -1,14 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-# -------------------------------------------------------------------------
-# Base Class
-# -------------------------------------------------------------------------
-class Factory:
-  def __init__(self, logger, name='Factory'):
-    # logger.debug('Init for class %s' % name)
-    self.logger = logger
-    self.name = name
diff --git a/Tools/RunTimeTester/src/Factory_BaseDirectoryMaker.py b/Tools/RunTimeTester/src/Factory_BaseDirectoryMaker.py
deleted file mode 100755
index c3af2c8f6588b135a95a14d49b1610865d2160cd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_BaseDirectoryMaker.py
+++ /dev/null
@@ -1,24 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-
-from Factory_BaseClass     import Factory
-from BaseDirectoryMaker   import BaseDirectoryMaker, CERNSLC4BaseDirectoryMaker    
-class BaseDirectoryMakerFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'CERNslc4':  CERNSLC4BaseDirectoryMaker
-      }
-    
-  def create(self, paths): 
-    key = paths.site
-    if key in self.dict:
-        return self.dict[key](paths,self.logger)
-    else:
-        return BaseDirectoryMaker(paths, self.logger)
diff --git a/Tools/RunTimeTester/src/Factory_BigFileIO.py.obsolete b/Tools/RunTimeTester/src/Factory_BigFileIO.py.obsolete
deleted file mode 100644
index 630787d2b1649d161ec15ac2d08829f537730cd3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_BigFileIO.py.obsolete
+++ /dev/null
@@ -1,26 +0,0 @@
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-
-from Factory_BaseClass     import Factory
-from BigFileIO   import BigFileIO, CERNBigFileIO
-
-class BigFileIOFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'CERNslc4':  CERNBigFileIO
-      }
-    
-  def create(self, paths, doNothing=True): 
-    key = paths.site
-    self.logger.debug('BigFileFactory: grabbing key: %s' % key)
-    if key in self.dict:
-      inst = self.dict[key](paths, doNothing)
-      self.logger.debug('Key in dict: returning inst class: %s' % inst.__class__.__name__)
-      return inst
-    else:
-      return BigFileIO(paths, doNothing)
diff --git a/Tools/RunTimeTester/src/Factory_CMTConfigurer.py b/Tools/RunTimeTester/src/Factory_CMTConfigurer.py
deleted file mode 100755
index 99b55090676599c3fae40d8b6f0343361100baa4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_CMTConfigurer.py
+++ /dev/null
@@ -1,32 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-
-from Factory_BaseClass import Factory
-from CMTConfigurer     import CMTConfigurerBuild
-from CMTConfigurer     import CMTConfigurerLocal
-
-class CMTConfigurerFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-    self.dict = {
-      'afs_build'     : CMTConfigurerBuild,
-      'localPackages' : CMTConfigurerLocal
-      }
-
-  def makeKey(self, paths):    
-    if paths.localRTTRun:     return 'localPackages'
-    return 'afs_build'
-    
-  def create(self, paths, configArgDict): 
-    key = self.makeKey(paths)
-
-    configurer =  self.dict[key](configArgDict, self.logger)
-
-    self.logger.debug('Instantiated a CMTConfigurer of type %s' % (configurer.__class__.__name__))
-
-    return configurer
-  
-
-
diff --git a/Tools/RunTimeTester/src/Factory_CmtLines.py b/Tools/RunTimeTester/src/Factory_CmtLines.py
deleted file mode 100755
index d156e06649a24224d627f64255ccd3f7a0dc312f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_CmtLines.py
+++ /dev/null
@@ -1,38 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass    import Factory
-from CmtLines import CmtLinesBuild, CmtLinesLocal, CmtLinesBuildHLT
-
-import sys
-
-class CmtLinesFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-    self.dict = {
-      'build':         CmtLinesBuild,
-      'hlt':           CmtLinesBuildHLT,
-      'local':         CmtLinesLocal
-      }
-    
-  def create(self, paths, argDict):
-    """
-    it is overridden by local RTT runs
-    """
-
-
-    mode = paths.runType # now only build
-    if paths.localRTTRun: mode = 'local'
-
-    if mode in self.dict.keys():
-      if paths.topProject == 'AtlasHLT': mode='hlt'
-      cmtl =  self.dict[mode](self.logger, argDict, useRunTime=True)
-      self.logger.debug('Creating a %s' % cmtl.__class__.__name__)
-      return cmtl
-    else:
-      msg = 'Unknown mode, could not instantiate a CmtLines: '+str(mode)
-      self.logger.info(msg)
-      sys.exit(0)
- 
-
-
diff --git a/Tools/RunTimeTester/src/Factory_CmtLinesForScripts.py b/Tools/RunTimeTester/src/Factory_CmtLinesForScripts.py
deleted file mode 100755
index df4ab5015b877f5e9482d2032c432a6c811e3d94..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_CmtLinesForScripts.py
+++ /dev/null
@@ -1,43 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-from Factory_BaseClass  import Factory
-from CmtLines           import CmtLinesBuild, CmtLinesLocal, CmtLinesBuildHLT
-
-
-class CmtLinesForScriptsFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-    self.dict = {
-      'build':         CmtLinesBuild,
-      'hlt':           CmtLinesBuildHLT,
-      'local':         CmtLinesLocal
-      }
-    
-
-    # self.HLTJobGroups = ['AthenaHLTTesting']
-    
-  def create(self, jDescriptor):
-    """
-    mode release only .
-    it is overridden by local RTT runs
-    """
-
-    paths = jDescriptor.paths
-
-    mode = paths.runType # now only  build
-    if paths.localRTTRun: mode = 'local'
-
-    if mode in self.dict.keys():
-      if paths.topProject == 'AtlasHLT': mode = 'hlt'
-      cmtl =  self.dict[mode](self.logger, paths, useRunTime=True)
-      self.logger.debug('Creating a %s' % cmtl.__class__.__name__)
-      return cmtl
-    else:
-      msg = 'Unknown mode, could not instantiate a CmtLines: '+str(mode)
-      self.logger.info(msg)
-      sys.exit(0)
- 
-
-
diff --git a/Tools/RunTimeTester/src/Factory_Commander.py b/Tools/RunTimeTester/src/Factory_Commander.py
deleted file mode 100755
index 55280010cedebf5e937c3cb5e27d434284231b3b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_Commander.py
+++ /dev/null
@@ -1,41 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-from Factory_BaseClass    import Factory
-from Commander            import LinuxInteractiveCommander
-from Commander            import DummyCommander
-from Commander            import LSFBatchCommander
-from RTTSException        import RTTCodingError
-from exc2string2          import exc2string2
-
-import sys
-
-
-class CommanderFactory(Factory):
-  def __init__(self, mode, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-    self.mode      = mode
-    
-    self.dict = {
-      'LinuxInteractive': LinuxInteractiveCommander,
-      'LSFBatch':         LSFBatchCommander,
-      'QueuedDummy':      DummyCommander,
-      'RunningDummy':     DummyCommander,
-      'TransparentDummy': DummyCommander,
-      'Dummy':            DummyCommander,
-      }
-    
-    
-  def create(self):
-    'Returns a Commander instance according to mode'
-    
-    if self.mode in self.dict.keys():
-      return self.dict[self.mode]()
-    else:
-      msg = 'CommanderFactory: Unknown mode: %s, could not instantiate a Commander\nTraceback:\n%s' % (self.mode,
-                                                                                                       exc2string2())
-      self.logger.error(msg)
-      raise RTTCodingError(msg)
-      
-
diff --git a/Tools/RunTimeTester/src/Factory_DataSet.py b/Tools/RunTimeTester/src/Factory_DataSet.py
deleted file mode 100755
index 58125e67f4bde3a585d43ac44fc48367d6df661a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_DataSet.py
+++ /dev/null
@@ -1,147 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from exc2string           import exc2string
-from Factory_BaseClass    import Factory
-from Factory_IOUtils      import IOUtilsFactory
-from RTTSException        import RTTInputError
-
-import os
-import time
-
-class DataSetFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    
-    self.dict = {
-        'dc2'              : DC2DataSet,
-        'bytestream'       : ByteStreamDataSet,
-        'jobTransformData' : JobTransformDataSet
-        }
-    
-    
-  def create(self, datasetType, physicalName, logicalName, origin):
-    "Factory for contructing DataSet objects"
-    
-    try:
-        return self.dict[datasetType]((physicalName, logicalName), origin, self.logger)
-    except KeyError:
-        pass
-
-
-# ----------------------------------------------------------------
-
-class DataSet:
-  def __init__(self, (physical, logical), origin, logger):
-    self.physical = physical
-    self.logical  = logical 
-    self.type     = None
-    self.mtime    = 'Unknown'
-    self.size     = 'Unknown'
-    self.origin   = origin
-    self.exists   = False
-    self.logger   = logger
-    self.ioutils  = None
-    
-    #self.setIOutils()
-    #if not self.ioutils: return # some datasets have physical path set to Unknown
-    #self.setStats()
-    
-  def setStats(self):
-    if not self.ioutils: self.setIOutils()
-    
-    if self.ioutils and self.ioutils.exists(self.physical):
-        self.exists = True
-        self.mtime  = self.ioutils.lastmodtime(self.physical)
-        self.size   = str(self.ioutils.filesize(self.physical))+'B'
-
-  def exists_(self):
-    if not self.ioutils:
-      self.logger.error('Factory_Dataset => self.ioutils does not exist, cannot check if dataset exists_()')
-      return False
-    return self.ioutils.exists(self.physical)
-  
-  def whichAthenaList(self):
-    pass
-
-  def __str__(self):
-    return self.physical
-
-  def castorProtocol(self):
-    return 'castor:'
-
-  def isCastorBased(self):
-    return self.physical.startswith('/castor/cern.ch/')
-
-  def isAFSBased(self):
-    return self.physical.startswith('/afs/cern.ch/atlas')
-
-  def setIOutils(self):
-    if self.isCastorBased(): self.ioutils = IOUtilsFactory(self.logger).create('castor')
-    elif self.isAFSBased(): self.ioutils = IOUtilsFactory(self.logger).create('afs')
-    else: self.ioutils = None
-  
-class JobTransformDataSet(DataSet):
-  def __init__(self,(physical, logical), origin, logger):
-    DataSet.__init__(self, (physical, logical), origin, logger)
-    self.type = 'JobTransform'
-      
-class ByteStreamDataSet(DataSet):
-  def __init__(self, (physical, logical), origin, logger):
-    DataSet.__init__(self, (physical, logical), origin, logger)
-    self.type = 'ByteStream'
-
-  def whichAthenaList(self):
-    return 'ByteStreamInputSvc.FullFileName'
-
-  def castorProtocol(self):
-    return 'rfio:'
-    
-class DC2DataSet(DataSet):
-  def __init__(self, (physical, logical), origin, logger):
-    DataSet.__init__(self, (physical, logical), origin, logger)
-    self.type = 'DC2'
-        
-  def whichAthenaList(self):
-    return 'ServiceMgr.EventSelector.InputCollections'
-
-class DQ2DataSet:
-  "dataset from dq2"
-  def __init__(self, dsName, files, type, origin):
-    self.name  = dsName
-    self.files = files
-    self.type  = type
-    self.logicalFiles  = self.setLogicalFiles()
-    self.origin = origin
-    
-  def __str__(self):
-    self.logicalFiles.sort()
-    return str(self.logicalFiles)
-    
-  def setLogicalFiles(self):
-    lf = []
-    for file in self.files:
-      fn =os.path.join(self.name, file)
-      lf.append(fn)
-    return lf
-
-  def whichAthenaList(self):
-
-    lists ={
-      'dc2':        'ServiceMgr.EventSelector.InputCollections',
-      'bytestream': 'ByteStreamInputSvc.FullFileName'
-      }
-
-    if self.type in lists.keys():
-      return lists[self.type]
-    else:
-      print 'Unkown dataset type!'
-      
-      
-      
diff --git a/Tools/RunTimeTester/src/Factory_DirectoryMaker.py.OBSOLETE b/Tools/RunTimeTester/src/Factory_DirectoryMaker.py.OBSOLETE
deleted file mode 100755
index 6af293fc7e83ab143bfe4f6549df4fa9d9ba8ac8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_DirectoryMaker.py.OBSOLETE
+++ /dev/null
@@ -1,46 +0,0 @@
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from exc2string           import exc2string
-
-from Factory_BaseClass    import Factory
-from DirectoryMaker       import *
-
-# -------------------------------------------------------------------------
-# DirectoryMaker Factory
-# -------------------------------------------------------------------------
-
-class DirectoryMakerFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    
-    self.dict = {
-      'NoNewDirs':    DummyDirectoryMaker,
-      'NoNewRunDirs': NoWorkDirCreation,
-      'project':      DirectoryMakerProject,
-      'monolithic':   DirectoryMaker
-      }
-    
-    
-  def create(self, paths):
-    "Factory for contructing JobMinders"
-    
-    # no checks on legality of mode - that should be done in pathsMaker,
-    # and this dictionary should conform to that.
-    
-    mode = paths.testMode
-    if mode in self.dict.keys():
-      return self.dict[mode](paths, self.logger)
-
-    mode = paths.releaseType
-    if mode in self.dict.keys():
-      return self.dict[mode](paths, self.logger)
-
-    msg  = 'Cannot tell which DirectoryMaker to construct '
-    msg += 'testMode = %s releaseType = %s' % (paths.testMode,
-                                               paths.releaseType)
-
-    self.logger.error(msg)
diff --git a/Tools/RunTimeTester/src/Factory_HardCoded.py b/Tools/RunTimeTester/src/Factory_HardCoded.py
deleted file mode 100755
index 3335ece08c06b281345aa6cc74336d02610a5085..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_HardCoded.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass   import Factory
-from HardCoded           import HardCodedProjectRelease
-from HardCoded           import HardCodedProjectKit
-from HardCoded           import HardCodedBase
-
-class HardCodedFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      ('project', 'build'):         HardCodedProjectRelease,
-      ('project', 'kit'):           HardCodedProjectKit,
-      ('base', None):               HardCodedBase
-      }
-  def create(self, releaseType, runType):
-
-    mode = (releaseType, runType)    
-    return self.dict[mode]()
- 
-
-
diff --git a/Tools/RunTimeTester/src/Factory_IOUtils.py b/Tools/RunTimeTester/src/Factory_IOUtils.py
deleted file mode 100644
index 0296d7495ce6569f889c8cda275c4b7569a07085..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_IOUtils.py
+++ /dev/null
@@ -1,161 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-import os.path
-import time
-import shutil
-from stat import ST_MTIME
-from exc2string2 import exc2string2
-
-from Factory_BaseClass import Factory
-from RTTSException     import RTTCodingError, RTTInputError, RTTTimeOut
-from Tools             import timed_cmd
-
-class IOUtilsFactory(Factory):
-    def __init__(self, logger):
-        Factory.__init__(self,  logger, self.__class__.__name__)
-            
-    def create(self, type_):
-        dict = {'afs':AFSIOUtils, 'castor':CastorIOUtils}
-        try:
-            return dict[type_](self.logger)
-        except KeyError:
-            m  = 'Unknown IO utils type %s, should be one of %s\n' % (type_, dict.keys())
-            m += 'This illegal type was likely input through your top level cfg file.'
-            raise RTTInputError(m)
-        except Exception, e:
-            raise RTTCodingError(str(e))
-        
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-# -----------------------------------------------------------------
-
-class AFSIOUtils:
-    def __init__(self, logger):
-        self.logger = logger
-        
-    def mkdir(self, dirToMake):
-        try:
-            timed_cmd('mkdir -p %s' % dirToMake)
-            return self.exists(dirToMake)
-        except:
-            return False
-    def exists(self, d):
-        return os.path.exists(d)
-    def isdir(self, d):
-        return os.path.isdir(d)
-    def isfile(self, f):
-        return os.path.isfile(f)
-    def filesize(self, fpath):
-        if not self.isfile(fpath): raise RTTInputError('%s: not a file' % fpath)
-        return os.path.getsize(fpath)
-    def lastmodtime(self, fpath, asSecsSinceEpoch=False):
-        if not self.isfile(fpath): raise RTTInputError('%s: not a file' % fpath)
-        stats = time.localtime(os.lstat(fpath)[ST_MTIME])
-        if asSecsSinceEpoch: return time.mktime(stats)
-        return time.strftime('%y/%m/%d@%H:%M', stats)
-    
-    def deleteStuffIn(self, path):
-        # Deletes everything inside directory path
-        if not self.exists(path) or not self.isdir(path): return
-        timed_cmd('rm -rf %s/*' % path)
-    def copyfile(self,src,dst):
-        dst = os.path.join(dst, os.path.basename(src))
-        try:
-            shutil.copyfile(src,dst)
-            return True
-        except:
-            self.logger.error(exc2string2())
-            return False
-    
-# -----------------------------------------------------------------
-
-class CastorIOUtils:
-    def __init__(self, logger):
-        self.logger = logger
-
-    def mkdir(self, dirToMake):
-        try:
-            timed_cmd('rfmkdir -p %s' % dirToMake, 30)
-            return True
-        except:
-            return False
-
-    def path_info(self, fpath):
-        try:
-            return timed_cmd('nsls -ld %s' % fpath, 30)
-        except:
-            return None
-
-    def isfile(self, fpath):
-        answer = self.path_info(fpath)
-        if not answer: return False
-        return len(answer)==1 and (not answer[0].strip().startswith('d') and not answer[0].strip().startswith('l'))
-        
-    def exists(self, fpath):
-        answer = self.path_info(fpath)
-        if not answer: return False
-        return len(answer)>1 or (len(answer)==1 and "No such file or directory" not in answer[0])
-
-    def isdir(self, fpath):
-        answer = self.path_info(fpath)
-        if not answer: return False
-        return len(answer)==1 and answer[0].strip().startswith('d')
-
-    def filesize(self, fpath):
-        if not self.isfile(fpath): raise RTTInputError('%s: not a file' % fpath)
-        info = self.path_info(fpath)
-        return int(info[0].strip().split()[4])
-
-    def lastmodtime(self, fpath, asSecsSinceEpoch=False):
-        """Return the last modifed time of fpath, either (default) as a string YY/MM/DD@HH:MM,
-        or as seconds since the Unix epoch."""
-        if not self.isfile(fpath): raise RTTInputError('%s: not a file' % fpath)
-        info = self.path_info(fpath)
-        toks = info[0].strip().split()
-
-        fmonth = toks[5]
-        fday   = toks[6]
-        
-        if len(toks[7].split(':'))==1: # year, not time
-            fyear = toks[7]
-            ftime = '?'
-        else:
-            fyear = '?'
-            ftime = toks[7]
-
-        if asSecsSinceEpoch:
-            from datetime import datetime
-            if fyear == '?': fyear = time.localtime()[0]
-            if ftime == '?': ftime = '00:00'
-            fmonth = {'Jan':1,'Feb':2,'Mar':3,'Apr':4,'May':5,'Jun':6,'Jul':7,'Aug':8,'Sep':9,'Oct':10,'Nov':11,'Dec':12}[fmonth]
-            dt = datetime(int(fyear),int(fmonth),int(fday),int(ftime.split(':')[0]), int(ftime.split(':')[1]))
-            return time.mktime(dt.timetuple())
-            
-        return '%s/%s/%s@%s' % (fyear, fmonth, fday, ftime)
-        
-    def copyfile(self, src, dst):
-        timed_cmd('rfcp %s %s;' % (src, dst), 120)
-        if not self.exists(os.path.join(dst,os.path.basename(src))):
-            m = 'Unable to copy file %s to %s' % (src, dst)
-            self.logger.error(m)
-            return False
-        return True
-    
-    def deleteStuffIn(self, path):
-        # Deletes everything inside directory path
-        if not self.exists(path) or not self.isdir(path): return            
-        timed_cmd('yes | rfrm -r %s;rfmkdir %s;' % (path, os.path.basename(path))) # rfrm -r asks if we are sure, hence pipe yes
-        print 'Deleted everything inside %s' % path
-
-if __name__ == '__main__':
-    path = '/castor/cern.ch/grid/atlas/atlasgroupdisk/proj-sit/rtt/moniFile.db'
-    path = '/castor/cern.ch/grid/atlas/atlasgroupdisk/proj-sit/rtt/brinick/archive/0/nightly/15.0.X.Y/build/i686-slc4-gcc34-opt/p1hlt'
-    # path = '/afs/cern.ch/atlas/project/RTT/prod/time_stamp'
-    cu = CastorIOUtils(None)
-    #cu = AFSIOUtils(None)
-    
-    # cu.path_info(path)
-    print cu.exists(path)
-    cu.deleteStuffIn(path)
-    # print cu.lastmodtime(path, True)
diff --git a/Tools/RunTimeTester/src/Factory_JobDescriptor.py b/Tools/RunTimeTester/src/Factory_JobDescriptor.py
deleted file mode 100755
index d0bb1c6a44dfc6ecc58c2a3352ec3279c2b3e488..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_JobDescriptor.py
+++ /dev/null
@@ -1,65 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from Factory_BaseClass      import Factory
-from AthenaJobDescriptor    import AthenaJobDescriptor
-from AthenaJobDescriptor    import AthenaPilotJobDescriptor
-from AthenaJobDescriptor    import AthenaATNJobDescriptor
-from Descriptor             import Descriptor
-from PseudoDescriptor       import PseudoDescriptor
-from RTTSException          import RTTCodingError
-from WorkerJobDescriptor    import WorkerJobDescriptor
-from WatcherJobDescriptor   import WatcherJobDescriptor
-from JobTransformDescriptor import JobTransformDescriptor
-
-from SequentialDescriptor   import SequentialDescriptor
-from ParallelDescriptor     import ParallelDescriptor
-from ChainJobDescriptor     import ChainJobDescriptor
-from exc2string2            import exc2string2
-
-from copy                 import deepcopy
-
-class DescriptorFactory(Factory):
-  def __init__(self, argBag, logger):
-    Factory.__init__(self,  logger, self.__class__.__name__)
-    self.argBag = argBag
-    self.dict = {
-      'Athena':        AthenaJobDescriptor,
-      'AthenaPilot':   AthenaPilotJobDescriptor,
-      'AthenaATN':     AthenaATNJobDescriptor,
-      'JobTransform':  JobTransformDescriptor, 
-      'Sequential':    SequentialDescriptor,
-      'Parallel':      ParallelDescriptor,
-      'Chain':         ChainJobDescriptor,
-      'Pseudo':        PseudoDescriptor
-     }
-
-    self.jobSerialNumber = -1
-  
-  def create(self, selector, argDict={}):
-    paths = self.argBag
-    if selector in self.dict:
-      # limit the information flow,
-      # and allow the descriptors to update their own paths object
-      # descPaths   = PathsForDescriptors(paths)
-      # clonedPaths = deepcopy(descPaths)
-      # desc        = self.dict[selector](clonedPaths, self.logger)
-      self.jobSerialNumber += 1
-      sn = self.jobSerialNumber
-      try:
-        desc        = self.dict[selector](paths, self.logger, sn, argDict)
-      except Exception, e:
-        msg = 'Error while instantiating  Descriptor type %s\nException: %s\nTraceback:\n%s' % (str(selector), str(e), exc2string2())
-        self.logger.error(msg)
-        raise RTTCodingError(msg)
-      else:
-        return desc
-    else:
-      msg = 'Unknown Descriptor type %s' % str(selector)
-      self.logger.error(msg)
-      raise RTTCodingError(msg)
-
diff --git a/Tools/RunTimeTester/src/Factory_JobGroupDirectoryMaker.py b/Tools/RunTimeTester/src/Factory_JobGroupDirectoryMaker.py
deleted file mode 100755
index e3bdbeeabdf01a2a831a59a7fee0e38b5df6ee44..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_JobGroupDirectoryMaker.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from exc2string                   import exc2string
-
-from Factory_BaseClass            import Factory
-from JobGroupDirectoryMaker       import JobGroupDirectoryMaker, JobGroupDirectoryMakerWatcher
-from RTTSException                import RTTCodingError
-from exc2string2                  import exc2string2
-# -------------------------------------------------------------------------
-# DirectoryMaker Factory
-# -------------------------------------------------------------------------
-
-class JobGroupDirectoryMakerFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    
-    self.dict = {
-      'WorkerJobMinder':           JobGroupDirectoryMaker,
-      'LSFBatchJobMinder':         JobGroupDirectoryMaker,
-      'PBSJobMinder':              JobGroupDirectoryMaker,
-      'LinuxInteractiveJobMinder': JobGroupDirectoryMaker,
-      'ErrorJobMinder':            JobGroupDirectoryMaker,
-      'PseudoJobMinder':           JobGroupDirectoryMaker,
-      'ChainJobMinder':            JobGroupDirectoryMaker,
-      'SequentialMinder':          JobGroupDirectoryMaker,
-      'ParallelMinder':            JobGroupDirectoryMaker,
-      'WatcherJobMinder':          JobGroupDirectoryMakerWatcher,
-      }
-    
-    
-  def create(self, descSetUpRunDirFn, minderClassName):
-    
-    if minderClassName in self.dict.keys():
-      return self.dict[minderClassName](descSetUpRunDirFn, self.logger)
-    try:
-      directoryMaker = self.dict[minderClassName](descSetUpRunDirFn, self.logger)
-    except KeyError:
-      msg  = 'JobGroupDirectoryMakerFactory: key not in dictionary: %s' % minderClassName
-      raise RTTCodingError(msg)
-    except Exception, e:
-      'Error constructing DirectoryMaker from minder type %s \nException: %s\Traceback:\n%s' % (minderClassName,
-                                                                                                str(e),
-                                                                                                exc2string2()
-                                                                                                )
-  
diff --git a/Tools/RunTimeTester/src/Factory_JobMaker.py.obsolete b/Tools/RunTimeTester/src/Factory_JobMaker.py.obsolete
deleted file mode 100755
index 443a8732c1007f0b447bc92e036ccfdc4df5ea44..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_JobMaker.py.obsolete
+++ /dev/null
@@ -1,21 +0,0 @@
-from Factory_BaseClass    import Factory
-from JobMaker             import RTTJobMaker, ATNJobMaker
-
-
-# -------------------------------------------------------------------------
-# JobMaker Factory
-# -------------------------------------------------------------------------
-class JobMakerFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    self.dict = {'RTT': RTTJobMaker,
-                 'ATN': ATNJobMaker}
-    
-  def create(self, confMode, jobListElement, paths):
-
-    if not self.dict.has_key(confMode):
-        self.logger.error('jobMakerFactory dictionary has no key '+confMode)
-        return None
-    
-    return self.dict[confMode](jobListElement, paths, self.logger)
-
diff --git a/Tools/RunTimeTester/src/Factory_JobsXMLReader.py b/Tools/RunTimeTester/src/Factory_JobsXMLReader.py
deleted file mode 100755
index 7f17758cefbd048ac35c6b8d95d950ae1e4769f9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_JobsXMLReader.py
+++ /dev/null
@@ -1,37 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass    import Factory
-from JobsXMLReader        import JobsXMLReader
-from JobsXMLReader        import JobsXMLReaderATN
-from exc2string2          import exc2string2
-from RTTSException        import RTTCodingError
-# -------------------------------------------------------------------------
-
-class JobsXMLReaderFactory(Factory):
-    def __init__(self, logger):
-        Factory.__init__(self, logger, self.__class__.__name__)
-        self.dict = {
-            'PBS':              JobsXMLReader,
-            'Dummy':            JobsXMLReader,
-            'FullDummy':        JobsXMLReader,
-            'EmptyDummy':       JobsXMLReader,
-            'LinuxInteractive': JobsXMLReader,
-            'LSFBatch':         JobsXMLReader,
-            'ATN':              JobsXMLReaderATN
-            }
-
-    def create(self, mode, element, rttPilotJobs, packageName, dataSetCatalog, chainStoreMaker,  descFactory):
-        if not self.dict.has_key(mode):
-            msg = 'JobsXMLReaderFactory dictionary has no key: %s' % mode
-            self.logger.error(msg)
-            return None
-        
-        try:
-            reader    = self.dict[mode](element, rttPilotJobs, packageName, dataSetCatalog, chainStoreMaker,  self.logger, descFactory)
-        except:
-            msg  = 'Error making JobsXMLReader\n'
-            msg += exc2string2()
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-        return reader
-
diff --git a/Tools/RunTimeTester/src/Factory_Launcher.py b/Tools/RunTimeTester/src/Factory_Launcher.py
deleted file mode 100755
index 97891763fb724203851e25d33c4182a40c77b4a2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_Launcher.py
+++ /dev/null
@@ -1,14 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Launchers govern the rate at which jobs are started.
-BatchLauncher,       submits all the jobs in rapid succession.
-InteractiveLauncher, waits for a job to finish before submitting the next.
-"""
-
-from Launcher               import ThrottleLauncher
-
-def launcherFactory(mode, minders, logger):
-    noToLaunch = 40
-    if mode == 'LinuxInteractive': noToLaunch = 1
-    return ThrottleLauncher(noToLaunch, minders, logger)
diff --git a/Tools/RunTimeTester/src/Factory_Legals.py b/Tools/RunTimeTester/src/Factory_Legals.py
deleted file mode 100755
index ce489cf3634c380577751879ec64f91dd6847ac8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_Legals.py
+++ /dev/null
@@ -1,16 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass   import Factory
-from Legals              import LegalsProject
-from Legals              import LegalsBase
-from Tools2              import releaseNumberComponents
-
-import sys
-
-class LegalsFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-  def create(self, releaseType, runType, hardCoded):
-    return LegalsProject(hardCoded)
-  
diff --git a/Tools/RunTimeTester/src/Factory_Minder.py b/Tools/RunTimeTester/src/Factory_Minder.py
deleted file mode 100755
index bf51a4b6d786f9a773ac85facab3cfa4e6b095c4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_Minder.py
+++ /dev/null
@@ -1,152 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-
-# some imports have been placed at end of
-# module to prevent circular import problems
-from Factory_BaseClass         import Factory
-from WatcherJobMinder          import WatcherJobMinder
-from ErrorJobMinder            import ErrorJobMinder
-from ErrorDescriptor           import ErrorDescriptor
-from PseudoJobMinder           import PseudoJobMinder
-from LinuxInteractiveJobMinder import LinuxInteractiveJobMinder
-from LSFBatchJobMinder         import LSFBatchJobMinder
-from ProductionManagerVetoer   import ProductionManagerVetoer
-from exc2string2               import exc2string2
-from getJobGroupKit            import getJobGroupKit
-import sys
-
-from RTTSException             import RTTCodingError
-
-from RTTConstructionFactoryHelpers import ArgBag
-from RTTConstructionFactoryHelpers import makeArgBag
-
-# -------------------------------------------------------------------------
-# Minder Factory
-# -------------------------------------------------------------------------
-
-class MinderFactory(Factory):
-  def __init__(self, jobVetoer, scriptWriterFactory, jobGroupDirectoryMakerFactory, jobGroupKits, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractive':      LinuxInteractiveJobMinder,
-      'Watcher':               WatcherJobMinder,
-      'Error':                 ErrorJobMinder,
-      'LSFBatch':              LSFBatchJobMinder,
-      'ChainJobDescriptor':    ChainJobMinder,
-      'SequentialDescriptor':  SequentialMinder,
-      'ParallelDescriptor':    ParallelMinder,
-      'PseudoDescriptor':      PseudoJobMinder
-      }
-
-    self.jobVetoer = jobVetoer
-    self.jgDirectoryMakerFactory = jobGroupDirectoryMakerFactory
-    self.scriptWriterFactory     = scriptWriterFactory
-    self.jobGroupKits            = jobGroupKits
-    
-  def makeKey(self, desc):
-    if desc.jobGroup == "RTTMonitor": return desc.jobGroup
-    if desc.constructionError:        return 'Error'
-    
-    keyByDesc = ['SequentialDescriptor', 'ParallelDescriptor', 'ChainJobDescriptor',
-                 'PseudoDescriptor']
-
-    descName = desc.__class__.__name__
-    if descName in  keyByDesc: return descName
-
-    if self.jobVetoer.toBeVetoed(desc): 
-      desc.errorMessages.append('Production Manager Veto')
-      return 'Error'
-
-  
-  def createMinder(self, argBag, jDescriptor):
-    self.logger.debug("Coming into createMinder: desc of type %s" % jDescriptor.__class__.__name__)
-    key = self.makeKey(jDescriptor)
-    self.logger.debug("Made key, value is: %s" % key)
-    if not key: key = argBag.mode
-    self.logger.debug("Now have key, value is: %s" % key)
-    if key not in self.dict.keys():
-        m = 'MinderFactory: Unknown key for minder type. key: %s ' % key
-        self.logger.error(m)
-        raise RTTCodingError(m)
-
-    minderClass = self.dict[key]
-    msg  = 'Instantiating a  %s from a %s with name %s for package %s' % (minderClass.__name__,
-                                                                          jDescriptor.__class__.__name__,
-                                                                          jDescriptor.identifiedName,
-                                                                          argBag.package)
-    self.logger.info(msg)
-    labels = ['jobGroup', 'runPath', 'castorSvcClass', 'jobTransformCmd', 'jobOptions']
-    descArgsForScripts = makeArgBag(labels, jDescriptor)
-    try:
-      descArgsForScripts.athenaCommand = jDescriptor.athenaCommand()
-    except:
-      pass
-
-    # add in things  jDescriptor _might_ have these will typically be collected
-    # by minders low in the Minder hierarchy
-    labels = ['outChainFiles', 'missingDatasets']
-    tempArgBag  = makeArgBag(labels, jDescriptor)
-    argBag.__dict__.update(tempArgBag.__dict__)
-
-    try:
-      argBag.jobGroupKit = getJobGroupKit(jDescriptor.jobGroup, self.jobGroupKits)
-    except Exception, e:
-      m  = 'MinderFactory: Error extracting job group kit descriptor '
-      m += 'type: %s, job group: %s, error message: %s\nTraceback:%s' % (jDescriptor.__class__.__name__,
-                                                                         jDescriptor.jobGroup,
-                                                                         str(e),
-                                                                         exc2string2())      
-      # Not sure if this is really a coding error: might be a configuration problem.
-      # In any case, the exception is caught by the calling method
-      raise RTTCodingError(m)
-
-    argBag.scriptWriter           = self.scriptWriterFactory.create(descArgsForScripts, self.logger)
-    argBag.errorMessages          = jDescriptor.errorMessages
-    argBag.resPath                = jDescriptor.resPath
-    argBag.runPath                = jDescriptor.runPath
-    argBag.identifiedName         = jDescriptor.identifiedName
-    argBag.log                    = jDescriptor.log
-    argBag.elog                   = jDescriptor.elog
-
-    minderClass                   = self.dict[key]
-    argBag.jobGroupDirectoryMaker = self.jgDirectoryMakerFactory.create(jDescriptor.setupRunDir, minderClass.__name__)
-
-    minder = minderClass(argBag, jDescriptor)
-    return minder
-
-  def createErrorMinder(self, argBag, jDescriptor):
-    desc = ErrorDescriptor(jDescriptor)
-    minder = ErrorJobMinder(argBag, desc)
-    return minder
-
-  def create(self, argBag, jDescriptor):
-    # creating normal minders can fall over,
-    # but creating ErrorMidners must be bullet-proof
-    # therefore keep very simple!
-    try:
-      minder = self.createMinder(argBag, jDescriptor)
-    except Exception, e:
-      msg  = 'Error constructing minder from a Descriptor of type %s, '   % jDescriptor.__class__.__name__
-      msg += 'replacing with ErrorMinder:\nException:\n%s\nTraceback\n%s' % (str(e), exc2string2())
-      self.logger.error(msg)
-      try:
-        minder = self.createErrorMinder(argBag, jDescriptor)
-      except Exception, e:
-        msg  = 'Error constructing ErrorMinder\n Traceback:\n  %s' % exc2string2()
-        self.logger.error(msg)
-        raise RTTCodingError(msg)
-      else:
-        minder.errorMessages.append(msg)
-        
-    self.logger.debug('Returning a minder of type %s' % minder.__class__.__name__)
-    return minder
-
-from ContainerMinder           import SequentialMinder
-from ContainerMinder           import ParallelMinder
-from ContainerMinder           import ChainJobMinder
diff --git a/Tools/RunTimeTester/src/Factory_MinderStateEngine.py b/Tools/RunTimeTester/src/Factory_MinderStateEngine.py
deleted file mode 100644
index fe44ee94fee63c54ccfdabe486a45ff77d000925..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_MinderStateEngine.py
+++ /dev/null
@@ -1,41 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass                 import Factory
-from LinuxInteractiveMinderStateEngine import LinuxInteractiveMinderStateEngine
-from WatcherMinderStateEngine          import WatcherMinderStateEngine
-from ErrorMinderStateEngine            import ErrorMinderStateEngine
-from BatchMinderStateEngine            import BatchMinderStateEngine
-from ContainerMinderStateEngine        import ContainerMinderStateEngine
-from ChainMinderStateEngine            import ChainMinderStateEngine
-
-from exc2string2                       import exc2string2
-from RTTSException                     import RTTCodingError
-
-class MinderStateEngineFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractiveJobMinder': LinuxInteractiveMinderStateEngine,
-      'WatcherJobMinder':          WatcherMinderStateEngine,
-      'ErrorJobMinder':            ErrorMinderStateEngine,
-      'LSFBatchJobMinder':         BatchMinderStateEngine,
-      'ContainerMinder':           ContainerMinderStateEngine,
-      'SequentialMinder':          ContainerMinderStateEngine,
-      'ParallelMinder':            ContainerMinderStateEngine,
-      'ChainJobMinder':            ChainMinderStateEngine,
-      'PseudoJobMinder':           BatchMinderStateEngine
-      }
-
-  def create(self, minder, state='queued'): 
-
-    key = minder.__class__.__name__
-
-    try:
-      engine =  self.dict[key](minder, state)
-    except:
-      msg = 'MinderStateEngineFactory: Cannot create a State engine for minder type %s:\nTraceback:\n%s' % (key, exc2string2())
-      print msg
-      raise RTTCodingError(msg)
-
-    return engine
diff --git a/Tools/RunTimeTester/src/Factory_MinderStateEngineDummy.py.obsolete b/Tools/RunTimeTester/src/Factory_MinderStateEngineDummy.py.obsolete
deleted file mode 100644
index 37b5c2f1a159af113f14b40a0c8f9c135f08bd2f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_MinderStateEngineDummy.py.obsolete
+++ /dev/null
@@ -1,38 +0,0 @@
-from Factory_BaseClass                 import Factory
-from ErrorMinderStateEngine            import ErrorMinderStateEngine
-from ContainerMinderStateEngine        import ContainerMinderStateEngine
-from MinderStateEngineDummy            import MinderStateEngineDummy
-
-from exc2string2                       import exc2string2
-from RTTSException                     import RTTCodingError
-
-class MinderStateEngineFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractiveJobMinder': MinderStateEngineDummy,
-      'Minder':                    MinderStateEngineDummy, 
-      'ErrorJobMinder':            ErrorMinderStateEngine,
-      'BatchJobMinder':            MinderStateEngineDummy, 
-      'LSFBatchJobMinder':         MinderStateEngineDummy, 
-      'ContainerMinder':           ContainerMinderStateEngine,
-      'SequentialMinder':          ContainerMinderStateEngine,
-      'ParallelMinder':            ContainerMinderStateEngine,
-      'ChainJobMinder':            ContainerMinderStateEngine,
-      'PseudoJobMinder':           MinderStateEngineDummy,
-      'WorkerMinder':              MinderStateEngineDummy
-      }
-
-  def create(self, minder, state='queued'): 
-
-    key = minder.__class__.__name__
-
-    try:
-      engine =  self.dict[key](minder, state)
-    except:
-      msg = 'MinderStateEngineFactory: Cannot create a State engine for minder type %s\nTraceback:\n' % (key, exc2string2())
-      print msg
-      raise RTTCodingError(msg)
-
-    return engine
diff --git a/Tools/RunTimeTester/src/Factory_MinderStateEngineReal.py.obsolete b/Tools/RunTimeTester/src/Factory_MinderStateEngineReal.py.obsolete
deleted file mode 100644
index 7675130672a64e74dd28c2e5f9c5a3a60fe54741..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_MinderStateEngineReal.py.obsolete
+++ /dev/null
@@ -1,41 +0,0 @@
-from Factory_BaseClass                 import Factory
-from LinuxInteractiveMinderStateEngine import LinuxInteractiveMinderStateEngine
-from MinderStateEngine                 import MinderStateEngine
-from ErrorMinderStateEngine            import ErrorMinderStateEngine
-from BatchMinderStateEngine            import BatchMinderStateEngine
-from ContainerMinderStateEngine        import ContainerMinderStateEngine
-from WorkerMinderStateEngine           import WorkerMinderStateEngine
-
-from exc2string2                       import exc2string2
-from RTTSException                     import RTTCodingError
-
-class MinderStateEngineFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractiveJobMinder': LinuxInteractiveMinderStateEngine,
-      'Minder':                    MinderStateEngine,         # used by WatcherJobMinder through this base class
-      'ErrorJobMinder':            ErrorMinderStateEngine,
-      'BatchJobMinder':            BatchMinderStateEngine, # used by LSFBatchJobMinder through this base class
-      'LSFBatchJobMinder':         BatchMinderStateEngine, # used by LSFBatchJobMinder through this base class
-      'ContainerMinder':           ContainerMinderStateEngine,
-      'SequentialMinder':          ContainerMinderStateEngine,
-      'ParallelMinder':            ContainerMinderStateEngine,
-      'ChainJobMinder':            ContainerMinderStateEngine,
-      'PseudoJobMinder':           MinderStateEngine,
-      'WorkerMinder':              WorkerMinderStateEngine
-      }
-
-  def create(self, minder, state='queued'): 
-
-    key = minder.__class__.__name__
-
-    try:
-      engine =  self.dict[key](minder, state)
-    except:
-      msg = 'MinderStateEngineFactory: Cannot create a State engine for minder type %s:\Traceback:\n%s' % (key, exc2string2())
-      print msg
-      raise RTTCodingError(msg)
-
-    return engine
diff --git a/Tools/RunTimeTester/src/Factory_PathNameBuilder.py b/Tools/RunTimeTester/src/Factory_PathNameBuilder.py
deleted file mode 100755
index 9bd118ea8a7adab3b64db3a175e12fa057f7cb99..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_PathNameBuilder.py
+++ /dev/null
@@ -1,42 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass    import Factory
-from PathNameBuilder      import PathNameBuilder
-from RTTSException        import RTTCodingError
-from exc2string2          import exc2string2
-
-class PathNameBuilderFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'project': PathNameBuilder
-      }
-
-    self.logger = logger
-    
-  def create(self, mode, argBag, hardCoded, legals, siteSpecificsObj): 
-
-    try:
-      pnb = self.dict[mode](argBag,
-                            hardCoded,
-                            legals,
-                            siteSpecificsObj,
-                            self.logger)
-    except KeyError:
-      msg = '%s unknown key %s, could not instantiate PathNameBuilder: '% (self.__class__.__name__,str(mode))
-      self.logger.error(msg)
-      raise RTTCodingError(msg)
-    except Exception, e:
-      msg = '%s Error constructing PathNameBuilder from key  %s.\nException: %s\nTraceback:\n%s '% (self.__class__.__name__,
-                                                                                                    str(mode),
-                                                                                                    str(e),
-                                                                                                    exc2string2()
-                                                                                                    )
-      self.logger.error(msg)
-      raise RTTCodingError(msg)
-    else:
-      return pnb
- 
-
-
diff --git a/Tools/RunTimeTester/src/Factory_RunTimer.py.obsolete b/Tools/RunTimeTester/src/Factory_RunTimer.py.obsolete
deleted file mode 100755
index 859b6c536fe3026766f418896e2c38a9a29f1a93..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_RunTimer.py.obsolete
+++ /dev/null
@@ -1,30 +0,0 @@
-from Factory_BaseClass import Factory
-from RunTimer          import N0XNightlyRunTimer, NX0NightlyRunTimer
-from RunTimer          import NumberedRelRunTimer
-# -------------------------------------------------------------------------
-
-class RunTimerFactory(Factory):
-    def __init__(self, logger):
-        Factory.__init__(self, logger, self.__class__.__name__)
-        self.dict = {'bugfix':    N0XNightlyRunTimer,
-                     '10.0.X':    N0XNightlyRunTimer,
-                     'dev':       NX0NightlyRunTimer,
-                     'devval':    NX0NightlyRunTimer,
-                     'devmig6':   NX0NightlyRunTimer,
-                     'mig6':      NX0NightlyRunTimer,
-                     '13.2.0.Y':  NX0NightlyRunTimer,
-                     '14.0.0.Y':  NX0NightlyRunTimer,
-                     '14.0.10.Y': NX0NightlyRunTimer,
-                     '14.1.0.Y':  N0XNightlyRunTimer,
-                     '14.2.0.Y':  N0XNightlyRunTimer,
-                     'val':       NX0NightlyRunTimer,
-                     'pcache':    NX0NightlyRunTimer,
-                     'Numbered':  NumberedRelRunTimer}
-
-    
-    def create(self, isNightlyRelease,branch):
-        if isNightlyRelease:
-            return self.dict.get(branch, NX0NightlyRunTimer)(self.logger)
-    
-        return self.dict['Numbered'](self.logger)
-
diff --git a/Tools/RunTimeTester/src/Factory_ScriptWriter.py b/Tools/RunTimeTester/src/Factory_ScriptWriter.py
deleted file mode 100755
index 9d347e05eaa0c9a6d3b2352525fd76f7cf470f08..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_ScriptWriter.py
+++ /dev/null
@@ -1,373 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from Factory_BaseClass    import Factory
-from ScriptWriter         import *
-from RTTSException        import RTTCodingError
-
-# -------------------------------------------------------------------------
-# ScriptWriter Factory
-# -------------------------------------------------------------------------
-
-class ScriptWriterFactory(Factory):
-  def __init__(self, argsForScripts, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    self.argsForScripts = argsForScripts
-    
-    self.dict = {
-      'ContainerGroup'                                                                           : NullScriptWriter,
-      'AthenaJunkJob'                                                                            : JunkScriptWriter,
-      'Athena'                                                                                   : AthenaScriptWriter,
-      'AthenaAtlfastWatcher'                                                                     : NullScriptWriter,
-      'AthenaegammaRecWatcher'                                                                   : NullScriptWriter,
-      'Monitoring'                                                                               : MonitoringScriptWriter,
-      'AthenaATN'                                                                                : AthenaScriptWriter,
-      'AthenaDCubeClient'                                                                        : AthenaScriptWriter,
-      'AthenaCoolDozer'                                                                          : AthenaScriptWriter,
-      'AthenaHelloWorld'                                                                         : AthenaScriptWriter,
-      'AthenaAtlfast'                                                                            : AthenaScriptWriter,
-      'AthenaAtlfastPredicateUnitTest'                                                           : AthenaScriptWriter,
-      'AthenaFastShower'                                                                         : AthenaScriptWriter,
-      'AthenaGenerateFastSim'                                                                    : AthenaScriptWriter,
-      'AthenaCscClusterValidation_misal1_csc11.005145.PythiaZmumu.digit.RDO.v12000601_tid006791' : RecExCommonScriptWriter,
-      'AthenaCscSegmentValidation_misal1_csc11.005145.PythiaZmumu.digit.RDO.v12000601_tid006791' : RecExCommonScriptWriter,
-      'AthenaJiveXML'                                                                            : JiveXMLScriptWriter,
-      'AthenaTriggerRelease'                                                                     : TriggerReleaseScriptWriter,
-      'AthenaTriggerReleaseBS'                                                                   : TriggerReleaseScriptWriter,
-      'AthenaTrigP1TestMonitorHistos'                                                            : TriggerReleaseScriptWriter,
-      'AthenaTrigMenuValidation'                                                                 : TriggerReleaseScriptWriter,
-      'AthenaTriggerTestBS'                                                                      : TriggerReleaseScriptWriter,
-      'AthenaTriggerTestReco'                                                                    : TriggerReleaseScriptWriter,
-      'AthenaTriggerTestMonitorHistos'                                                           : TriggerReleaseScriptWriter,
-      'AthenaTrigAnalysisTestIOTests'                                                            : TriggerReleaseScriptWriter,
-      'AthenaTriggerReleaseReco'                                                                 : TriggerReleaseScriptWriter,
-      'AthenaL2Time1E31MC'                                                                       : HLTScriptWriter,
-      'AthenaEFTime1E31MC'                                                                       : HLTScriptWriter,
-      'AthenaTriggerReleaseID'                                                                   : TrigReleaseIDScriptWriter,
-      'AthenaTriggerReleaseCalo'                                                                 : TrigReleaseCaloScriptWriter,
-      'AthenaBackTrack_K0S_Et10'                                                                 : RecExCommonScriptWriter,
-      'AthenaBackTrack_ttbar'                                                                    : RecExCommonScriptWriter,
-      'AthenaBackTrack_Gamma_Conv_Et20'                                                          : RecExCommonScriptWriter,
-      'AthenaRecExCommon'                                                                        : RecExCommonScriptWriter,
-      'AthenaTestingRecExCommission'                                                             : RecExCommissionScriptWriter,
-      'AthenaTestingRecExCommissionSim'                                                          : RecExCommissionSimScriptWriter,
-      'AthenaPrimaryVertexingLowLumi'                                                            : RecExCommonScriptWriter,
-      'AthenaConversionVertexingLowLumi'                                                         : RecExCommonScriptWriter,
-      'AthenaConversionVertexing'                                                                : RecExCommonScriptWriter,
-      'AthenaAtlfastAlgs'                                                                        : RecExCommonScriptWriter,
-      'AthenaPileupDigiTest'                                                                     : RecExCommonScriptWriter,
-      'AthenaAMAValidationRTT-ttbar-mc08-geo-02-01-00'                                           : RecExCommonScriptWriter,
-      'AthenaPileupOverlay'                                                                      : RecExCommonScriptWriter,
-      'AthenaTrigBjetValidation'                                                                 : RecExCommonScriptWriter,
-      'AthenaHIValidation'                                                                       : RecExCommonScriptWriter,
-      'AthenaegammaRec'                                                                          : RecExCommonScriptWriter,
-      'AthenaegammaPerformance'                                                                  : RecExCommonScriptWriter,
-      'AthenaJetRec_J5'                                                                          : RecExCommonScriptWriter,
-      'AthenaJetRec_Pythia'                                                                      : RecExCommonScriptWriter,
-      'AthenaJetRec_PythiaFixedHardScatter'                                                      : RecExCommonScriptWriter,
-      'AthenaJetRec_ESD_to_CBNT'                                                                 : RecExCommonScriptWriter,
-      'AthenaJetRec_SinglePion'                                                                  : RecExCommonScriptWriter,
-      'AthenaSUSYValidation'                                                                     : RecExCommonScriptWriter,
-      'AthenaSCTDigiRTT'                                                                         : RecExCommonScriptWriter,
-      'AthenaH4lAnalysis'                                                                        : RecExCommonScriptWriter,
-      'AthenaIDAlignMonRTT'                                                                      : RecExCommonScriptWriter,
-      'AthenaIDPerfMonRTT'                                                                       : RecExCommonScriptWriter,
-      'AthenaIDAlignMon_Zmumu_perfect'                                                           : RecExCommonScriptWriter,
-      'AthenaIDAlignMon_Zmumu_aligned'                                                           : RecExCommonScriptWriter,
-      'AthenaIDAlignMon_Zmumu_misaligned'                                                        : RecExCommonScriptWriter,
-      'AthenaInDetRecValidation'                                                                 : RecExCommonScriptWriter,
-      'AthenaInDetRecValidation_Rome-Initial'                                                    : RecExCommonScriptWriter,
-      'AthenaInDetRecValidation_DC3'                                                             : RecExCommonScriptWriter,
-      'AthenaPixelTrackValidation'                                                               : RecExCommonScriptWriter,
-      'AthenaPixelDigitizationValidation'                                                        : RecExCommonScriptWriter,
-      'AthenaPixelTrackValidation_ttbar'                                                         : RecExCommonScriptWriter,
-      'AthenaPixelTrackValidation_mu100'                                                         : RecExCommonScriptWriter,
-      'AthenaPixelTrackValidation_mu3'                                                           : RecExCommonScriptWriter,
-      'AthenaMuonDigiExample'                                                                    : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance'                                                                  : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance-SingleMuonsPt10'                                                  : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance-SingleMuonsPt100'                                                 : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance-Zmumu'                                                            : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance-T1'                                                               : RecExCommonScriptWriter,
-      'AthenaMuGirlPerformance-directJpsi'                                                       : RecExCommonScriptWriter,   
-      'AthenaMuonEvtValidator'                                                                   : RecExCommonScriptWriter,
-      'AthenaMuonGeomRTT-HitRelocation'                                                          : RecExCommonScriptWriter,
-      'AthenaMuonGeomRTT-MuonGMCheck'                                                            : RecExCommonScriptWriter,
-      'AthenaMuonGeomRTT-AmdcMGM'                                                                : RecExCommonScriptWriter,
-      'AthenaTrigEgammaValidation'                                                               : RecExCommonScriptWriter,
-      'AthenaTrigEgammaNewRingerValidation'                                                      : RecExCommonScriptWriter,
-      'AthenaTrigEgammaNewValidation'                                                            : RecExCommonScriptWriter,
-      'AthenaTrigEgammaNewNavValidation'                                                         : RecExCommonScriptWriter,
-      'AthenaTrigEgammaNewValidationMTPT'                                                        : RecExCommonScriptWriter,
-      'AthenaInDetTrigRecExample'                                                                : RecExCommonScriptWriter,
-      'AthenaMuonHitTest'                                                                        : RecExCommonScriptWriter,
-      'AthenaMuonDigitTest'                                                                      : RecExCommonScriptWriter,
-      'AthenatauRec'                                                                             : RecExCommonScriptWriter,
-      'AthenaTauValidation'                                                                      : RecExCommonScriptWriter,
-      'AthenaTrigTauPerformanceTest'                                                             : RecExCommonScriptWriter,
-      'AthenaMuonRecValidator'                                                                   : RecExCommonScriptWriter,
-      'AthenaMuonRecValidator-Zmumu-v13002003'                                                   : RecExCommonScriptWriter,
-      'AthenaMuonInSituPerformance'                                                              : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-singleMu500gev-csc-01-00-00'                            : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-singleMu100gev-csc-01-00-00'                            : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Momu-singleMu500gev-csc-01-00-00'                            : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Momu-singleMu100gev-csc-01-00-00'                            : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-zmumu-sf05-csc-01-02-00'                                : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Momu-zmumu-sf05-csc-01-02-00'                                : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-Jpsi-csc-01-02-00'                                      : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-singleMu100gev-geo-02-00-00'                            : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Mboy-ttbar-geo-02-01-00'                                     : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Momu-zmumu-sf05-csc-01-02-00'                                : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Moore-Jpsi-csc-01-02-00'                                     : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Moore-singleMu100gev-geo-02-00-00'                           : RecExCommonScriptWriter,
-      'AthenaMuonHolesValidationRTT-Moore-ttbar-geo-02-01-00'                                    : RecExCommonScriptWriter,
-      'AthenaMooPerformance'                                                                     : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-geo-02-01-00'                                            : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-hyb-geo-02-01-00'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-TGmat-geo-02-01-00'                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt100-geo-02-01-00'                                           : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-Zprime-geo-02-01-00'                                            : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-Jpsi-geo-02-01-00'                                              : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-zmumu-geo-02-01-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-ttbar-geo-02-01-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuOnly-mupt100-geo-02-01-00'                                       : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-h4lep-geo-02-01-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-cosmics-comm-09-00-00'                                          : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-hyb-csc-01-02-00'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-TGmat-csc-01-02-00'                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt100-geo-02-00-00'                                           : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-segmreco-zmumu-csc-01-02-00-sf05'                               : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt10-dc3-02'                                                       : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt100-dc3-02'                                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt300-dc3-02'                                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-ttbar-geo-02-00-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-geo-02-00-00'                                            : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-hyb-geo-02-00-00'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-TGmat-geo-02-00-00'                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuOnly-mupt100-geo-02-00-00'                                       : RecExCommonScriptWriter,
-      'AthenaMooPerformance-zmumu-dc3-02'                                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-ttbar-csc-01-02-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuOnly-mupt100-csc-01-02-00'                                       : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuOnly-zmumu-csc-01-02-00-sf05'                                    : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-Zprime-csc-01-02-00'                                            : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-Jpsi-csc-01-02-00'                                              : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt10-csc-01-02-00'                                            : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt100-csc-01-02-00'                                           : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-mupt500-csc-01-02-00'                                           : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-zmumu-csc-01-02-00'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMu-zmumu-csc-01-02-00-sf05'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MooreOnly-zmumu-csc-01-02-00'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMooreOnly-zmumu-csc-01-02-00'                                    : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMoore-mupt10-csc-01-02-00'                                       : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMoore-mupt100-csc-01-02-00'                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMoore-mupt500-csc-01-02-00'                                      : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMoore-zmumu-csc-01-02-00'                                        : RecExCommonScriptWriter,
-      'AthenaMooPerformance-MoMuMoore-zmumu-csc-01-02-00-sf05'                                   : RecExCommonScriptWriter,
-      'AthenaMooPerformance-zmumu-csc-01-02-00-sf05'                                             : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt10-csc-01-02-00'                                                 : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt100-csc-01-02-00'                                                : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt500-csc-01-02-00'                                                : RecExCommonScriptWriter,
-      'AthenaMooPerformance-zmumu-csc-01-02-00'                                                  : RecExCommonScriptWriter,
-      'AthenaMooPerformance-mupt100-csc-01-00-00'                                                : RecExCommonScriptWriter,
-      'AthenaMooPerformance-zmumu-csc-01-02-00'                                                  : RecExCommonScriptWriter,
-      'AthenaMuonRecRTT-zmumu-csc-01-02-00'                                                      : RecExCommonScriptWriter,
-      'AthenaMuonRecRTT-T1-csc-01-02-00'                                                         : RecExCommonScriptWriter,
-      'AthenaMuonRecRTT'                                                                         : RecExCommonScriptWriter,
-      'AthenaMuonDataRec'                                                                        : RecExCommissionScriptWriter,
-      'AthenaMuonDataRec_mon'                                                                    : RecExCommissionScriptWriter,
-      'AthenaMuonSimRec'                                                                         : RecExCommissionScriptWriter,
-      'AthenaMuonSimRec_mon'                                                                     : RecExCommissionScriptWriter,
-      'AthenaMuonCommRec'                                                                        : RecExCommissionScriptWriter,
-      'AthenaMuonCommRec_mon'                                                                    : RecExCommissionScriptWriter,
-      'AthenaMuonCommRecSim'                                                                     : RecExCommonScriptWriter,
-      'AthenaMuonCommRecSim_mon'                                                                 : RecExCommonScriptWriter,
-      'AthenaCaloRecEx'                                                                          : RecExCommonScriptWriter,
-      'AthenaGammaE50'                                                                           : RecExCommonScriptWriter,
-      'AthenaGammaE500'                                                                          : RecExCommonScriptWriter,
-      'AthenaGammaE5'                                                                            : RecExCommonScriptWriter,
-      'AthenaElectronE5'                                                                         : RecExCommonScriptWriter,
-      'AthenaElectronSliceDNAKalman'                                                             : RecExCommonScriptWriter,
-      'AthenaEle5IDPRTTKalmanDNA'                                                                : RecExCommonScriptWriter,
-      'AthenaEle60IDPRTTKalmanDNA'                                                               : RecExCommonScriptWriter,
-      'AthenaEle5IDPRTTGaussianSumFilter'                                                        : RecExCommonScriptWriter,
-      'AthenaEle60IDPRTTGaussianSumFilter'                                                       : RecExCommonScriptWriter,      
-      'AthenaTopJimmy'                                                                           : RecExCommonScriptWriter,
-      'AthenaElectronE10'                                                                        : RecExCommonScriptWriter,
-      'AthenaCaloTests'                                                                          : RecExCommonScriptWriter,
-      'AthenaCaloTestsCTB'                                                                       : RecExCommonScriptWriter,
-      'AthenaCaloTestsAANT'                                                                      : RecExCommonScriptWriter,
-      'AthenaCaloAnaExTestBeam'                                                                  : RecExCommonScriptWriter,
-      'AthenaCaloAnaExESDAOD'                                                                    : RecExCommonScriptWriter,
-      'AthenaCaloDigEx'                                                                          : RecExCommonScriptWriter,
-      'AthenaCaloSimEx'                                                                          : RecExCommonScriptWriter,
-      'AthenaCaloCosEx'                                                                          : RecExCommonScriptWriter,
-      'AthenaCaloRec'                                                                            : RecExCommonScriptWriter,
-      'AthenaCaloRecExAANT'                                                                      : RecExCommonScriptWriter,
-      'AthenaCaloRecExAANTReal'                                                                  : CaloRecExAANTRealScriptWriter,
-      'AthenaBPhysValidation'                                                                    : RecExCommonScriptWriter,
-      'AthenaBPhysTrigger'                                                                       : RecExCommonScriptWriter,
-      'AthenaSimDig'                                                                             : RecExCommonScriptWriter,
-      'AthenaBPhysReco_iPatRec'                                                                  : RecExCommonScriptWriter,
-      'AthenaBPhysReco_xKalman'                                                                  : RecExCommonScriptWriter,
-      'AthenaBPhysReco_newTrk'                                                                   : RecExCommonScriptWriter,
-      'AthenaBPhysTrackValidation'                                                               : RecExCommonScriptWriter,
-      'AthenaBPhysTrackValidation_DC3'                                                           : RecExCommonScriptWriter,
-      'AthenaAnalysisExamples'                                                                   : RecExCommonScriptWriter,
-      'AthenaJetTagRTT'                                                                          : RecExCommonScriptWriter,
-      'AthenaDigitization'                                                                       : RecExCommonScriptWriter,
-      'AthenaG4AtlasApps'                                                                        : AthenaScriptWriter,
-      'AthenaG4AtlasAppsRelease'                                                                 : RecExCommonScriptWriter,
-      'AthenaGeneratorsRTT'                                                                      : RecExCommonScriptWriter,
-      'AthenaHiggsToGamGam'                                                                      : RecExCommonScriptWriter,
-      'AthenaQuickGenCheck'                                                                      : RecExCommonScriptWriter,
-      'AthenaInDetRTT'                                                                           : RecExCommonScriptWriter,
-      'AthenaV0FindingLowLumi'                                                                   : RecExCommonScriptWriter,
-      'AthenaPi1IDPRTT'                                                                          : RecExCommonScriptWriter,
-      'AthenaPi100IDPRTT'                                                                        : RecExCommonScriptWriter,
-      'AthenaEle1IDPRTT'                                                                         : RecExCommonScriptWriter,
-      'AthenaEle100IDPRTT'                                                                       : RecExCommonScriptWriter,      
-      'AthenaMu1IDPRTT'                                                                          : RecExCommonScriptWriter,
-      'AthenaMu1IDPRTTKalman'                                                                    : RecExCommonScriptWriter,
-      'AthenaMu100IDPRTT'                                                                        : RecExCommonScriptWriter,
-      'AthenaZtomumuIDPRTT'                                                                      : RecExCommonScriptWriter,
-      'AthenaZtomumuIDPRTTKalman'                                                                : RecExCommonScriptWriter,
-      'AthenaMu100IDPRTTKalman'                                                                  : RecExCommonScriptWriter,
-      'AthenaMu1GlobalChiSq'                                                                     : RecExCommonScriptWriter,
-      'AthenaMu100GlobalChiSq'                                                                   : RecExCommonScriptWriter,
-      'AthenaMu1Kalman'                                                                          : RecExCommonScriptWriter,
-      'AthenaMu100Kalman'                                                                        : RecExCommonScriptWriter,
-      'AthenaPi1'                                                                                : RecExCommonScriptWriter,
-      'AthenaPi100'                                                                              : RecExCommonScriptWriter,
-      'AthenaLowPt'                                                                              : RecExCommonScriptWriter,
-      'AthenaLowPtKalman'                                                                        : RecExCommonScriptWriter,
-      'AthenaInDetSimRTT'                                                                        : RecExCommonScriptWriter,      
-      'AthenaPi1IDPRTTKalman'                                                                    : RecExCommonScriptWriter,
-      'AthenaPi100IDPRTTKalman'                                                                  : RecExCommonScriptWriter,
-      'AthenaEle1IDPRTTKalman'                                                                   : RecExCommonScriptWriter,
-      'AthenaEle100IDPRTTKalman'                                                                 : RecExCommonScriptWriter,
-      'AthenaTtbarIDPRTT'                                                                        : RecExCommonScriptWriter,
-      'AthenaTtbarIDPRTTKalman'                                                                  : RecExCommonScriptWriter,
-      'AthenaZtomumuPileupIDPRTT'                                                                : RecExCommonScriptWriter,
-      'AthenaZtomumuPileupIDPRTTKalman'                                                          : RecExCommonScriptWriter,      
-      'AthenaSingleParticle'                                                                     : RecExCommonScriptWriter,
-      'AthenaSingleE'                                                                            : RecExCommonScriptWriter,
-      'AthenaSingleMu'                                                                           : RecExCommonScriptWriter,
-      'AthenaSingleMuKalman'                                                                     : RecExCommonScriptWriter,
-      'AthenaSinglePi'                                                                           : RecExCommonScriptWriter,
-      'AthenaExoticsPhysValJet'                                                                  : RecExCommonScriptWriter,
-      'AthenaMinBias'                                                                            : RecExCommonScriptWriter,
-      'AthenaMinBiasIDPRTTKalman'                                                                : RecExCommonScriptWriter,
-      'AthenaMinBiasIDPRTT'                                                                      : RecExCommonScriptWriter,
-      'AthenaMinBiasTrig'                                                                        : RecExCommonScriptWriter,
-      'AthenaMinbiasValidation'                                                                  : RecExCommonScriptWriter,
-      'AthenaZee'                                                                                : RecExCommonScriptWriter,
-      'AthenaZmumu'                                                                              : RecExCommonScriptWriter,
-      'AthenaZmumuKalman'                                                                        : RecExCommonScriptWriter,
-      'AthenaTop'                                                                                : RecExCommonScriptWriter,
-      'AthenaTopRTT'                                                                             : RecExCommonScriptWriter,
-      'AthenaJ5'                                                                                 : RecExCommonScriptWriter,
-      'AthenaJ8'                                                                                 : RecExCommonScriptWriter,
-      'AthenaHighPtMu'                                                                           : RecExCommonScriptWriter,
-      'AthenaHighPtMuKalman'                                                                     : RecExCommonScriptWriter,
-      'AthenaHiggsToGamGam'                                                                      : RecExCommonScriptWriter,
-      'AthenaHighPtMuMisaligned'                                                                 : RecExCommonScriptWriter,
-      'AthenaZmumuMisaligned'                                                                    : RecExCommonScriptWriter,
-      'AthenaSingleMuMisaligned'                                                                 : RecExCommonScriptWriter,
-      'AthenaMissingET'                                                                          : RecExCommonScriptWriter,
-      'AthenaTRT_RTT'                                                                            : RecExCommonScriptWriter,
-      'AthenaTRTstandalone_Gamma_Conv_Et20'                                                      : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_BS'                                                                         : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_Monitoring'                                                                 : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_minus'                                                                    : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_plus'                                                                     : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_minus'                                                                   : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_plus'                                                                    : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_minus'                                                                   : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_plus'                                                                    : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_minus_DIG'                                                                : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_plus_DIG'                                                                 : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_minus_DIG'                                                               : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_plus_DIG'                                                                : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_minus_DIG'                                                               : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_plus_DIG'                                                                : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_minus_SIM'                                                                : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_e_plus_SIM'                                                                 : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_minus_SIM'                                                               : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_mu_plus_SIM'                                                                : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_minus_SIM'                                                               : RecExCommonScriptWriter,
-      'AthenaTRT_RTT_pi_plus_SIM'                                                                : RecExCommonScriptWriter,
-      'AthenaLArMonTools'                                                                        : RecExCommonScriptWriter,
-      'AthenaPhysicsAnalysisTools'                                                               : RecExCommonScriptWriter,
-      'AthenaElectronValidation'                                                                 : RecExCommonScriptWriter,
-      'AthenaMuonboyAth'                                                                         : RecExCommonScriptWriter,
-      'AthenaMboyPerformance'                                                                    : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt100-csc-01-00-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt100-csc-01-02-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt10-csc-01-02-00'                                                : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt500-csc-01-02-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-zmumu-csc-01-02-00'                                                 : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-zmumu-csc-01-02-00-sf05'                                            : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt5-geo-02-00-00'                                                 : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt10-geo-02-00-00'                                                : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt100-geo-02-00-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt500-geo-02-00-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-mupt100-csc-01-02-00'                                               : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-zmumu-csc-01-02-00'                                                 : RecExCommonScriptWriter,
-      'AthenaMboyPerformance-zmumu-csc-01-02-00-sf05'                                            : RecExCommonScriptWriter,
-      'AthenaTauRec'                                                                             : RecExCommonScriptWriter,
-      'AthenaTauDPDMaker'                                                                        : RecExCommonScriptWriter,
-      'AthenaTrigInDetValidation'                                                                : RecExCommonScriptWriter,
-      'AthenaTestingTileRecEx'                                                                   : TileRecExScriptWriter,
-      'AthenaTestingTileSimEx'                                                                   : TileSimExScriptWriter,
-      'AthenaAtlasHLT'                                                                           : HLTScriptWriter,
-      'AthenaAtlasTrigConf'                                                                      : HLTTrigConfScriptWriter,
-      'AthenaAtlasTrigConfPTOnly'                                                                : HLTTrigConfScriptWriter,
-      'AthenaAtlasTrigConfMTOnly'                                                                : HLTTrigConfScriptWriter,
-      'AthenaAtlasTrigConfMC'                                                                    : HLTTrigConfScriptWriter,
-      'JobTransform'                                                                             : JobTransformScriptWriter,
-      'SimuJobTransform'                                                                         : JobTransformScriptWriter,
-      'PATJobTransform'                                                                          : JobTransformScriptWriter,
-      'TCTTransform'                                                                             : JobTransformScriptWriter,
-      'FCTTransform'                                                                             : JobTransformScriptWriter,
-      'OnlineRecoTests'                                                                          : JobTransformScriptWriter,
-      'Digitization_JobTransforms'                                                               : JobTransformScriptWriter,
-      'GeneratorJobTransforms'                                                                   : JobTransformScriptWriter,
-      'AtlfastJobTransforms'                                                                     : JobTransformScriptWriter,
-      'Valgrind_JobTransforms'                                                                   : JobTransformScriptWriter,
-      'Overlay_JobTransforms'                                                                    : JobTransformScriptWriter,
-      'OverlayAOD_JobTransforms'                                                                 : JobTransformScriptWriter,
-      'OverlayAODMC_JobTransforms'                                                               : JobTransformScriptWriter,
-      'OverlayReco_JobTransforms'                                                                : JobTransformScriptWriter,
-      'OverlayRecoMC_JobTransforms'                                                              : JobTransformScriptWriter,
-      'OverlayBS_JobTransforms'                                                                  : JobTransformScriptWriter,
-      'OverlayPOOL_JobTransforms'                                                                : JobTransformScriptWriter,
-      'RecTrf'                                                                                   : JobTransformScriptWriter,
-      'TrfInDetTrigRecExample'                                                                   : JobTransformScriptWriter,
-      'InDetSiLevel1Alignment'                                                                   : JobTransformScriptWriter
-      }    
-    
-  def create(self, descArgsForScripts, logger):
-    'Returns a  ScriptWriter instance according to jobGroup'
-    jobGroup = descArgsForScripts.jobGroup
-    if jobGroup not in self.dict.keys():
-      msg  = 'Error instantiating Script: unknown job jobGroup: %s ' % jobGroup
-      self.logger.error(msg)
-      print msg
-      raise RTTCodingError(msg)
-
-    argsForScripts = descArgsForScripts
-    argsForScripts.__dict__.update(self.argsForScripts.__dict__)
-
-    try:
-      script = self.dict[jobGroup](argsForScripts, logger)
-    except Exception, e:
-      m = 'Error constricting a runscript:\nException:\n%s\nTraceback:\n%s' % (str(e), exc2string2())
-      logger.error(m)
-      raise RTTCodingError(m)
-    
-    return script
-
diff --git a/Tools/RunTimeTester/src/Factory_ScriptWriterGrid.py.obsolete b/Tools/RunTimeTester/src/Factory_ScriptWriterGrid.py.obsolete
deleted file mode 100755
index 6381cfb8b71d4d3be79fa6730d37e0204fcc13df..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_ScriptWriterGrid.py.obsolete
+++ /dev/null
@@ -1,94 +0,0 @@
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from Factory_BaseClass    import Factory
-from ScriptWriterGrid     import *
-
-
-import sys
-
-
-# -------------------------------------------------------------------------
-# ScriptWriter Factory
-# -------------------------------------------------------------------------
-class ScriptWriterGridFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-    self.dict = {
-      'AthenaHelloWorld':         AthenaScriptWriterGrid,
-      'AthenaRecExCommon':        RecExCommonScriptWriterGrid,
-      'AthenaAtlfast':            AthenaScriptWriterGrid,
-      'AthenaAtlfastAlgs':        RecExCommonScriptWriterGrid,        
-      'AthenaJiveXML':            JiveXMLScriptWriterGrid,
-      'AthenaTriggerRelease':     TriggerReleaseScriptWriterGrid,
-      'AthenaTriggerReleaseID':   TrigReleaseIDScriptWriterGrid,
-      'AthenaTriggerReleaseCalo': TrigReleaseCaloScriptWriterGrid,
-      'AthenaTriggerReleaseBS':   TriggerReleaseScriptWriterGrid,
-      'AthenaTriggerReleaseReco': TriggerReleaseScriptWriterGrid,      
-      'AthenaegammaRec':          RecExCommonScriptWriterGrid,
-      'RuleChecker':              RuleCheckerScriptWriterGrid,
-      'QAmetrics':                QAmetricsScriptWriterGrid,
-      'AthenaAtlfastWatcher':     NullScriptWriterGrid,
-      'AthenaegammaRecWatcher':   NullScriptWriterGrid,
-      'RuleCheckerWatcher':       NullScriptWriterGrid,
-      'AthenaInDetRecValidation': RecExCommonScriptWriterGrid,
-      'AthenaMuonDigiExample':    RecExCommonScriptWriterGrid,
-      'KitValidation':            KitValidationScriptWriterGrid,
-      'AthenaMooPerformance':     RecExCommonScriptWriterGrid,
-      'AthenaCaloRecEx':          RecExCommonScriptWriterGrid,
-      'AthenaCaloDigEx':          RecExCommonScriptWriterGrid,
-      'AthenaCaloSimEx':          RecExCommonScriptWriterGrid,
-      'AthenaCaloRec':            RecExCommonScriptWriterGrid,
-      'AthenaBPhysValidation':    RecExCommonScriptWriterGrid,
-      'AthenaAnalysisExamples':   RecExCommonScriptWriterGrid,
-      'AthenaDigitization':       RecExCommonScriptWriterGrid,
-      'AthenaG4AtlasApps':        RecExCommonScriptWriterGrid,
-      'AthenaG4AtlasAppsRelease': RecExCommonScriptWriterGrid,
-      'AthenaGeneratorsRTT':      RecExCommonScriptWriterGrid,
-      'AthenaInDetRTT':           RecExCommonScriptWriterGrid,      
-      'AthenaSingleParticle':     RecExCommonScriptWriterGrid,
-      'AthenaSingleE':            RecExCommonScriptWriterGrid,
-      'AthenaSingleMu':           RecExCommonScriptWriterGrid,
-      'AthenaSinglePi':           RecExCommonScriptWriterGrid,
-      'AthenaMinBias':            RecExCommonScriptWriterGrid,
-      'AthenaZee':                RecExCommonScriptWriterGrid,
-      'AthenaZmumu':              RecExCommonScriptWriterGrid,
-      'AthenaTop':                RecExCommonScriptWriterGrid,
-      'AthenaJ5':                 RecExCommonScriptWriterGrid,
-      'AthenaJ8':                 RecExCommonScriptWriterGrid,      
-      'AthenaMissingET':          RecExCommonScriptWriterGrid,
-      'AthenaPhysicsAnalysisTools': RecExCommonScriptWriterGrid,
-      'Monitoring':               MonitoringScriptWriterGrid,
-      'AthenaMuonboyAth':         RecExCommonScriptWriterGrid,
-      'AthenaMboyPerformance':    RecExCommonScriptWriterGrid,
-      'AthenatauRec':             RecExCommonScriptWriterGrid,
-      'AthenaInDetSimRTT':        RecExCommonScriptWriterGrid,
-      'AthenaLArMonTools':        RecExCommonScriptWriterGrid,
-      'AthenaRecExAnaTest':       RecExCommonScriptWriterGrid,
-      'AthenaRecExTrigTest':      RecExCommonScriptWriterGrid,
-      'AthenaQuickGenCheck':      RecExCommonScriptWriterGrid,
-      'AthenaSimDig':             RecExCommonScriptWriterGrid,
-      'AthenaCaloRecExESDAOD':    RecExCommonScriptWriterGrid,
-      'AthenaCaloRecExTestBeam':  RecExCommonScriptWriterGrid,
-      'AthenaHighPtMu':           RecExCommonScriptWriterGrid,
-      'RTTMonitor':               NullScriptWriterGrid
-      #'RTTSelfTest':              RTTSelfTestScriptWriterGrid
-      }
-    
-
-
-
-  def create(self, jDescriptor):
-    'Returns a  ScriptWriter instance according to jobGroup'
-    group = jDescriptor.jobGroup
-    if group in self.dict.keys():
-      return self.dict[group](jDescriptor)
-    else:
-      msg = 'Unknown mode, could not instantiate a ScriptWriter: '+group
-      self.logger.error(msg)
-      sys.exit(0)
-
-
diff --git a/Tools/RunTimeTester/src/Factory_SiteSpecifics.py b/Tools/RunTimeTester/src/Factory_SiteSpecifics.py
deleted file mode 100755
index e9cf59d8e68ca09e3057e21f450bb45024fb8111..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_SiteSpecifics.py
+++ /dev/null
@@ -1,32 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass    import Factory
-from  SiteSpecifics       import * 
-
-
-# -------------------------------------------------------------------------
-# SiteSpeocifics Factory
-# -------------------------------------------------------------------------
-
-class SiteSpecificsFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      "CERN"     : CERNSiteSpecifics,
-      "CERNslc4" : CERNslc4SiteSpecifics,
-      "PetersMac": PetersMacSiteSpecifics,
-      "UCLslc3"  : UCLslc3SiteSpecifics,
-      "Lancaster": LancasterSiteSpecifics
-      }
-
-
-
-  def  create(self, site):
-    """ provide an instance of a SiteSpecifics:
-You give the site of a group - and you get an object with
-details for that site."""
-      
-    return self.dict[site]()
-
-# -------------------------------------------------------------------------
diff --git a/Tools/RunTimeTester/src/Factory_StampObject.py.obsolete b/Tools/RunTimeTester/src/Factory_StampObject.py.obsolete
deleted file mode 100755
index c22346ef6374cb12be856147980dd9365bf177f4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_StampObject.py.obsolete
+++ /dev/null
@@ -1,29 +0,0 @@
-from Factory_BaseClass    import Factory
-from StampObject          import StampObject
-from StampObject          import StampObjectKit
-import sys
-
-class StampObjectFactory(Factory):
-    def __init__(self,logger):
-        Factory.__init__(self, logger, self.__class__.__name__)
-        
-        self.dict = {
-            'kit':           StampObjectKit,
-            'build':         StampObject
-            }
-        
-    def create(self, paramsForStampObject):
-        """
-        mode is kit or release.
-        it is overridden by local RTT runs
-        """
-        key = paramsForStampObject.runType # kit or build
-        
-        if key in self.dict.keys():
-            sto =  self.dict[key](paramsForStampObject)
-            self.logger.debug('Creating a %s' % sto.__class__.__name__)
-            return sto
-        else:
-            msg = 'Unknown release type, could not instantiate a StampObject: '+str(key)
-            self.logger.info(msg)
-            sys.exit(0)
diff --git a/Tools/RunTimeTester/src/Factory_TestRun.py.obsolete b/Tools/RunTimeTester/src/Factory_TestRun.py.obsolete
deleted file mode 100755
index 68cd099e705383e113c272c4ce8755965b9a42b6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_TestRun.py.obsolete
+++ /dev/null
@@ -1,42 +0,0 @@
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-
-from exc2string2          import exc2string2
-from Factory_BaseClass    import Factory
-from TestRun              import UnifiedConfigurationTestRun, FailureTestRun
-
-# -------------------------------------------------------------------------
-# TestRun Factory
-# -------------------------------------------------------------------------
-
-class TestRunFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-  def create(self, cvsPackage, packageName, paths, rttRunStartTime, closeDownKeyMaker, logger):
-    if not cvsPackage.runJobs:
-      self.logger.warning('Package %s instantaiting FailureTestRun due to RTT operator veto' % packageName)
-      return FailureTestRun(packageName, paths, rttRunStartTime, closeDownKeyMaker, logger, 'RTT operator veto', True)
-
-    if not cvsPackage.noError():
-      m = 'Package %s instantaiting FailureTestRun due to CVS package error: %s' % (packageName,
-                                                                                    cvsPackage.errorMessage)
-      self.logger.warning(m)
-      return FailureTestRun(packageName, paths, rttRunStartTime, closeDownKeyMaker, logger, cvsPackage.errorMessage, False)
-          
-    try:
-      tr = UnifiedConfigurationTestRun(packageName, paths, rttRunStartTime, closeDownKeyMaker, logger, '')
-    except Exception, e:
-      self.logger.warning('Failed to make a TestRun for package %s' % paths.package)
-      # eText = exc2string2()
-      self.logger.warning(exc2string2())
-      self.logger.warning('Replacing with a Failure TestRun')
-      text  = 'Package rejected!\n'
-      text += str(e)
-      return FailureTestRun(packageName, paths, rttRunStartTime, closeDownKeyMaker, logger, '', False)
-    else:
-      self.logger.debug('Instantiated a TestRun for package %s' % paths.package)
-      return tr
-
diff --git a/Tools/RunTimeTester/src/Factory_TestRunXMLConverter.py b/Tools/RunTimeTester/src/Factory_TestRunXMLConverter.py
deleted file mode 100644
index 0cde0faa5f1d136e97d89b52fa5673c88d7b6549..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_TestRunXMLConverter.py
+++ /dev/null
@@ -1,60 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This module bundles the RTT factory functions.
-These functions are written as classes to help write test code.
-"""
-import os.path
-
-from exc2string2          import exc2string2
-from Factory_BaseClass    import Factory
-from TestRunXMLConverter2 import TestRunXMLConverter
-from TestRunXMLConverter2 import FailureTestRunXMLConverter
-from RTTSException        import RTTCodingError
-
-# -------------------------------------------------------------------------
-# TestRun Factory
-# -------------------------------------------------------------------------
-
-class UnknownTestRunException: pass
-
-class TestRunXMLConverterFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-
-  def create(self,
-             testRun,
-             packageResultsDir,
-             packageName,
-             packageTag,
-             containerPackage,
-             callBackConverter,
-             elementCreator,
-             textNodeCreator,
-             logger):
-
-    trClassName = testRun.__class__.__name__
-    # outputFile = os.path.join(packageResultsDir, packageName, 'rttpackageinfo.xml')
-    outputFile = os.path.join(packageResultsDir, 'rttpackageinfo.xml')
-
-    if trClassName == 'UnifiedConfigurationTestRun':
-      try:
-        converter = TestRunXMLConverter(outputFile, packageName, packageTag, containerPackage,
-                                        elementCreator, textNodeCreator, callBackConverter, logger)
-      except Exception, e:
-        m = 'Error creating TestRun: key %s\nException:\n%s\nTraceback:\n%s' % (trClassName, str(e), exc2string2())
-        raise RTTCodingError(m)
-      
-    elif trClassName == 'FailureTestRun':
-      try:
-        converter = FailureTestRunXMLConverter(outputFile, packageName, packageTag, containerPackage,
-                                               elementCreator, textNodeCreator, callBackConverter, logger)
-      except Exception, e:
-        m = 'Error creating TestRun: key %s\nException:\n%s\nTraceback:\n%s' % (trClassName, str(e), exc2string2())
-        raise RTTCodingError(m)
-      
-    else:
-      raise RTTCodingError('UnknownTestRunException %s' % trClassName)
-    
-
-    return converter
diff --git a/Tools/RunTimeTester/src/Factory_UserStuffRetriever.py b/Tools/RunTimeTester/src/Factory_UserStuffRetriever.py
deleted file mode 100755
index 90b13df42abf516126a6b54af72b444b4d4e58fb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_UserStuffRetriever.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass          import Factory
-
-from UserStuffRetriever  import UserStuffRetrieverLocal
-from UserStuffRetriever  import UserStuffRetrieverNonLocal
-
-import sys
-
-class UserStuffRetrieverFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    #'monolithic':    UserStuffRetriever,
-    self.dict = {
-                 'release' :      UserStuffRetrieverNonLocal,
-                 'dev' :          UserStuffRetrieverNonLocal,
-                 'bugfix' :       UserStuffRetrieverNonLocal,
-                 'val':           UserStuffRetrieverNonLocal,
-                 'localPackages': UserStuffRetrieverLocal
-                 }
-
-  def makeKey(self, paths):
-    
-    if paths.localRTTRun:                     return 'localPackages'
-    if paths.release in ['11.3.0', '11.4.0']: return '11301140'
-    if paths.branch=='dev':                   return 'dev'
-    if paths.branch=='bugfix':                return 'bugfix'
-    if paths.branch=='val':                   return 'val'
-
-    return 'release'
-
-  def create(self, paths, argDict): 
-
-    key = self.makeKey(paths)
-        
-    if key not in self.dict.keys():
-      msg  = 'Unknown key, could not instantiate a UserStuffRetriever: '
-      msg += str(key)
-      self.logger.error(msg)
-      sys.exit(0)
-
-
-    usr =  self.dict[key](argDict)
-
-    self.logger.debug('Instantiated a USR of type %s' % usr.__class__.__name__)
-
-    return usr
-  
-
-
diff --git a/Tools/RunTimeTester/src/Factory_XMLConverter.py b/Tools/RunTimeTester/src/Factory_XMLConverter.py
deleted file mode 100644
index 1d590cfb99b594085ce1bb1b4645917b4d6bed53..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_XMLConverter.py
+++ /dev/null
@@ -1,46 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseClass             import Factory
-from MinderToXML2                  import MinderToXML
-from LinuxInteractiveMinderToXML2  import LinuxInteractiveMinderToXML
-from NullXMLConverter              import NullXMLConverter
-from ContainerXMLConverter         import ContainerXMLConverter
-from ErrorMinderToXML              import ErrorMinderToXML
-
-from RTTSException                 import RTTCodingError
-from exc2string2                   import exc2string2
-
-class XMLConverterFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractiveJobMinder': LinuxInteractiveMinderToXML,# eventually should be renamed and replace MinderToXML
-      'LSFBatchJobMinder':         MinderToXML,
-      'Minder':                    MinderToXML,
-      'ErrorJobMinder':            ErrorMinderToXML,
-      'BatchJobMinder':            MinderToXML,
-      'LSFBatchJobMinder':         MinderToXML,
-      'ContainerMinder':           ContainerXMLConverter,
-      'SequentialMinder':          ContainerXMLConverter,
-      'ParallelMinder':            ContainerXMLConverter,
-      'ChainJobMinder':            ContainerXMLConverter,
-      'PseudoJobMinder':           MinderToXML,
-      'WorkerMinder':              MinderToXML
-      }
-
-  def create(self, minder): 
-
-    key = minder.__class__.__name__
-
-    try:
-      converter =  self.dict[key](minder)
-    except Exception, e:
-      msg = 'MinderStateEngineFactory: Cannot create an XML converter for minder type %s Reason: %s\nTraceback\n%s ' % (key,
-                                                                                                                        str(e),
-                                                                                                                        exc2string2()
-                                                                                                                        )
-      print msg
-      raise RTTCodingError(msg)
-
-    return converter
diff --git a/Tools/RunTimeTester/src/Factory_XMLConverter.py.obsolete b/Tools/RunTimeTester/src/Factory_XMLConverter.py.obsolete
deleted file mode 100644
index b543d3f231356f86dbd6356ab1be700081c66e26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Factory_XMLConverter.py.obsolete
+++ /dev/null
@@ -1,43 +0,0 @@
-from Factory_BaseClass         import Factory
-from MinderToXML2              import MinderToXML
-from NullXMLConverter          import NullXMLConverter
-from ContainerXMLConverter     import ContainerXMLConverter
-from ErrorMinderToXML          import ErrorMinderToXML
-
-from RTTSException             import RTTCodingError
-from exc2string2               import exc2string2
-
-class XMLConverterFactory(Factory):
-  def __init__(self, logger):
-    Factory.__init__(self, logger, self.__class__.__name__)
-    
-    self.dict = {
-      'LinuxInteractiveJobMinder': MinderToXML,
-      'LSFBatchJobMinder':         MinderToXML,
-      'Minder':                    MinderToXML,
-      'ErrorJobMinder':            ErrorMinderToXML,
-      'BatchJobMinder':            MinderToXML,
-      'LSFBatchJobMinder':         MinderToXML,
-      'ContainerMinder':           ContainerXMLConverter,
-      'SequentialMinder':          ContainerXMLConverter,
-      'ParallelMinder':            ContainerXMLConverter,
-      'ChainJobMinder':            ContainerXMLConverter,
-      'PseudoJobMinder':           MinderToXML,
-      'WorkerMinder':              MinderToXML
-      }
-
-  def create(self, minder): 
-
-    key = minder.__class__.__name__
-
-    try:
-      converter =  self.dict[key](minder)
-    except Exception, e:
-      msg = 'MinderStateEngineFactory: Cannot create an XML converter for minder type %s Reason: %s\nTraceback\n%s ' % (key,
-                                                                                                                        str(e),
-                                                                                                                        exc2string2()
-                                                                                                                        )
-      print msg
-      raise RTTCodingError(msg)
-
-    return converter
diff --git a/Tools/RunTimeTester/src/FailureTestRun.py b/Tools/RunTimeTester/src/FailureTestRun.py
deleted file mode 100644
index 4b0e8b0b50f2090310e5c6d90249e628a0316061..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/FailureTestRun.py
+++ /dev/null
@@ -1,35 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-from MethodTimer              import MethodTimer
-from MethodTracer        import MetaTracer
-
-from TestRun import TestRun
-import threading
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-# -------------------------------------------------------------------------
-# -------------------------------------------------------------------------
-
-class FailureTestRun(TestRun):
-    'Created when a UnifiedTestRun cannot be created (due to some error).'
-    
-    # def __init__(self, packageName, paths, rttRunStartTime, closeDownKeyMaker, logger, failText, vetoFlag):
-    def __init__(self, packageName, closedownKey, rttRunStartTime, closeDownKeyMaker, logger, failText, vetoFlag):
-        # pass minders as an empty list
-        TestRun.__init__(self, [], packageName, closedownKey, closeDownKeyMaker, logger, failText)
-
-    # def groupsInTestRun(self):
-    #    return []
-
-
-
-# ------------------------------------------------------------------------- 
-
-    def run(self):
-        """split the jobs into various phases. Each phase runs to completion
-         before starting the next one."""
-        self.logger.info("FailureTestRun run method starts")
-        self.xmlConverter.pushData()
-        self.logger.info("FailureTestRun run method ends")
-
diff --git a/Tools/RunTimeTester/src/FindInstalledDirsMonolithic.py.obsolete b/Tools/RunTimeTester/src/FindInstalledDirsMonolithic.py.obsolete
deleted file mode 100755
index 1617319b74cdb5a5c5a23b45271adc9b14335fb3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/FindInstalledDirsMonolithic.py.obsolete
+++ /dev/null
@@ -1,102 +0,0 @@
-import os, os.path
-from os.path import join
-
-from Factory_HardCoded import HardCodedFactory
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-class FindInstalledDirsMonolithic:
-    "A class holding functions to find various installed monolithic build directories"
-
-    def __init__(self, local, releasePath, runType):
-        """
-        local - must be a boolean
-        releasePath - string that represents a path
-        local = true: find local installed area from CMTPATH env variable
-        """
-        self.local = local # boolean
-        self.releasePath = releasePath
-        self.releaseType = 'monolithic'
-        self.runType     = runType
-        
-    def findInstallAreas(self):
-        """
-        returns a list of strings representing paths
-        Always add the release installed area.
-        """
-    
-        paths = []
-        hardCoded   = HardCodedFactory(logger).create(self.releaseType,
-                                                      self.runType)
-        installArea = hardCoded.installAreaDirName()
-        
-        if self.local:
-            if 'CMTPATH' in os.environ.keys():
-                cmtpaths = os.environ['CMTPATH'].split(':')
-                cmtpaths = [join(path, installArea) for path in cmtpaths]
-                paths.extend(cmtpaths)
-
-
-        paths.append(join(self.releasePath, installArea))
-        paths = [path for path in paths if os.path.exists(path)]
-    
-        logger.info('installed dirs'+str(paths))
-        return paths
-
-    def findInstalledDirs(self, directory):
-        """
-        Returns a list of strings representing full directory paths
-        hanging from Installed.
-        The installed directories are looked for in the directories given
-        in CMTPATH."""
-        
-        paths = self.findInstallAreas()
-        paths = [join(path, directory) for path in paths]
-        paths = [path for path in paths if os.path.exists(path)]
-        
-        logger.debug('local dirs included:'+str(paths))
-        return paths
-
-    def findOptionsPaths(self):
-        """
-        Returns a list of paths to be used to find option files.
-        The most local path is given first, so in locating option files
-        the path should be traversed from beginning to end.
-        """
-        
-        return self.findInstalledDirs('jobOptions')
-
-    def findSharePaths(self):
-        """
-        Returns a list of paths that will act as a search list.
-        """
-        return self.findInstalledDirs('share')
-
-
-if __name__ == '__main__':
-    "Test function"
-
-    import sys
-    from formatCollection import formatCollection
-    
-    if  'CMTPATH' not in os.environ.keys():
-        print 'run home CMT setup.sh before this test'
-        sys.exit()
-    codeStatuses=['Installed', 'CheckedOut']
-    dirs = ['jobOptions', 'share']
-    release = '/afs/cern.ch/atlas/software/dist/9.0.2'
-
-
-    fids = [FindInstalledDirsMonolithic(True,  release),
-            FindInstalledDirsMonolithic(False, release)]
-    for fid in fids:
-        print '\n\njobOptions paths'
-        print formatCollection(fid.findOptionsPaths())
-
-
-        print '\n\nshare paths'
-        print formatCollection(fid.findSharePaths())
-
-        
-        
diff --git a/Tools/RunTimeTester/src/FindInstalledDirsProject.py.obsolete b/Tools/RunTimeTester/src/FindInstalledDirsProject.py.obsolete
deleted file mode 100755
index f3205a29fa5d36a3d04c71f03f093113d409bd3f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/FindInstalledDirsProject.py.obsolete
+++ /dev/null
@@ -1,105 +0,0 @@
-import os, os.path
-from os.path import join
-
-from Factory_HardCoded import HardCodedFactory
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-class FindInstalledDirsProject:
-    "A class holding functions to find various installed project build directories"
-
-
-    def __init__(self, local, releasePath, runType):
-        """
-        local - must be a boolean
-        releasePath - string that represents a path
-        local = true: find local installed area from CMTPATH env variable
-        """
-        self.local = local # boolean
-        self.releasePath = releasePath
-        self.releaseType = 'project'
-        self.runType     = runType
-        
-    def findInstallAreas(self):
-        """
-        returns a list of strings representing paths
-        Always add the release installed area.
-        """
-    
-        paths = []
-        
-        hardCoded   = HardCodedFactory(logger).create(self.releaseType,
-                                                      self.runType)
-        installArea = hardCoded.installAreaDirName()
-        
-        if self.local:
-            if 'CMTPATH' in os.environ.keys():
-                cmtpaths = os.environ['CMTPATH'].split(':')
-                cmtpaths = [join(path, installArea) for path in cmtpaths]
-                paths.extend(cmtpaths)
-
-
-        paths.append(join(self.releasePath, installArea))
-        paths = [path for path in paths if os.path.exists(path)]
-    
-        logger.info('installed dirs'+str(paths))
-        return paths
-
-    def findInstalledDirs(self, directory):
-        """
-        Returns a list of strings representing full directory paths
-        hanging from Installed.
-        The installed directories are looked for in the directories given
-        in CMTPATH."""
-        
-        paths = self.findInstallAreas()
-        paths = [join(path, directory) for path in paths]
-        paths = [path for path in paths if os.path.exists(path)]
-        
-        logger.debug('local dirs included:'+str(paths))
-        return paths
-
-    def findOptionsPaths(self):
-        """
-        Returns a list of paths to be used to find option files.
-        The most local path is given first, so in locating option files
-        the path should be traversed from beginning to end.
-        """
-        
-        return self.findInstalledDirs('jobOptions')
-
-    def findSharePaths(self):
-        """
-        Returns a list of paths that will act as a search list.
-        """
-        return self.findInstalledDirs('share')
-
-
-if __name__ == '__main__':
-    "Test function"
-
-    import sys
-    from formatCollection import formatCollection
-    
-    if  'CMTPATH' not in os.environ.keys():
-        print 'run home CMT setup.sh before this test'
-        sys.exit()
-    codeStatuses=['Installed', 'CheckedOut']
-    dirs = ['jobOptions', 'share']
-    release = '/afs/cern.ch/atlas/software/dist/9.0.2'
-
-
-    fids = [FindInstalledDirsProject(True,  release),
-            FindInstalledDirsProject(False, release)]
-    for fid in fids:
-        print '\n\njobOptions paths'
-        print formatCollection(fid.findOptionsPaths())
-
-
-        print '\n\nshare paths'
-        print formatCollection(fid.findSharePaths())
-
-        
-        
diff --git a/Tools/RunTimeTester/src/Geneology.py b/Tools/RunTimeTester/src/Geneology.py
deleted file mode 100755
index 57df3069d0d8456d4ed92564c76e13489a97bb47..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Geneology.py
+++ /dev/null
@@ -1,371 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Base class dealing with geanelogy of tree structures
-"""
-# class Logger:
-#     def __init__(self):
-#         pass
-#     def critical(self, m):
-#         print str(m)
-#     def error(self, m):
-#         print str(m)
-# 
-# logger = Logger();
-# 
-#
-from copy import copy
-
-
-from formatCollection import formatCollection
-from Tools            import unique
-from RTTSException    import RTTCodingError
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class Tree:
-    def __init__(self, nodes, topLabel=None):
-
-        logger.info('no. of nodes on entering Tree constructer = %d',len(nodes))
-        self.nodeDict       = {}
-        self.removedNodes   = []
-        self.topLabel       = 'Top'
-        self.topParentLabel = 'TopParent'
-
-        self.sanity(nodes)
-        
-        for n in nodes: self.nodeDict[str(n.myLabel())]=n
-        self.nodeDict[self.topParentLabel] = None
-        self.lineage(nodes)
-
-    def nodeString(self, nodes):
-        def concoctString(node):
-            return 'parent: '+n.parentLabel()+' group: '+ n.myLabel()+'\n'
-        temp = [concoctString(n) for n in nodes]
-        return reduce(lambda x,y:x+y, temp, ' ')
-    
-    def sanity(self, nodes):
-        m = 'Geneology sanity check: nodes have:' + self.nodeString(nodes)
-        logger.debug(m)
-        self.areNodes(nodes)
-        self.removeDuplicates(nodes)
-        self.parentsPresent(nodes)
-        self.hasTop(nodes)
-
-    def areNodes(self, nodes):
-        nonNodes = [n for n in nodes if not self.isNode(n)]
-        for n in nonNodes:
-            logger.error('removing non Node %s from Tree' % str(n))
-            nodes.remove(n)
-
-    def removeDuplicates(self, nodes):
-        multiplicity = {}
-        [multiplicity.setdefault(n.myLabel(), []).append(n) for n in nodes]
-
-        for label in multiplicity.keys():
-            nDup = len(multiplicity[label])
-            if nDup >1:
-                msg  = 'removing %d nodes with '
-                msg += 'duplicate label %s' % (nDup, label)
-                logger.error(msg)
-                [nodes.remove(n) for n in multiplicity[label]]
-
-
-    def parentsPresent(self, nodes):
-        """
-        remove a node if its parent is not present.
-        if >= 1 node is removed, re-call the method.
-        """
-        
-        labels = [n.myLabel() for n in nodes]
-        parentMissing = [n for n in nodes if n.parentLabel() not in labels]
-        tops = [n for n in parentMissing if n.myLabel() == self.topLabel]
-        [parentMissing.remove(t) for t in tops]
-        if parentMissing:
-            [nodes.remove(n) for n in parentMissing]
-            self.removedNodes.extend(parentMissing)
-            msg  = 'No parents for the following nodes; the nodes have been removed\n'
-            msg += self.nodeString(parentMissing)
-            logger.debug(msg)
-            self.parentsPresent(nodes)
-            
-    def hasTop(self, nodes):
-
-        tops = [n for n in nodes if n.myLabel() == self.topLabel]
-
-        assert (len(tops)<2)
-
-        if len(tops) == 0:
-            msg  = 'No top node in Tree, raising exception\n'
-            msg += 'top node label %s \n' % self.topLabel
-            msg += formatCollection([n.myLabel() for n in nodes])
-            logger.error(msg)
-            raise RTTCodingError(msg)
-
-        top = tops[0]
-        top.parent = self.topParentLabel
-
-    def lineage(self, nodes):
-
-        for n in nodes:
-
-            # do not remove the top node
-            if n.myLabel() == self.topLabel: continue
-            
-            aLabels =[a.myLabel() for a in self.ancestors(n)]
-            if self.topLabel not in aLabels:
-                msg  = 'removing node %s ' % n.myLabel()
-                msg += 'which cannot be traced back to the top node'
-                msg += 'ancestors are:\n%s\n ' % (formatCollection(aLabels))
-                msg += 'top label is: %s' % self.topLabel
-                logger.error(msg)
-                nodes.remove(n)
-
-                
-
-    def nNodes(self):
-        return len(self.nodeDict.keys())
-
-    def nodeWasRemoved(self,nodeName):
-        for rn in self.removedNodes:
-            if nodeName==rn.myLabel():
-                return True
-        return False
-        
-    def getNodeNames(self):
-        return self.nodeDict.keys()
-
-    def getNode(self, nodeName):
-
-        node = self.nodeDict.get(nodeName, None)
-        if node == None  and nodeName != self.topParentLabel:
-            if self.nodeWasRemoved(nodeName):
-                msg = 'Unknown node: %s was removed because it had no known parent node.' % nodeName
-            else:
-                msg =  'Unknown node:'
-                msg +=' %s\n possible nodes %s:' % (nodeName,
-                                                    str(self.nodeDict.keys()))
-            logger.error(msg)
-        return node
-
-    def isNode(self, node):
-        try:
-            parent = node.parentLabel()
-            me     = node.myLabel()
-        except:
-            logger.error(str(node)+' is not a node object')
-            return False
-        return True
-    
-    def ancestors(self, nnode):
-        """
-        returns a list of Nodes which are parents of the Node node.
-        Stops at the Top node if this an ancestor
-        """
-        
-        ancestors          = []
-        node               = copy(nnode)
-        label              = node.myLabel()
-
-        #if passed node is the top node, return it.
-        if label == self.topLabel: return [node]
-        
-        while label != self.topParentLabel:
-
-            ancestors.append(node)
-            node = self.nodeDict[node.parentLabel()]
-            label = node.myLabel()
-            if label == self.topLabel:
-                ancestors.append(node)
-                break
-
-
-        # ancestors.remove(self.nodeDict[self.topLabel])
-        return ancestors
-
-    def getAllNodes(self):
-        nodes = []
-
-        keys = self.nodeDict.keys()
-        if self.topParentLabel in keys: keys.remove(self.topParentLabel)
-
-        [nodes.append(self.nodeDict[k]) for k in keys]
-
-        return nodes 
-
-
-class KitTree(Tree):
-    def __init__(self, ourKits, theirKits=[]):
-
-        self.removeOverlaps(ourKits, theirKits)
-        self.removeDuplicates(ourKits)
-        self.removeDuplicates(theirKits)
-
-        kits = copy(ourKits)
-        kits.extend(theirKits)
-        
-
-
-        topParent   = str(None)
-        Tree.__init__(self, kits, topParent)
-        self.propagate()
-
-    def propagate(self):
-        """
-        propagate parent kit to child kit down the tree for each generation
-        """
-        generations = {}
-        nodes = self.getAllNodes()
-
-        for n in nodes:
-            nGen = len(self.ancestors(n))
-            generations.setdefault(nGen, []).append(n)
-
-        nGen = range(1, max(generations.keys())+1)
-        for gen in nGen:
-            thisGeneration = generations[gen]
-            for node in thisGeneration:
-                parentNode = self.nodeDict[node.parentLabel()]
-                #add the parent is is not the top parent
-                if parentNode: node.add(parentNode)
-
-    def getKit(self, label):
-        return self.getNode(label)
-    
-    def removeOverlaps(self, on, tn):
-        onLabels = [n.myLabel() for n in on]
-        overlap  = [n for n in tn if n.myLabel() in onLabels]
-        if len(overlap)>1:
-            msg  = 'the following nodes overlap with class A nodes'
-            msg += ' and have been removed'
-            logger.error(msg)
-            msg = ''
-            for n in overlap: msg+=n.myLabel()+' '
-            logger.error(msg)
-            [tn.remove(n) for n in overlap]
-        
-    def removeDuplicates(self, nodes):
-        """
-        Identify kits with the same lable. Remove all such kits.
-        """
-        dict = {}
-        [dict.setdefault(n.myLabel(),[]).append(n)  for n in nodes]
-
-        for label in dict.keys():
-            list = dict[label]
-            if len(list)>1:
-                testKit = list[0]
-                for kit in list:
-                    if not testKit.equivalent(kit):
-                        msg  = 'Node with label %s has non-equivalent '% label
-                        msg += 'duplicates:  removing all such nodes'
-                        logger.error(msg)
-                        [nodes.remove(n) for n in list]
-                        break
-        
-    def getKitNames(self):
-        return self.getNodeNames()
-                       
-            
-    def nKits(self): return self.nNodes()
-    def getAllKits(self): return self.getAllNodes()
-class TreeMember:
-    def __init__(self, name, parent):
-        self.name       = name
-        self.parent     = parent
-        self.parentNode = None
-    def parentLabel(self):
-        return str(self.parent)
-    def myLabel(self):
-        return str(self.name)
-    def setParent(self, node): self.parentNode = node
-    def __str__(self):
-        msg  = 'Tree Member: parent = '+self.parentLabel()
-        msg += ' me = '+self.myLabel()
-        return msg  
-
-class A(TreeMember):
-    def __init__(self, name, parent):
-        TreeMember.__init__(self, name, parent)
-            
-
-
-def fill(table):
-    nodes = []
-    for item in table.items():
-        nodes.append(A(item[0], item[1]))
-
-    return nodes
-
-def TreeTest(logger):
-    table= {
-        'Top': 'TopParent',
-        'Athena': 'Top',
-        }
-
-
-    nodes = fill(table)    
-    tree = Tree(nodes)
-    
-    for n in [m.myLabel() for m in nodes]:
-        logger.debug('node '+n)
-        if not tree.getNode(n):
-            logger.debug ('  error retrieving node '+str(n))
-            
-
-    if tree.nNodes() != 2:
-        m = 'Expected 2 nodes, found %d' % tree.nNodes()
-        return False
-
-
-    table = {
-        'c': 'Athena',
-        'd': 'Top',
-        'e': 'd'
-        }
-
-
-    nodes = fill(table)
-    tree = Tree(nodes)
-
-    if tree.nNodes() != 2:
-        m = 'Expected 2 nodes, found %d' % tree.nNodes()
-        return False
-
-
-
-    table= {
-        'Top': 'TopParent',
-        'Athena': 'Top',
-        'omega':  'chi'
-        }
-
-    nodes = fill(table)
-    tree = Tree(nodes)
-
-    if tree.nNodes() != 2:
-        m = 'Expected 2 nodes, found %d' % tree.nNodes()
-        return False
-
-
-    return True
-
-def KitTest(logger, kits):
-    tree = KitTree(kits)
-
-    kits = tree.getAllKits()
-    for k in kits:
-        if k.myLabel() == 'Top': print k
-
-    cur = 0
-    for k in kits:
-        cur += 1
-        logger.debug( '-------------%d-------------', cur)
-        logger.debug(str(k))
-
-if __name__ == "__main__":
-    pass
-    
diff --git a/Tools/RunTimeTester/src/GetUserConfigData.py b/Tools/RunTimeTester/src/GetUserConfigData.py
deleted file mode 100755
index 17da9d24f7c45bb407654dee8693c5628d2f6bbd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/GetUserConfigData.py
+++ /dev/null
@@ -1,188 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class that takes in an top level RTT top configuration file,
-and returns a list of tuples of form (configurationStyle, dict)
-where config style is a string with values
-['unifiedConfiguration | rttConfiguration]
-and is a dictionary with key = package name, value = path to conf file
-"""
-#----- RTT modules -----
-
-from Factory_UserStuffRetriever  import UserStuffRetrieverFactory
-from formatCollection            import formatCollection
-from CVSPackage                  import CVSPackage
-from RTTpath                     import RTTpath
-# -------------------------------------------------------------------------
-import os.path, xml.dom.minidom, copy
-# -------------------------------------------------------------------------
-
-class GetUserConfigData:
-    def __init__(self, argBag):
-
-        self.usr                  = argBag.pop('usr')
-        self.logger               = argBag.pop('logger')
-        self.pnb                  = argBag.pop('pnb')
-
-        self.installedConfFiles   = argBag.pop('installedConfFiles', None)
-        self.localConfFiles       = argBag.pop('localConfFiles', None)
-        self.rttProvidesConfFiles = argBag.pop('rttProvidesConfFiles', None)
-        self.confFileSuffix       = argBag.pop('confFileSuffix', None)
-        self.confFiles            = argBag.pop('confFiles', None)
-        self.packagesAccept       = argBag.pop('packagesAccept', None)
-        self.packagesVeto         = argBag.pop('packagesVeto', None)
-        self.cmtHomeDir           = argBag.pop('cmtHomeDir', None)
-        self.cmtProjDir           = argBag.pop('cmtProjDir', None)
- 
-        
-        self.installedPackages    = []
-        self.localPackages        = []
-        self.rttProvidesPackages  = []
-
-        self.packages        = []
-        
-        self.releaseConfig()
-        self.rttProvidesConfig()
-        self.packageDomination()
-
-    def __str__(self):
-        str  = 'GetUserConfigData - packages:\n'
-        str += formatCollection([p.name for p in self.packages])
-        return str 
-        
-    def userConfigData(self):
-        return self.packages
-
-    def packageInfo(self, packageName):
-
-        # style = UnifiedConfiguration
-        for p in self.packages:
-            if p.name == packageName:
-                m = 'Found package %s' % (packageName)
-                self.logger.debug(m)
-                return p
-
-        m = 'GetUserDConfigData has no info about package %s ' % packageName
-        self.logger.warning(m)
-        return CVSPackage(packageName, '', '', '', '', '', '', '', '',
-                          'No GetUserConfigData data for this package')
-    
-    def releaseConfig(self):
-        # --------------------------------------------------
-        # Get the user requested jobs:
-        # handle packages with unified configuration files
-        # --------------------------------------------------
-
-        if not (self.installedConfFiles or self.localConfFiles ): return
-    
-        macroSubString    = self.confFileSuffix
-        dicts             = {macroSubString:[]}
-    
-        # instlledPackages will contain a list of all CVSPackage objects
-        # (installed, local, rttProvided)    allPackages       = [] 
-
-        # -------------------------------------------------
-        # ----- RELEASE PACKAGES --------------------------
-        # -------------------------------------------------
-
-        # get a list of CVSPackage objects
-        self.installedPackages.extend(self.usr.userStuff())
-
-    # -------------------------------------------------
-    # ----- RTT PROVIDED PACKAGES ---------------------
-    # -------------------------------------------------
-    
-    def rttProvidesConfig(self):        
-        if not self.rttProvidesConfFiles: return
-
-        topLevelPackageDict = self.confFiles
-
-
-        # set the path given to the xml file
-        # packagename is a cvs offset eg 'Simulation/Atlfast/TestAtlfast'
-        # {packagename: {'confFilePath': xxxx,
-        #                'cmtDir':      yyyy}
-        #
-        #    .
-        #    .
-        #    .}
-
-
-        for packageOffset  in topLevelPackageDict.keys():
-            innerDict        = topLevelPackageDict[packageOffset]
-            pathToConfigFile = innerDict['confFilePath']
-            #
-            # cmtPath, version, root are package terms used by CMT
-            # examples of values are documented in UserStuffRetriever
-            #
-            # the cmt mechanisms for determinging
-            # packageCmtPath, packageCmtDir, packageCmtDir, packageRoot
-            # are not currently working. The variable packageSharePath
-            # is the only one currently used in CVSPackage.
-            # kludge this to point to the user provided directory containing
-            # the _TestConfiguration file
-            #
-            packageCmtPath   = innerDict['cmtPath']
-            packageCmtDir    = str(self.pnb.cmtDir(packageCmtPath))
-            packageVersion   = innerDict['version']
-            packageRoot      = str(RTTpath(packageCmtPath, packageOffset))
-            packageSharePath = innerDict['shareArea']
-            # RTT provides files are not 'local'  as no CMT actions
-            # have been done by hand
-            local         = False 
-            errMsg = ''
-            if not os.path.exists(pathToConfigFile):
-                errMsg = '"RTT provides" declares a non-existant configuration'
-                errMsg += 'file %s' % (pathToConfigFile)                
-                self.logger.warning(errMsg)
-
-
-            self.rttProvidesPackages.append(
-                CVSPackage(
-                os.path.basename(packageOffset),
-                pathToConfigFile,
-                packageOffset,
-                packageVersion,
-                packageCmtPath,
-                packageRoot,
-                packageSharePath,
-                # self.cmtHomeDir,
-                # self.cmtProjDir,
-                errMsg,
-                local)
-                )
-                       
-    def packageDomination(self):
-        # add all RTT provided "packages" to the list
-        packages = copy.copy(self.rttProvidesPackages)
-        packagesNames = [package.name for package in packages]
-        packages.extend([p for p in self.localPackages if p.name not in packagesNames])
-
-        # add installled packages to the list
-        packagesNames = [package.name for package in packages]
-        packages.extend([p for p in self.installedPackages if p.name not in packagesNames])
-
-        # keep configuration files only if they decalre RTT jobs   
-        noRTT = [p.name for p in packages if not p.wantsRTT]
-        msg = 'The following packages do not have RTT jobs %s' % str(noRTT)
-        self.logger.debug(msg)
-        packages = [p for p in packages if p.wantsRTT]
-        
-        # now handle user accept and veto requests
-        # Do not remove any packages - just flag if jobs are not to be run
-        # RunTimeTester is needed for reasons other than running jobs.
-        
-        accept    = self.packagesAccept
-        if accept: # do nothing if the user does not ask for package selection
-            [p.setNoRun() for p in packages if p.name not in accept]
-
-        veto      = self.packagesVeto
-        [p.setNoRun() for p in packages if p.name in veto]
-
-        self.packages = packages
-
-
-        
-if __name__ == "__main__":
-    pass
-
diff --git a/Tools/RunTimeTester/src/GroupKits.py.obsolete b/Tools/RunTimeTester/src/GroupKits.py.obsolete
deleted file mode 100755
index 92362917353e844f2a18752922c7b681112d9b5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/GroupKits.py.obsolete
+++ /dev/null
@@ -1,261 +0,0 @@
-"""
-A class that contains information and behaviour specific to different groups.
-For now subclasses contain only names of functions to run when all
-Subclasses overide the dummy fucntions to give the real functionality.
-
-Sub classes:
-
-AtlfastKit
-ArtemisKit
-RecExCommonKit
-HelloWorldKit
-TriggerReleaseKit              
-TriggerReleaseKit1
-"""
-
-import os, sys
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def groupKitFactory(group):
-    dict = {
-        'AthenaAtlfast':          AtlfastGroupKit,
-        'AthenaAtlfastAlgs':      GroupKit,
-        'AthenaHelloWorld':       HelloWorldGroupKit,
-        'AthenaJiveXML':          JiveXMLGroupKit,
-        'AthenaTriggerRelease':   TriggerReleaseGroupKit,
-        'AthenaTriggerReleaseID':    TrigReleaseIDGroupKit,
-        'AthenaTriggerReleaseCalo':  TrigReleaseCaloGroupKit,
-        'AthenaTriggerRelease3':  TriggerReleaseGroupKit3,
-        'AthenaTriggerReleaseReco':  TriggerReleaseRecoGroupKit,
-        'AthenaTriggerReleaseBS':    TriggerReleaseBSGroupKit,
-        'AthenaRecExCommon':      RecExCommonGroupKit,
-        'AthenaegammaRec':        EgammaGroupKit,
-        'AthenaInDetRecValidation': InDetRecValidationGroupKit,
-        'AthenaMuonDigiExample':   MuonDigiExampleGroupKit,
-        #        'RuleChecker':            RuleCheckerGroupKit,
-        # 'RuleCheckerWatcher':     RuleCheckerWatcherGroupKit,
-        # 'QAmetrics':              QAmetricsGroupKit,
-        'AthenaAtlfastWatcher':   GroupKit,
-        'AthenaegammaRecWatcher': GroupKit,
-        'KitValidation':          GroupKit,
-        'AthenaMooPerformance':   GroupKit,
-        'AthenaCaloRecEx':        GroupKit,
-        'AthenaCaloDigEx':        GroupKit,
-        'AthenaCaloSimEx':        GroupKit,
-        'AthenaCaloRec':          GroupKit,
-        'AthenaBPhysValidation':  GroupKit,
-        'AthenaAnalysisExamples': GroupKit,
-        'AthenaDigitization':     GroupKit,
-        'AthenaG4AtlasApps':      GroupKit,
-        'AthenaG4AtlasAppsRelease':      GroupKit,
-        'AthenaGeneratorsRTT':    GroupKit,
-        'AthenaInDetRTT':         GroupKit,
-        'AthenaMissingET':        GroupKit,
-        'AthenaPhysicsAnalysisTools': GroupKit,
-        'Monitoring':             GroupKit,
-        'AthenaMuonboyAth':       GroupKit,
-        'AthenaMboyPerformance':       GroupKit,
-        'AthenaTauRec':           GroupKit
-        }
-
-    if group in dict.keys():
-        return dict[group]()
-    else:
-        logger.error('unknown group: '+group+' exiting')
-        sys.exit(0)
-
-#################################################                    
-class GroupKit:
-    """ Manage functionality that depends on job groups, by group
-    rather than by job,
-    """
-    
-    def __init__(self):
-        "Set default values for all groups"
-        self.pScripts = []
-        self.displayFiles = []
-        self.setPostScripts()
-        
-
-    def setPostScripts(self):
-        'Specific postScripts are set by derived classes'
-        pass
-
-    def runPostScripts(self, srcDir, resDir):
-        """
-        Runs postscripts.
-
-        PostScripts:
-        1. Take their inputs from the srcDir
-        2. Write their output file to the resDir
-        3. return the name of the output file. This is appended to
-           the files to display
-        """
-        for script in self.pScripts:
-            self.displayFiles.append(script(srcDir, resDir))
-
-    def displayFiles(self):
-        """
-        Return the list of files to display.
-        These are typically produced by a postScript that
-        summarises all the jobs of a group
-        """
-        return self.displayFiles
-#################################################                    
-class AthenaGroupKit(GroupKit):
-    """Base class for providing shared behaviour for all Athena jobs"""
-    def __init__(self):
-        "Supplies a checker that looks for the Athena success message"
-        
-        GroupKit.__init__(self)
-#################################################                    
-class AtlfastGroupKit(GroupKit):
-    def __init__(self):
-        "Set defaults for Atlfast"
-        GroupKit.__init__(self)
-#################################################                    
-class HelloWorldGroupKit(GroupKit):
-    def __init__(self):
-        "Default values for HelloWorld"
-        GroupKit.__init__(self)
-#################################################                    
-class RecExCommonGroupKit(GroupKit):
-    
-    def __init__(self):
-        "Default values for RecExCommon" 
-        GroupKit.__init__(self)
-#################################################                    
-class JiveXMLGroupKit(GroupKit):
-    def __init__(self):
-        "Default values for JiveXML" 
-        GroupKit.__init__(self)
-#################################################                    
-class RuleCheckerWatcherGroupKit(GroupKit):
-    def __init__(self):
-        "Default values for RuleCheckerWatcher" 
-        GroupKit.__init__(self)
-#################################################
-
-from viol2xml import viol2xml
-
-class RuleCheckerGroupKit(GroupKit):
-    
-    
-    def __init__(self):
-        "Default values for RuleChecker" 
-        GroupKit.__init__(self)
-
-    # def setPostScripts(self):
-    #    self.pScripts.append(viol2xml)
-        
-#################################################
-from qaMetrics2OnePage import qaMetrics2OnePage
-
-class QAmetricsGroupKit(GroupKit):
-        
-    def __init__(self):
-        "Default values for QAmetrics" 
-        GroupKit.__init__(self)
-    def setPostScripts(self):
-        self.pScripts.append(qaMetrics2OnePage)
-        
-#################################################
-class TriggerReleaseGroupKit(GroupKit):
-    "Default values for TriggerRelease" 
-    def __init__(self):
-        '''default values for TriggerRelease '''
-        GroupKit.__init__(self)
-#################################################                    
-class EgammaGroupKit(GroupKit):
-    def __init__(self):
-        "Set defaults for egamma"
-        GroupKit.__init__(self)
-#################################################                    
-
-class TrigReleaseIDGroupKit(GroupKit):
-    '''Default values for Trigger sub group
-    '''
-    def __init__(self):
-        GroupKit.__init__(self)
-################################################
-
-class TrigReleaseCaloGroupKit(GroupKit):
-    '''Default values for Trigger subgroup
-    T2Calo regression test performed
-    '''
-    def __init__(self):
-        GroupKit.__init__(self)
-
-#################################################
-
-class TriggerReleaseBSGroupKit(GroupKit):
-    "Default values for TriggerRelease"
-    def __init__(self):
-        '''default values for TriggerRelease '''
-        GroupKit.__init__(self)
-
-#################################################
-
-class TriggerReleaseRecoGroupKit(GroupKit):
-    "Default values for TriggerRelease"
-    def __init__(self):
-        '''default values for TriggerRelease '''
-        GroupKit.__init__(self)
-
-#################################################
-        
-class TriggerReleaseGroupKit3(GroupKit):
-    '''ref files:
-       idscan_readBS.reference
-       sitrack_readBS.reference
-    '''
-    def __init__(self):
-        GroupKit.__init__(self)
-
-
-#################################################                    
-class InDetRecValidationGroupKit(GroupKit):
-    def __init__(self):
-        "Default values for InDetRecValidation"
-        GroupKit.__init__(self)
-
-class MuonDigiExampleGroupKit(GroupKit):
-    def __init__(self):
-        "Default values for MuonDigiExample"
-        GroupKit.__init__(self)
-
-if __name__ == "__main__":
-    kitnames = ['AthenaAtlfast',
-                'AthenaHelloWorld',
-                'AthenaJiveXML',
-                'AthenaTriggerRelease',
-                'AthenaTrigReleaseID',
-                'AthenaTrigReleaseCalo',
-                'AthenaRecExCommon',
-                'Athenaegamma',
-                'RuleChecker',
-                'QAmetrics',
-                'AthenaAtlfastWatcher',
-                'AthenaegammaWatcher',
-                'KitValidation']
-
-    kits = [groupKitFactory(kitname) for kitname in kitnames]
-
-
-    def kitContent(kit):
-        print ''
-        print ''
-        print '*********  kit: ',kit.__class__.__name__+'  *********'
-        print ''
-        print 'postScripts:'
-        print str(kit.pScripts)
-        print ''
-        print 'display files'
-        print str(kit.displayFiles)
-
-    [kitContent(kit) for kit in kits]
-    
-        
diff --git a/Tools/RunTimeTester/src/GroupsStatus.py.obsolete b/Tools/RunTimeTester/src/GroupsStatus.py.obsolete
deleted file mode 100755
index 654faddfb7dc06c9b2ff001423c51d82a088c11c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/GroupsStatus.py.obsolete
+++ /dev/null
@@ -1,89 +0,0 @@
-from formatCollection import formatCollection
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class GroupsStatus:
-    
-    def __init__(self, jobs):
-        'Sets up dictionary of lists of jobs keyed on jobg group'
-        
-        self.groups = {}
-        for job in jobs:
-            self.groups.setdefault(job.jDescriptor.jobGroup,[]).append(job)
-
-    def __str__(self):
-
-        s = 'Group Status dict:\n'
-        s += formatCollection(self.groups)
-        return s
-        
-    def allGroups(self):
-        return self.groups.keys()
-
-    def doneGroups(self):
-        """
-        Return a list stating of all jobs in a group have finished.
-        """
-
-        # first remove finished jobs:
-        for group in self.groups.keys():
-            jobsToRemove = [job for job in self.groups[group] if job.done]
-            [self.groups[group].remove(job) for job in jobsToRemove]
-
-        # return a dictionary stating if all jobs in a group are done.
-
-        # return a list of all groups which no longer have running jobs
-        # i.e. for which all jobs have been removed from the status dictionary
-        done = []
-        [done.append(group) for group in self.groups.keys() if not self.groups[group]]
-
-        return done
-
-if __name__ == '__main__':
-
-    from minderFactory            import minderFactory
-    from descriptorFactory        import descriptorFactory
-
-    from Paths                    import Paths
-    things = []
-    ajd = descriptorFactory('AthenaJobDescriptor')
-    rjd = descriptorFactory('ReleaseJobDescriptor')
-
-    
-    ajd.jobGroup = 'Athenaegamma'
-
-    paths = Paths()
-    paths.refDBenable = False
-
-    ajd.package = 'blah1'
-    rjd.package = 'blah2'
-
-    ajd.runPath = 'run1'
-    rjd.runPath = 'run2'
-    
-    ajd.initialise(paths)
-    rjd.initialise(paths)
-
-
-    ajd.dump()
-    rjd.dump()
-    things.append(minderFactory('Dummy',ajd,''))
-    things.append(minderFactory('Dummy',ajd,''))
-    things.append(minderFactory('Dummy',rjd,''))
-    things.append(minderFactory('Dummy',ajd,''))
-    things.append(minderFactory('Dummy',rjd,''))
-    things.append(minderFactory('Dummy',rjd,''))
-
-    expect= ['FF', 'FF', 'FF', 'TF', 'TF', 'TT']
-
-    gs = GroupsStatus(things)
-    print gs.doneGroups(),' expect ',expect[0]
-    for i in range(6):
-        print 'setting done for job position ',i
-        things[i].done = True
-        print gs.doneGroups(),' expect ',expect[i]
-
-
diff --git a/Tools/RunTimeTester/src/HTMLWriter.py.obsolete b/Tools/RunTimeTester/src/HTMLWriter.py.obsolete
deleted file mode 100755
index 6598177e6bf605f9bb4f5a6a45b728f045999311..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/HTMLWriter.py.obsolete
+++ /dev/null
@@ -1,1260 +0,0 @@
-#===== Python imports ==
-from os.path                 import join
-import                       os, copy, string
-
-#===== RTT imports =====
-from Coloriser               import *
-from GroupKits               import groupKitFactory
-from getInheritanceHeirarchy import isAthenaJob
-from Tools2                  import fixSortBranches, getAllResultDirs
-from Tools2                  import jobStatusFromFile, setProtectionFile
-from NICOS                   import NICOS
-from Factory_HardCoded       import HardCodedFactory
-from RTTDateTime             import RTTDateTime
-from RTTpath                 import RTTpath
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-def HTMLheader():
-    return """
-    <!DOCTYPE html PUBLIC \"-//W3C//DTD XHTML 1.0 Strict//EN\" \"http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd\">
-    <html>
-    <head>
-
-    <meta http-equiv="Pragma" content="no-cache">
-    <!-- Pragma content set to no-cache tells the browser not to cache the page
-    This may or may not work in IE -->
-
-    <meta http-equiv="expires" content="0">
-    <!-- Setting the page to expire at 0 means the page is immediately expired
-    Any vales less then one will set the page to expire some time in past and
-    not be cached. This may not work with Navigator -->
-
-    <style>
-    body{background-color:#ddddff}
-    acronym{border-bottom: dotted 2px green; cursor:help}
-    .level0TableHeader{font-family:arial;font-size:20pt;font-weight:bold;color:gray;background-color:#666666}
-    .level1TableHeader{font-family:arial;font-size:14pt;font-weight:bold;color:black;background-color:#8888ff}
-    .level2TableHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;background-color:#aaaaff}
-    .parentTableCellColor{background-color:#ddddff}
-    .tableEntry{font-family:arial;font-size:14pt;color:#333333;background-color:#bbbbff}
-    .failedTableEntry{font-family:arial;font-size:14pt;color:#333333;background-color:#9999cc}
-
-    .nightlyTableEntry{font-family:arial;font-size:14pt;color:#333333;background-color:#aaaaee}
-    
-    a.tableLink{font-family:arial;font-size:14pt;color:#333333;text-decoration:none;border-bottom:1px solid}
-    a.tableLink:hover{font-family:arial;font-size:14pt;color:orange;text-decoration:none;border-bottom:1px solid}
-    a.tableLinkWithInfo{font-family:arial;font-size:14pt;color:#333333;text-decoration:none;border-bottom:1px dashed}
-    a.tableLinkWithInfo:hover{font-family:arial;font-size:14pt;color:orange;text-decoration:none;border-bottom:1px solid;cursor:help}
-    a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-    a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-    
-    .pageHeader{font-family:arial;font-size:20pt;font-weight:bold;color:black;}
-    a.mail{font-family:arial;font-size:12pt;color:#777777;text-decoration:none}
-    a.mail:hover{font-family:arial;font-size:12pt;color:black;text-decoration:underline}
-    .keepFilesHeader{font-family:arial;font-size:14pt;font-weight:bold;color:black;background-color:#aaaaff}
-    .sectionHeader{font-family:arial;font-size:14pt;font-weight:bold;color:black;}
-    .reporterParamEntry{font-family:arial;font-size:10pt;font-weight:bold;color:black;}
-    .reporterValueEntry{font-family:arial;font-size:10pt;font-weight:plain;color:black;}
-    .latestNews{font-family:arial;font-size:10pt;font-weight:plain;color:black;background-color:#cccccc}
-    
-    .TabCommon {FONT: 18px Verdana; COLOR: #6D6D6D; PADDING: 5px; FONT-WEIGHT: bold; TEXT-ALIGN: center; HEIGHT: 20px; WIDTH: 300px;}
-    .TabContent {PADDING: 5px;}
-    .TabContentBottom {PADDING: 10px; BORDER-BOTTOM: 2px outset #99ccff;}
-    .TabOff {CURSOR: hand; BACKGROUND-COLOR: #cccccc; BORDER-LEFT: 1px solid #BBBBBB;}
-    .TabOn {CURSOR: default; BORDER-TOP: 2px outset #D1D1D1; COLOR: #000000;}
-    .TabBorderBottom{BORDER-BOTTOM: 2px inset #D1D1D1;}
-    .TabActiveBorderLeftRight{BORDER-RIGHT: 2px outset #D1D1D1; BORDER-LEFT: 2px outset #D1D1D1;}
-    .TabActiveBackground {BACKGROUND-COLOR: #ccccff;}
-    </style>
-
-    <script>
-    function TabClick( nTab ){
-    Col = document.getElementsByName(\"Content\");
-    for (i = 0; i < document.getElementsByName(\"Content\").length; i++)
-    {
-    document.getElementsByName(\"tabs\")[i].className = \"TabBorderBottom TabCommon TabOff\";
-    document.getElementsByName(\"Content\")[i].style.display = \"none\";
-    }
-    document.getElementsByName(\"Content\")[nTab].style.display = \"block\";  
-    document.getElementsByName(\"tabs\")[nTab].className = \"TabCommon TabOn TabActiveBackground TabActiveBorderLeftRight\";
-    }
-    </script>
-    
-    </head>
-    <body onload=\"TabClick(0);\">    
-    """
-
-#=================================================================================================
-
-def pageTitle(title):
-    return '<center><span class=\"pageHeader\">' + title + '</span></center><br/><br/>\n'
-
-#=================================================================================================
-
-def sectionTitle(title):
-    return '<center><span class=\"sectionHeader\">' + title + '</span></center><br/><br/>\n'
-
-#=================================================================================================
-
-def blockquote(type='open'):
-    if type=='close':
-        return '</blockquote>'
-
-    return '<blockquote>'
-
-#=================================================================================================
-	
-def table(tagType='begin',border=1,bkgdColor='',width=''):
-    if tagType == 'end':
-        return '</table>' + endLine()
-
-    bkgd = ''
-    if bkgdColor != '':
-        bkgd=' bgcolor=\"' + bkgdColor + '\"'
-
-    widthString = ''
-    if width != '':
-        widthString = 'width=' + width
-    return '<table ' + widthString + ' BORDER=\"'+str(border)+'\" CELLSPACING=\"0\" CELLPADDING=\"4\"' + bkgd + '>' + endLine()
-
-#=================================================================================================
-
-def tableRow(styleClass='tableEntry',colSpan=1,tableCells=[]):
-    html1 = '<tr class=\"' + styleClass + '\">' + endLine()
-    for cell in tableCells:
-        
-        values = {'content':'','color':'','align':'center','style':''}
-        for value in values:
-            try:
-                values[value] = str(cell[value])
-            except KeyError:
-                pass #do nothing, default holds
-
-        styleText = 'style=\"background-color:' + values['color'] + '\"'
-        if values['style']:
-            styleText = 'style=' + values['style']
-
-        html1 += '<td ' + styleText + ' align=\"' + values['align'] + '\" valign=\"top\" colspan=\"'+str(colSpan)+'\">' + values['content'] + '</td>'
-    html1 += '</tr>' + endLine()
-    return html1 
-	
-#=================================================================================================
-        
-def tableEnd():
-    return '</table>' + endLine()
-		
-#=================================================================================================
-
-def HTMLfooter():
-    return """
-    </body>
-    </html>
-    """
-
-#=================================================================================================
-
-def breakTag(number=1):
-    """Return the HTML code for the break tag."""
-    return '<br/>'*number
-
-#=================================================================================================
-
-def endLine(number=1):
-    """Return the end of line character \n."""
-    return '\n'*number
-
-#=================================================================================================
-
-def pageRule():
-    """Return the HTML code for a horizontal rule."""
-    return '<hr/>'
-
-#=================================================================================================
-
-def center(tagType='open'):
-    if tagType == 'close':
-        return '</center>'
-    return '<center>'
-
-#=================================================================================================
-	
-def toFile(htmlFile, htmlContent):
-    page = open(htmlFile,'w', 777)
-    page.write(htmlContent)
-
-#=================================================================================================
-
-def writeTempPage2(resPath):
-    text = center('open') + '<div class=\"pageHeader\">The RTT is currently in initialisation phase...please be patient...and check back to this page later</div>' + center('close') + endLine()
-    content = HTMLheader() + text + HTMLfooter()
-    page2Path = join(resPath,"page2.html")
-    toFile(page2Path,content)    
-
-#=================================================================================================
-
-def writePage2GlobalFailure(resPath):
-    text = '<center><div class=\"pageHeader\"><font color="red">The RTT was unable to run because of a fatal error during initialisation.<br/>More details may be found in the&nbsp;</font><a class=\"warningLink\" href=\"RTT_failure_report.html\">RTT Failure Report</a></div></center>\n'
-    content = HTMLheader() + text + HTMLfooter()
-    page2Path = join(resPath,"page2.html")
-    toFile(page2Path,content)
-
-#=================================================================================================
-
-def writePage2NothingAvailable(resPath):
-    text = '<center><div class=\"pageHeader\">No packages found that required testing! Twiddling thumbs...</div></center>\n'
-    content = HTMLheader() + text + HTMLfooter()
-    page2Path = join(resPath,"page2.html")
-    toFile(page2Path,content)
-
-#=================================================================================================
-
-def writePage2Shutdown(resPath):
-    text = '<center><div class=\"pageHeader\"><font color="red">The RTT process was shut down by the user-typed command: <code>kill -15</code></div></center>\n'
-    content = HTMLheader() + text + HTMLfooter()
-    page2Path = join(resPath,"page2.html")
-    toFile(page2Path,content)
-    
-#=================================================================================================
-
-def link(styleClass, pointsAt, name):
-    """Return the HTML code for a link with the given style class, href adress and name in the page."""
-    return '<a class=\"' + styleClass + '\" href=\"' + pointsAt  + '\">' + name + '</a>'
-
-#=================================================================================================
-
-def linkWithInfoString(styleClass, pointsAt, name, info):
-    """Return the HTML code for a link with the given style class, href adress and name in the page."""    
-    return '<a class=\"' + styleClass + '\" title=\"' + info + '\" href=\"' + pointsAt  + '\">' + name + '</a>'
-
-#=================================================================================================
-
-def linkSep():
-    """Return the character used to separate links in the web page."""
-    return ' | '
-    
-#=================================================================================================
-
-def page1EmailLinks():
-    email = pageRule() + endLine()
-    email += link('mail', 'mailto:en@hep.ucl.ac.uk','Eric Nzuobontane') + linkSep()
-    email += link('mail', 'mailto:peter.sherwood@cern.ch','Peter Sherwood') + linkSep()
-    email += link('mail', 'mailto:brinick.simmons@cern.ch','Brinick Simmons') + breakTag() + endLine()
-
-    return email
-
-#=================================================================================================
-
-def page1HeaderLinks_Javascript(resBaseDir):
-    theScript = 'RTTpage1_LinksAtTopOfPage.js'
-    fullPath  = os.path.join(resBaseDir,theScript)
-    if not os.path.exists(fullPath):
-        pass
-        # createPage1LinksJavaScriptFile(fullPath)
-                
-    return '<script src="' + theScript + '"></script>' + endLine() + breakTag()
-    
-#=================================================================================================
-
-def createPage1LinksJavaScriptFile(outputPath):
-    js  = "/*********************/\n"
-    js += "/***** Base URLs *****/\n"
-    js += "/*********************/\n\n"    
-    js += "uclHEP    = 'http://www.hep.ucl.ac.uk/atlas/AtlasTesting/';\n"
-    js += "atlasComp = 'http://atlas-computing.web.cern.ch/atlas-computing/';\n\n"
-    js += "/*********************/\n"
-    js += "/***** Style CSS *****/\n"
-    js += "/*********************/\n\n"
-    js += "styleClass  = '<style>';\n"
-    js += "styleClass += 'a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}';\n"
-    js += "styleClass += 'a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}';\n"
-    js += "styleClass += '</style>';\n\n"
-    js += "/*********************/\n"
-    js += "/***** The links *****/\n"
-    js += "/*********************/\n\n"
-    js += "link1 = '<a class=\"pageLink\" href=\"http://www.hep.ucl.ac.uk/atlas/RTTprojects/RTTpage1.html\"><font size=\"2\">RTT project runs</font></a>';\n"
-    js += "link2 = '<a class=\"pageLink\" href=\"' + uclHEP + 'RTTstatus/RTTstatus.html\">RTT status reports</a>';\n"
-    js += "link3 = '<a class=\"pageLink\" href=\"' + uclHEP + '\">RTT documentation</a>';\n"
-    js += "link4 = '<a class=\"pageLink\" href=\"' + atlasComp + 'projects/releases/releases.php\">Release Status</a>';\n"
-    js += "link5 = '<a class=\"pageLink\" href=\"' + atlasComp +'links/distDirectory/nightlies/global/\">NICOS</a>';\n"
-    js += "link6 = '<a class=\"pageLink\" href=\"' + uclHEP + 'RTTstatus/Monitoring.html\">Monitoring plots</a>';\n"
-    js += "link7 = '<a class=\"pageLink\" href=\"' + uclHEP + 'DatasetManager.report\">Dataset Manager Report</a>';\n\n"
-    js += "document.write(\n"
-    js += "   styleClass\n"
-    js += " + '<div align=\"center\">'\n"
-    js += " + link1 + ' | '\n"
-    js += " + link2 + ' | '\n"
-    js += " + link3 + ' | '\n"
-    js += " + link4 + ' | '\n"
-    js += " + link5 + ' | '\n"
-    js += " + link6 + ' | '\n"    
-    js += " + link7\n"    
-    js += " + '</div>'\n"
-    js += " + '<br /><br />'\n"
-    js += ");\n"
-        
-    f = open(outputPath,'w')
-    f.write(js)
-    f.close()
-    setProtectionFile(outputPath, 02775)
-            
-
-#=================================================================================================
-
-def createMessageOfDayJavaScriptFile(outputPath):
-    js  = "/****************************/\n"
-    js += "/*** User stuff goes here ***/\n"
-    js += "/****************************/\n\n"
-    js += "theDate    = '9th May 2006';\n"
-    js += "theMessage = 'This is the start up message of the day. The JavaScript file that created this is located in: <br />' + '" + outputPath + "';\n"
-    js += "level      = 1;\n\n"
-    js += "/*************************************************/\n"
-    js += "/*** Do not touch what follows unless you want ***/\n"
-    js += "/*** to change the web page appearance         ***/\n"
-    js += "/*************************************************/\n\n"    
-    js += "if (level==0){\n"
-    js += "   message_style = 'font-family:arial;font-size:10pt;font-weight:plain;color:black;background-color:#cccccc';\n"
-    js += "}\n"
-    js += "else{\n"
-    js += "   message_style = 'font-family:arial;font-size:10pt;font-weight:bold;color:red;background-color:#cccccc';\n"
-    js += "}\n"
-    js += "document.write(\n"
-    js += "  '<div align=\"center\">'\n"
-    js += "+ '   <table width=\"600\" border=\"0\" cellspacing=\"0\" cellpadding=\"4\">'\n"
-    js += "+ '      <tr align =\"center\" style=\"' + message_style + '\">'\n"
-    js += "+ '         <td>' + theDate + '</td>'\n"
-    js += "+ '      </tr>'\n"
-    js += "+ '      <tr align=\"left\" style=\"' + message_style + '\">'\n"
-    js += "+ '         <td>' + theMessage + '</td>'\n"
-    js += "+ '      </tr>'\n"
-    js += "+ '   </table>'\n"
-    js += "+ '</div>'\n"
-    js += ");\n"
-        
-    f = open(outputPath,'w')
-    f.write(js)
-    f.close()
-    setProtectionFile(outputPath,02775)
-    
-#=================================================================================================
-
-def messageOfTheDay(resBaseDir):
-    theScript = 'messageOfTheDay.js'
-    fullPath  = os.path.join(resBaseDir,theScript)
-    if not os.path.exists(fullPath):
-        pass
-        # createMessageOfDayJavaScriptFile(fullPath)
-        
-    return '<script src="' + theScript + '"></script>' + endLine() + breakTag()
-
-#=================================================================================================
-
-def currentStatusNX0(resBaseDir):
-    theScript = 'currentStatusNX0.js'
-    fullPath  = os.path.join(resBaseDir,theScript)
-    if not os.path.exists(fullPath):
-        pass
-        
-    return '<script src="' + theScript + '"></script>' + endLine() + breakTag()
-
-#=================================================================================================
-
-
-def currentStatusN0X(resBaseDir):
-    theScript = 'currentStatusN0X.js'
-    fullPath  = os.path.join(resBaseDir,theScript)
-    if not os.path.exists(fullPath):
-        pass
-        
-    return '<script src="' + theScript + '"></script>' + endLine() + breakTag()
-
-#=================================================================================================
-
-def generateTableForBuild(paths, releasesRun):
-    """Accepts a list of releases run for the same build. Returns the HTML table code."""  
-
-    resBase        = paths.resultsDirs['resBase/']
-    branchesToVeto = paths.branchesNotToDisplay
-    platform       = paths.platform
-    releaseType    = paths.releaseType
-  
-    if len(releasesRun)==0: return ''
-
-    # get the releases by branch
-    releasesByBranch = {}
-    [releasesByBranch.setdefault(rel['branch'],[]).append(rel) for rel in releasesRun]
-    
-    # begin a new table
-    tab  = center('begin') + table(width='850')
-
-
-    # we only want to show the branches not in the veto list passed to this method
-    branchesToDisplay = {}
-    for branch in releasesByBranch:
-        if branch not in branchesToVeto:
-            branchesToDisplay[branch] = releasesByBranch[branch]
-        else:
-            logger.warning('Branch ' + str(branch) + ' is vetoed, and will thus not be displayed on the web pages.')
-
-    sortedBranchesToDisplayKeys = fixSortBranches(branchesToDisplay.keys())
-    
-    for theBranch in sortedBranchesToDisplayKeys:
-
-        hardCoded = HardCodedFactory(logger).create(paths.releaseType,
-                                                    paths.runType)
-        branchRowHeaderText = hardCoded.branchToWebPageText(theBranch)
-        
-        tab += tableRow(styleClass='level1TableHeader',colSpan=7, tableCells=[{'content':branchRowHeaderText,'align':'left'}])
-
-        # now add the column titles for the releases within this branch
-        theCells = [{'content':'Release'},{'content':'Kit/Release'},{'content':'Platform'},{'content':'Started<font color="yellow"> (GMT)</font>'},{'content':'Completed<font color="yellow"> (GMT)</font>'},{'content':'NICOS'},{'content':'Status'}]
-        tab += tableRow(styleClass='level2TableHeader', tableCells=theCells)
-
-        # add a table row for each of the valid result directories within this branch
-
-        #    index = 1
-        for release in branchesToDisplay[theBranch]:
-            # get details from the status file
-            fullPath   = join(resBase,release['build'],release['branch'],release['runType'],release['release'])
-            statusFile = join(fullPath, "status.txt")
-            
-            print jobStatusFromFile("Started", fullPath, statusFile)
-
-            startStatus, startDate, startTime, startPlatform, startRunType, startNICOStime = jobStatusFromFile("Started", fullPath, statusFile)
-            doneStatus, doneDate, doneTime, donePlatform, doneRunType, doneNICOSdate       = jobStatusFromFile("completed", fullPath, statusFile)
-
-            startDetails = ''
-            doneDetails  = ''
-
-            if startStatus == "unknown":
-                startDetails = 'n/a'
-            else:
-                startDetails = startDate + ' @ ' + startTime
-
-            if doneStatus == "unknown":
-                doneDetails = 'n/a'
-            else:
-                doneDetails = doneDate + ' @ ' + doneTime
-     
-            # add this release to the table
-            page2path             = join(release['build'], release['branch'], release['runType'], release['release'], "page2.html")
-            jsRunInfoRelativePath = join(release['build'], release['branch'], release['runType'], release['release'], "runInfo.js")
-            jsRunInfoAbsPath      = join(fullPath,'runInfo.js')
-            javaScriptLink = '<script src="' + jsRunInfoRelativePath  + '"></script>'
-            
-            if not os.path.exists(jsRunInfoAbsPath):
-                javaScriptLink = 'n/a'
-
-            nicos     = NICOS(paths)
-
-            # convert the time (a string giving as time.time() is read in as a
-            # string. convert to int
-            try:
-                startNICOSdate = RTTDateTime(float(startNICOStime))
-            except Exception:
-                startNICOSdate = RTTDateTime()
-                
-            nicosLink = nicos.toNICOSdateLink(startNICOSdate)
-
-            theCells = [{'content':link('tableLink',page2path,release['release'])},
-                        {'content':startRunType},
-                        {'content':startPlatform},
-                        {'content':startDetails},
-                        {'content':doneDetails},
-                        {'content':nicosLink},
-                        {'content':javaScriptLink,'style':'\"font-size:8pt;font-weight:bold;\"'}
-                        ]
-
-            tab += tableRow(tableCells=theCells)
-
-    tab += table(tagType='end') + center('close')
-  
-    return tab
-
-#=================================================================================================
-
-def writePage1(paths):
-    """Loops through the output subdirectories.
-    There is one subdirectory for each processed release.
-    If The subdirectory contains date and status files,
-    the overall status is displayed, and results, if any are pointed to"""
-
-    resBase        = paths.resultsDirs['resBase/']
-    branchesToVeto = paths.branchesNotToDisplay
-    platform       = paths.platform
-    releaseType    = paths.releaseType
-                              
-    # HTML code for top of page
-    topPage =  HTMLheader() + pageTitle('The ATLAS Nightly Run Time Tests')
-
-    # HTML code for top page links
-    links = page1HeaderLinks_Javascript(resBase)
-
-    # latest news section
-    latestNewsText = messageOfTheDay(resBase)
-
-    # current run status
-    currentStateNX0 = currentStatusNX0(resBase)
-    currentStateN0X = currentStatusN0X(resBase)
-    
-    # returns a sorted list (for all releases and all builds),
-    legalRunTypes = ['release','kit'] # what are we allowed to run on?
-    allReleasesRun = getAllResultDirs(resBase, paths.legals)
-
-
-    # split up this list into a dictionary whose keys are the different,
-    # unique values for 'build', and whose
-    # values are lists containing only entries with same build value. Get it?
-    nDict = {}
-    [nDict.setdefault(entry['build'],[]).append(entry)
-     for entry in allReleasesRun]
-
-    pageContent  = center('begin')
-    pageContent += table(border=0)
-
-    cells = []
-
-    pageContent += '<tr>' + endLine()
-
-    numbBuilds = 0
-    for build in nDict.keys():
-        pageContent += '<td class="TabBorderBottom TabCommon TabOff" ID="tabs" NAME="tabs" ONCLICK="TabClick(' + str(numbBuilds) + ');"><NOBR>'+ build + ' build</NOBR></td>' + endLine()
-        numbBuilds += 1
-        cells.append(generateTableForBuild(paths, nDict[build]))
-
-    pageContent += '<td CLASS="TabBorderBottom" STYLE="width: 50px;"> </td>' + endLine()
-    pageContent += '</tr>' + endLine()
-
-    pageContent += '<tr>' + endLine()
-    pageContent += '<td COLSPAN=7 class="TabContent TabActiveBackground TabActiveBorderLeftRight TabContentBottom">' + endLine()
-
-    for cellContent in cells:
-        pageContent += '<div ID="Content" NAME="Content">' + cellContent + '</div>' + endLine()
-
-    pageContent += '</td></tr>' + endLine()
-    
-    pageContent += table(tagType='end')
-    pageContent += center('end')
-
-    # HTML code for bottom of page
-    bottomPage = page1EmailLinks() + HTMLfooter()
-
-    # put it all together
-    content = topPage + links + latestNewsText + currentStateNX0 + currentStateN0X + pageContent + bottomPage
-
-    page1Path = join(resBase, "RTTpage1.html")
-    toFile(page1Path,content)
-
-    return
-
-#=================================================================================================
-
-def colorStatusText(text):
-    color1 = 'green'
-    color2 = 'red'
-    color3 = 'brown'
-    color4 = 'black'
-    
-    colorDict = {'success'          : color1,
-                 'completed'        : color1,
-                 'error'            : color2,
-                 'Started'          : color2,
-                 'queued'           : color3,
-                 'running'          : color3,
-                 'jobTimedOut'      : color2,
-                 'PBSfatalError'    : color2,
-                 'operatorCloseDown': color2
-                 }
-
-    if text in colorDict.keys():
-        return '<font color=\"' + colorDict[text] + '\">' + text + '</font>'
-
-    return '<font color=\"' + color4 + '\">' + text + '</font>'
-
-#=================================================================================================
-
-def colorText(color,text):
-    return '<font color=\"' + color + '\">' + text + '</font>'
-
-#=================================================================================================
-
-def jobStatsHTMLcode(jobs):
-    """Given the list of jobs in a job group, returns a dictionary holding statistics for those jobs."""
-
-    numbTotalJobs = len(jobs)    
-
-    doneJobs = completedJobs(jobs) # returns a tuple
-    numbDoneJobs = len(doneJobs)
-
-    okJobs = completedOKJobs(doneJobs) # returns a tuple
-    numbOKJobs = len(okJobs)
-    
-    stats = {}
-    stats['numjobs'] = str(numbTotalJobs)
-    stats['done']    = str(numbDoneJobs)
-
-    if numbDoneJobs <= 0:
-        stats['ok'] = 'n/a'
-        stats['testok'] = 'n/a'
-    else:
-        stats['ok']      = str(numbOKJobs)   + '/' + str(numbDoneJobs)        
-        stats['testok'] = 'n/a'
-
-        # only look at the first job in the list as all
-        # RTT jobs in the same group have the same checks
-        firstJob = jobs[0]
-
-        if firstJob.hasChecks():
-            if len(firstJob.checks)>1 or len(firstJob.tests)>0: # we do not count the simple log checker
-                testsOK = completedOKChecks(okJobs) # returns a tuple
-                numbTestsOK = len(testsOK)
-
-                stats['testok'] = str(numbTestsOK) + '/' + str(numbOKJobs)
-        
-    return stats
-
-#=================================================================================================
-
-def completedJobs(jobs):
-    """Returns a tuple of jobs that are marked as done."""
-
-    doneJobs = []
-    for job in jobs:
-        if job.done:
-            doneJobs.append(job)
-
-    return tuple(doneJobs)
-
-#=================================================================================================
-    
-def completedOKJobs(jobs):
-    """Assumes that it gets passed a tuple of donejobs.
-    Returns a tuple containing those jobs that completed ok."""
-
-    okJobs = []
-    for job in jobs:
-        if job.status.primary()=='success':
-            okJobs.append(job)
-
-    return tuple(okJobs)
-
-#=================================================================================================    
-
-def completedOKChecks(jobs):
-    """Assumes it receives a tuple of jobs that completed ok"""
-    """Returns a tuple of those jobs for whom all tests completed ok."""
-
-    testsOK = []
-    for job in jobs:
-        if job.status.secondary()=='success':
-            testsOK.append(job)
-
-    return tuple(testsOK)
-
-#=================================================================================================    
-        
-def writePage3(group,jobs,paths,page3Path):
-    
-    tableEntries = ''	
-    #    coloriser = Coloriser32("success", "error")
-
-    for job in jobs:
-        
-	ref = 'n/a'
-	dataSet="n/a"
-
-	if isAthenaJob(job.jDescriptor):
-            if job.hasRefFiles():
-	        ref = job.jDescriptor.refRelease+"-"+paths.build
-            if job.jDescriptor.hasData():
-                dataSet = ''
-                for datasetList in job.jDescriptor.datasets.values():
-                    for file in datasetList:
-                        dataSet += os.path.basename(file.name)+'<br/>'
-
-        # where is page 4 relative to this page (page3)?
-        relativePage4path = join(job.jDescriptor.name,str(job.jDescriptor.jobSerialNumber),'page4.html')
-        
-        cells = [{'content':link('tableLink',relativePage4path,job.jDescriptor.identifiedName)},{'content':dataSet}]
-
-        statii = job.status.status()
-        coloredStatii = []
-
-        if len(statii.split())==1: statii += ' n/a'
-        for thing in statii.split()[0:2]:
-            cells.append({'content':colorStatusText(thing)})
-
-        cells.append({'content':ref})
-        tableEntries += tableRow(tableCells=cells)
-        
-        if job.jDescriptor.mode != 'TransparentDummy':
-            absolutePage4path = str(RTTpath(job.jDescriptor.resPath,
-                                            'page4.html')) 
-            # absolutePage4path = join(paths.resultsDirs['resBase/build/branch/runType/release'],group,relativePage4path)
-            writePage4(paths, job, absolutePage4path)
-
-    stats = jobStatsHTMLcode(jobs)
-
-    # HTML for top of page
-    #    pageTitleText = paths.release + ' ' + paths.build +  ' (' + paths.branch + ' branch)' + ': ' + group
-
-    pageTitleText = paths.release + ' ' + paths.runType + paths.kitID + ' run --- ' + paths.build + ', ' + paths.branch + ' branch --- ' + paths.platform + ' platform: ' + group    
-    startPage =  HTMLheader() + pageTitle(pageTitleText)
-
-    # HTML for stats table at top of page
-    cellTitles  = [{'content':'Total jobs'},{'content':'Done jobs'},{'content':'Jobs completing OK'},{'content':'Jobs with all post-run tests OK'}]
-    statsTable  = center('begin')
-    statsTable += table('begin')
-    statsTable += tableRow(styleClass='level1TableHeader', tableCells=cellTitles)
-    statsTable += tableRow(tableCells=[{'content':stats['numjobs']},{'content':stats['done']},{'content':stats['ok']},{'content':stats['testok']}])
-    statsTable += table('end')
-    statsTable += center('end') + breakTag()
-
-    # HTML code for the list of jobs table
-    cellTitles  = [{'content':'Job Name'},{'content':'Data Set'},{'content':'Job Status'},{'content':'Combined Checks'},{'content':'Ref. Release'}]
-    jobsTable   = center('begin')
-    jobsTable  += table('begin')
-    jobsTable  += tableRow(styleClass='level1TableHeader',tableCells=cellTitles)
-    jobsTable  += tableEntries
-    jobsTable  += table('end')
-    jobsTable  += center('end')
-    
-    # HTML for bottom page
-    backlink = breakTag() + center('begin') + link('pageLink','../../../../../../RTTpage1.html','Back to home page') + center('end') + endLine()
-    endPage = backlink + HTMLfooter()
-    
-    content = startPage + statsTable + jobsTable + endPage
-    toFile(page3Path,content)
-
-    return 
-    
-#=================================================================================================
-
-def writePage4(paths, jobMinder, page4Path):
-
-    from Tools import fileNameTrim, getKeepFileType
-    from time import sleep
-    
-    keepFileTypes = {
-        "Log Files:":['log','elog','.py','.log','lsaltF.txt'],
-        "Job Report Files:":['failureReport.html'],
-        "Results Files:":['.hist','.ps','.root','.hbook','.xml'],
-        "Other Files:":[]
-        }
-    
-    keepFilesSortedByType = {"Log Files:":[],"Job Report Files:":[],"Results Files:":[],"Other Files:":[]}
-    
-    someKeepFilesAvailable = False
-    
-    # sort the keep files into the different categories
-    
-    for thing in jobMinder.keepFiles.values():
-        
-        file = thing['keepFileString'] # recall that keep file values are in a dictionary 
-        if os.path.exists(file):
-	    head,tail=os.path.split(file)
-            fileNameTrimmed = fileNameTrim(tail,jobMinder.jDescriptor.log[:-4])
-            keepFilesSortedByType[getKeepFileType(fileNameTrimmed,keepFileTypes)].append(thing)
-            someKeepFilesAvailable = True
-
-    keepFileList = 'No Files Available'
-    
-    if someKeepFilesAvailable:
-        keepFileList = table(tagType='begin',border=0)
-        
-        for category in keepFilesSortedByType.keys():
-            keepFilesSortedByType[category].sort()
-
-            #only display a category if it has files in it
-            if keepFilesSortedByType[category]:
-                
-                keepFileList += tableRow(styleClass='keepFilesHeader',tableCells=[{'content':category}]) + endLine()
-
-                for thing in keepFilesSortedByType[category]:
-                    file = thing['keepFileString']
-                    infoString = thing['infoString']
-
-                    head,tail=os.path.split(file)
-                    fileNameTrimmed = fileNameTrim(tail,jobMinder.jDescriptor.log[:-4])
-                    infoString.strip()
-                    if len(infoString)==0:# no info string for this file
-                        keepFileList += tableRow(tableCells=[{'content':link('tableLink',tail,fileNameTrimmed)}])
-                    else:
-                        keepFileList += tableRow(tableCells=[{'content':linkWithInfoString('tableLinkWithInfo',tail,fileNameTrimmed,infoString)}])
-
-        keepFileList += table('end')    
-   
-    # HTML code for top of page
-    pageTextTitle = 'ATLAS RTT --- ' + paths.runType + paths.kitID + ' run, ' + paths.platform + ' platform'
-    topPage =  HTMLheader() + pageTitle(pageTextTitle)
-
-
-
-    # make the checks summary table
-    statii = jobMinder.status.status()
-    coloredStatii = []
-
-    summaryCells=[{'content:':'n/a'},{'content':'n/a'}]
-
-    if len(statii.split())>1:
-        index=0
-        for thing in statii.split()[0:2]:
-            if thing.find('success')!=-1:
-                summaryCells[index]['content']='True'
-            elif thing.find('error')!=-1:
-                summaryCells[index]['content']='False'
-
-            index += 1
-                
-    summaryCellTitles  = [{'content':'Job completed OK'},{'content':'All post-run tests OK'}]
-    summaryTable  = center('begin')
-    summaryTable += table('begin')
-    summaryTable += tableRow(styleClass='level1TableHeader', tableCells=summaryCellTitles)
-    summaryTable += tableRow(tableCells=summaryCells)
-    summaryTable += table('end')
-    summaryTable += center('end') + breakTag()
-    
-    # HTML code job table
-    cellTitles = [{'content':'Job Name'},{'content':'Release'},{'content':'Build'},{'content':'Branch'},{'content':'Keep files'}]
-    jobTable   = center('begin')
-    jobTable  += table('begin')
-    jobTable  += tableRow(styleClass='level1TableHeader',tableCells=cellTitles)
-    theCells   = [{'content':jobMinder.jDescriptor.identifiedName}, {'content':paths.release}, {'content':paths.build}, {'content':paths.branch}, {'content':keepFileList}]
-    jobTable  += tableRow(tableCells=theCells)
-    jobTable  += table('end')
-    jobTable  += center('end')
-    
-    # HTML code for bottom page
-    bottomPage  = breakTag()
-    bottomPage += center('begin')
-    bottomPage += link('pageLink','../../../../../../../../RTTpage1.html','Back to home page') + linkSep()
-    bottomPage += link('pageLink','../../../../page2.html','Back to packages page')
-    bottomPage += center('end')
-    bottomPage += HTMLfooter()
-    
-    content = topPage + summaryTable + jobTable + bottomPage
-    
-    toFile(page4Path,content)
-
-#=================================================================================================
-
-def reporterMessage2html(message):
-    if message.find('@') == -1:
-        message = message.replace(' ', '&nbsp;')
-        message = message.replace('\n','<br/>')        
-        return '<table><tr class=\"reporterValueEntry\"><td>' + message + '</td></tr></table>'
-
-    params = []
-    values = []
-
-    arrayIndex = 0
-    currentIndex = 0
-
-    while not currentIndex >= len(message):
-        nextAtSymbol = message.find('@',currentIndex,len(message))
-        valueEndIndex = indexOfValueEnd(message,nextAtSymbol+1)
-
-        params.append(message[currentIndex:nextAtSymbol])
-        values.append(message[nextAtSymbol+1:valueEndIndex])
-
-        currentIndex = valueEndIndex + 1
-        arrayIndex += 1
-
-    index = 0
-    for param in params:
-        params[index] = param.replace(' ','&nbsp;')
-        params[index] = params[index].replace('\n','<br/>')
-        index += 1
-
-    index = 0
-    for value in values:
-        values[index] = value.replace(' ','&nbsp;')
-        values[index] = values[index].replace('\n','<br/>')
-        index += 1          
-    
-    htmlCode = '<table width=\"90\%\" border=\"0\" cellpadding=\"0\" cellspacing=\"3\">'
-    index = 0
-    for param in params:
-        htmlCode += '<tr><td width=\"20\%\" bgcolor=\"#bbbbff\" class=\"reporterParamEntry\" valign=\"top\">' + params[index] + '</td><td  bgcolor=\"#ccccff\" class=\"reporterValueEntry\" valign=\"top\">' + values[index] + '</td></tr>'
-        index += 1
-    htmlCode += '</table>'
-
-    return htmlCode
-
-#=================================================================================================
-
-def indexOfValueEnd(message,currAtSymbol):
-    nextAtSymbol = message.find('@',currAtSymbol,len(message))
-
-    if nextAtSymbol == -1:
-        return len(message)
-    else:
-        lastSlashNsymbol = message.rfind('\n',currAtSymbol,nextAtSymbol)
-
-        if lastSlashNsymbol != -1:
-            return lastSlashNsymbol
-        else:
-            subString = message[currAtSymbol:nextAtSymbol]
-            subString.rstrip()
-            lastSlashNsymbol = subString.rfind(' ')
-            if lastSlashNsymbol != -1:
-                return lastSlashNsymbol
-            else:
-                return currAtSymbol + (len(subString)/2)
-
-#=================================================================================================
-# from Borg import Borg
-
-# class Page2Writer(Borg):
-class Page2Writer:
-    def __init__(self, noPackagePaths=None,rttTagInRelease='',init=False):
-        """bool argument prevents overwrite of the singleton dictionary
-        if it is not intended to (re)initialise the object.
-        The Paths passed = instance of NoPackagePaths created in Tester."""
-        
-        # Borg.__init__(self)
-
-
-        # if not init: return
-    
-        
-        self.knownTestRuns   = {}
-        self.noPackagePaths  = noPackagePaths
-        self.rttTagInRelease = rttTagInRelease
-
-    # =================================================================================================
-    
-    def addTestRuns(self,tr=[]):
-        """Add a list of testruns to loop over and display."""
-        order = 0
-        for thing in tr:
-            try:
-                self.knownTestRuns[thing.name] = thing
-            except Exception, e:
-                logger.warning("Problem trying to add TestRun with name " + str(thing.name) + " to Page2Writer's list of known TestRuns. Skipping request...")
-                logger.warning("Problem was: " + str(e))
-
-    # =================================================================================================
-
-    def formatErrorMessageForWebDisplay(self,error,cutOffPoint=60):
-        errorToks   = error.split()
-        finalToks = []
-        for tok in errorToks:
-            if len(tok) <= cutOffPoint:
-                finalToks.append(tok)
-                continue
-
-            lines = self.wrapString(tok)                        
-            [finalToks.append(item) for item in lines]            
-
-        formattedError = ''
-        lineLength = 0
-        for thing in finalToks:
-            thing = thing + ' '
-            lineLength += len(thing)
-            if lineLength < cutOffPoint:
-                preString = ''
-            else:
-                preString  = '<br/>'
-                lineLength = len(thing)
-                
-            formattedError += preString + thing
-
-        return formattedError
-
-    # =================================================================================================
-
-    def wrapString(self,theString,maxLineLength=60):
-        lines = []
-        remainder = len(theString)
-        if remainder <= maxLineLength:
-            lines = [theString]
-        else:            
-            forwardSlashChar = theString.find('/')
-            if forwardSlashChar != -1:
-                lines = self.wrapPathString(theString)
-            else:
-                index = 0
-                start = 0
-                end   = 0
-                while remainder > maxLineLength:
-                    start = index*maxLineLength
-                    end   = start + maxLineLength
-                    lines.append(theString[start:end])
-                    index += 1
-                    remainder = len(theString[end:])
-                lines.append(theString[end:])
-                
-        return lines
-
-    # =================================================================================================
-    
-    def wrapPathString(self,thePath,maxLineLength=60):
-        forwardSlash = thePath.find('/')
-        if forwardSlash == -1:
-            return [thePath]
-
-        done  = False
-        start = 0
-        end   = maxLineLength
-        lines = []
-
-        while not done:
-            subString = thePath[start:end]
-            slashIndex = subString.rfind('/')
-
-            if slashIndex == -1:
-                stringToAppend = subString
-                lines.append(stringToAppend)
-                start = end
-            else:
-                stringToAppend = subString[0:slashIndex+1]
-                lines.append(stringToAppend)
-                start = start + slashIndex + 1
-
-            
-            end   = start + maxLineLength
-            if start >= len(thePath):
-                done = True
-
-        return lines    
-    
-    # =================================================================================================
-    
-    def write(self):
-    
-        """Loops over a list of testrun objects, interrogating them for certain information to display on the web page.
-        HTML code for the page is output to file in the results base directory."""
-
-        from Tools import now
-
-        #        if len(self.knownTestRuns) == 0:
-        #   return
-        
-        # table titles and entries
-        tableTitles  = ''
-        tableEntries = ''
-
-        resPath = ''
-        rttTag  = ''
-
-        # names of all the test runs
-        trNames = self.knownTestRuns.keys()
-
-        # now let's sort alphabetically the testrun names, so that page2 is nice and easy to navigate
-        trNames.sort()
-
-        for testRunName in trNames:
-
-            theCells = [{'content':'Package'},{'content':'Container'},{'content':'Status'},{'content':'JobGroups in package'}]
-            tableTitles =  tableRow(styleClass='level1TableHeader',tableCells=theCells)
-            theTestRun  = self.knownTestRuns[testRunName]
-            paths       = theTestRun.paths
-            jobMinders  = theTestRun.jobMinders
-            groupKits   = theTestRun.groupKits
-
-            # results path
-            resPath    = paths.resultsDirs[
-                'resBase/build/branch/runType/release']
-    
-            # set up an initial row for this testrun
-            jobGroupTable = 'n/a'
-
-            # sort by job group
-            byGroup = {}
-            for job in jobMinders:
-                byGroup.setdefault(job.jDescriptor.jobGroup, []).append(job)
-
-            # sort the keys
-            byGroupKeysSorted = byGroup.keys()
-            byGroupKeysSorted.sort()
-
-            if len(byGroupKeysSorted)>0:
-                jobGroupTable  = table(border=0)
-                #   theCells = [{'content':'Name'},{'content':'Summ. Files'},{'content':'Done jobs'},{'content':'Jobs OK'},{'content':'Tests OK'}]
-                theCells = [{'content':'Name'},{'content':'Done jobs'},{'content':'Jobs OK'},{'content':'Tests OK'}]
-                jobGroupTable += tableRow(styleClass='level2TableHeader',tableCells=theCells)
-
-                for group in byGroupKeysSorted:
-
-                    # page 3 goes into .../package/group        
-                    # the full path will be to the run path .../package/group/joboptions/id#
-                    # this hardwiring of where the pages will be written is extrmeely package-centric
-                    # and should be generalised for non-package oriented descriptors
-                    relativePage3path = 'page3.html'
-                    relativePage3path = join(testRunName, group, relativePage3path)
-                    page3link = link('tableLink',relativePage3path,string.replace(group, 'Athena', 'Athena: '))
-            
-                    absolutePage3path = join(resPath,relativePage3path)
-                    writePage3(group,byGroup[group],paths,absolutePage3path)
-            
-                    # add the summary files to the header.
-                    # ask the group kit instances for the display files
-                    # This list was updated when the postScripts were run,
-                    # so it is essential to use the same instance as that which was used to run the postScripts
-            
-                    #                    statusText = colorStatusText('completed')
-                    # startedStatusText = colorStatusText('Started')                
-
-                    # summaryFiles = ''
-
-                    # when a all jobs in a Group terminate, a group kit is put into the groupKit
-                    # dictionary. Testing this to decide if the the group has completed.
-
-                    #if group in groupKits.keys():
-                    #    if groupKits[group].displayFiles:
-                    #        for file in groupKits[group].displayFiles:
-                    #            fullpath = os.path.join(resPath, file)
-                    #            if os.path.exists(fullpath):                    
-                    #                summaryFiles += link('',file,file) + breakTag() + endLine()                    
-
-                    #if summaryFiles == '': summaryFiles = 'n/a'
-
-                    stats = jobStatsHTMLcode(byGroup[group]) # returns a dictionary
-
-                    # let us color stats text accordingly
-                    # if any jobs have failed to complete, or if some checks failed for
-                    # jobs that otherwise were successful, we color red.
-
-                    # set to true if one of the stats column is turned red
-                    problemsNoted = False
-            
-                    # let's peek first at the jobs completed ok stat
-                    okJobs  = str(stats['ok'])
-                    tokens = okJobs.split('/')
-                    if okJobs != 'n/a' and tokens[0] != tokens[1]: #color red then
-                        stats['ok'] = colorText('red',stats['ok'])
-                        problemsNoted = True
-                
-                    # now let's peek first at the tests completed ok stat for those jobs completed ok
-                    okTests = str(stats['testok'])
-                    tokens = okTests.split('/')
-                    if okTests!= 'n/a' and tokens[0] != tokens[1]: #color red then
-                        stats['testok'] = colorText('red',stats['testok'])
-                        problemsNoted = True
-
-                    # if one of the above two stats was turned red, then
-                    # must color the status text red if it is set to "completed"
-                    #                    if problemsNoted and statusText.find('completed')!=-1: 
-                    #   statusText = colorText('red','completed')
-
-                    doneJobs = stats['done'] + '/' + stats['numjobs']
-                    
-                    jobGroupTable += tableRow(tableCells=[{'content':page3link},
-                                                          #  {'content':summaryFiles},
-                                                          {'content':doneJobs},
-                                                          {'content':stats['ok']},
-                                                          {'content':stats['testok']}])                    
-
-                jobGroupTable += table('end')    
-
-
-            if theTestRun.status() == theTestRun.ps.failureState:
-                jobGroupTable  = table(border=0,width='500')
-
-                error = self.formatErrorMessageForWebDisplay(
-                    str(theTestRun.getError()))
-                
-                jobGroupTable += tableRow(tableCells=[{'content':error}])
-                jobGroupTable += table('end')
-                firstCell = theTestRun.name+'<br/>'+(theTestRun.paths.packageTag).replace(theTestRun.name+'-','')
-                theCells  = [{'content':firstCell,'color':'#dd9999'},
-                             {'content':theTestRun.paths.containerPackage},
-                             {'content':theTestRun.status(),'color':'#dd9999'},
-                             {'content':jobGroupTable}
-                             ]
-                tableEntries += tableRow(tableCells=theCells)
-            else:                
-                firstCell = theTestRun.name+'<br/>'+(theTestRun.paths.packageTag).replace(theTestRun.name+'-','')
-                theCells = [{'content':firstCell},
-                            {'content':theTestRun.paths.containerPackage},
-                            {'content':theTestRun.status()},
-                            {'content':jobGroupTable}
-                            ]
-                tableEntries += tableRow(tableCells=theCells)
-                
-        
-        # open the file to write out the html file with the status        
-        page2Path = join(self.noPackagePaths.resultsDirs['resBase/build/branch/runType/release'], "page2.html")
-        
-        
-        # HTML code top page
-        # NOTE: paths.kitID is an empty string if this is a release run
-        # We use noPackagePaths (passed to the Page2Writer constructor) rather than paths, so that this code still works even if there are no TestRuns in the list
-        pageTitleString  = self.noPackagePaths.release + ' '
-        pageTitleString += self.noPackagePaths.runType + self.noPackagePaths.kitID + ' run --- '
-        pageTitleString += self.noPackagePaths.build + ', ' + self.noPackagePaths.branch + ' branch --- '
-        pageTitleString += self.noPackagePaths.platform + ' platform'
-
-        srcRun   = self.noPackagePaths.rttSrcDir
-        libTools = self.libToolsFrom()
-
-        len1 = len('RTT src code run: ')
-        len2 = len('RttLibraryTools module from: ')        
-        firstHalfLine1  = '&nbsp;'*(len2-len1) + 'RTT src code run: '
-
-        len1 = len(str(srcRun))
-        len2 = len(str(libTools))        
-        secondHalfLine1 = str(srcRun) + '&nbsp;'*(len2-len1)
-        
-        firstLine  = firstHalfLine1 + secondHalfLine1
-        secondLine = 'RttLibraryTools module from: ' + str(libTools)
-        
-        rttVersionsString  = firstLine + '<br/>'  
-        rttVersionsString += secondLine
-
-        packagesFoundString = 'Packages found by RTT: ' + str(len(self.knownTestRuns))
-        
-        # create the html code for the top of the page        
-        topPage =  HTMLheader() + pageTitle(pageTitleString) + sectionTitle(rttVersionsString) + pageTitle(packagesFoundString)
-
-        # create the RTT Run Summariser link, if the file exists
-        if os.path.exists(os.path.join(self.noPackagePaths.resultsDirs['resBase/build/branch/runType/release'],'RTTSummary.txt')):
-            topPage += center('begin') + link('pageLink','RTTSummary.txt','Current Run Summary') + center('end') + breakTag(2)
-        
-        # HTML code group table
-        groupTable  = center('begin')
-        groupTable += table('begin')
-        groupTable += tableTitles
-        groupTable += tableEntries
-        groupTable += table('end')
-        groupTable += center('end')
-        
-        # HTML code bottom page
-        bottomPage  = breakTag()
-        bottomPage += center('begin')
-        bottomPage += link('pageLink','../../../../RTTpage1.html','Back to home page')
-        bottomPage += center('end')
-        bottomPage += HTMLfooter()
-        
-        content = topPage + groupTable + bottomPage
-        toFile(page2Path,content)        
-        
-        return
-    
-    # =================================================================================================
-
-    def libToolsFrom(self):
-        isFrom = "release based (" + self.rttTagInRelease + ")"
-        if self.noPackagePaths.rttLibDir:
-            isFrom = self.noPackagePaths.rttLibDir
-
-        return isFrom
-                
-    # =================================================================================================
-    
-    def getStatusFile(self, paths):
-        resPath=paths.resultsDirs['resBase/build/branch/runType/release']
-        statusFile = join(resPath, "status.txt")
-
-        if(os.path.exists(statusFile)): return statusFile
-
-        logger.error("Could not find status file!")
-        return None                
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/HardCoded.py b/Tools/RunTimeTester/src/HardCoded.py
deleted file mode 100755
index 8bfa5744d0709952ff1757716ccd2ff1d2261812..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/HardCoded.py
+++ /dev/null
@@ -1,102 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from os.path import join, basename, normpath
-
-class HardCodedBase:
-    def __init__(self):
-        self.installArea      = 'InstallArea'
-        self.shareArea        = 'share'
-        self.leafStampDir     = 'AtlasOffline'
-        self.cmtArea          = 'cmt'
-        self.binArea          = 'bin'
-        self.limitedNightlies = ['atlrel_0', 'atlrel_3']
-        self.projectsWithReleaseName = ['AtlasOffline','AtlasProduction']
-        
-    def installAreaDirName(self):
-        return self.installArea
-
-    def binAreaDirName(self):
-        return self.binArea
-    
-    def shareAreaDirName(self):
-        return self.shareArea
-   
-    def cmtAreaDirName(self):
-        return self.cmtArea
-
-    # noticed by PyLint
-    # commented out as attribute not seat and method never called.
-    # def projectWithRTT(self):
-    #    return self.RTTproject
-
-    def projectWithRecExCommon(self):
-        return 'AtlasReconstruction'
-
-    def relativeRecExCommonLinks(self,projectShareArea):
-        release = basename(normpath(join(projectShareArea,'../..')))
-        return normpath(join(projectShareArea,
-                             '../../../..',
-                             self.projectWithRecExCommon(),
-                             release,
-                             self.installAreaDirName(),
-                             self.shareAreaDirName(),
-                             self.binAreaDirName(),
-                             'RecExCommon_links.sh'))
-    
-    def nightlyDirIfAny(self, branch, isNightly):
-        """return the nightly offset (eg nightlies/dev) if the branch is in
-        the dictionary, otherwise retuens an empty string. Branch is set
-        to Numbered if the reelase is numbered, and as this is not in the
-        dictionary, numbered releases return an empty string"""
-
-        if not isNightly: return ''
-        # branch must be in the dictionaries - otherwise RTT error
-        return self.getNightlyPathFragment(branch)
-
-    def pathFragmentToReleaseNotes(self,release):
-        return join(release,'cmt/ReleaseNotes.txt')
-    
-    def getNightlyPathFragment(self, branch):
-        pass
-    
-    def relativeRelease(self, branch, release):
-        return join(self.getNightlyPathFragment(branch), release)
-
-    def getLatestCopiedReleaseLinkName(self, key):
-        pass
-        
-# ==========================================================================
-
-class HardCodedProjectRelease(HardCodedBase):
-
-    def __init__(self):
-        HardCodedBase.__init__(self)
-        # self.distBase    = '/afs/cern.ch/atlas/software/builds'
-        self.rttProject  = 'AtlasCore' # package containing the RunTimeTester package
-        
-    def getNightlyPathFragment(self, branch):
-        if branch is 'Numbered': return ''
-        return 'nightlies/%s' % branch.strip()
-        
-    def getLatestCopiedReleaseLinkName(self, key):
-        d = {'i686'  : '32B', 'x86_64' : '64B',
-             'slc4'  : 'S4' , 'slc5'   : 'S5',
-             'dbg'   : 'Dbg', 'opt'    : 'Opt',
-             'gcc43' : 'G4' , 'gcc34'  : ''}
-        i, os, gcc, bin = key[0].split('-')
-        return 'latest_copied_release%s%s%sAll%s' % (d[i], d[os], d[gcc], d[bin])
-
-# ==========================================================================
-
-class HardCodedProjectKit(HardCodedBase):
-    def __init__(self):
-        HardCodedBase.__init__(self)
-        self.nightlyPathFragment = {'Numbered':'',
-                                    'val':'',
-                                    'bugfix':'',
-                                    'pcache':'',
-                                    'dev':''}
-
-    def nightlyStamp(self,branch):
-        return None
-
diff --git a/Tools/RunTimeTester/src/HtoPy.py.obsolete b/Tools/RunTimeTester/src/HtoPy.py.obsolete
deleted file mode 100755
index 2d6af6fade76044b258205cc6ec73b7f10f3c658..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/HtoPy.py.obsolete
+++ /dev/null
@@ -1,79 +0,0 @@
-#! /usr/local/bin/python
-"""
-Some kind of attempt to convert Together C++ output to python...
-"""
-
-import sys
-import os
-import re
-
-for iname in sys.argv[1:]:
-    head = iname.split(".")[0]
-    oname=head+".py"
-    print "inp file = "+iname
-    print "out file = "+oname
-    ifile = open(iname, "r")
-    ofile = open(oname, "w")
-
-    ofile.write("#! /usr/local/bin/python")
-    for line in ifile.readlines():
-        if line.find("Together")!=-1:
-            nline=""
-        elif line.find("#ifndef")!=-1:
-            nline=""
-        elif line.find("#define")!=-1:
-            nline=""
-        elif line.find("#endif")!=-1:
-            nline=""
-        elif line.find("public:")!=-1:
-            nline=""
-        elif line.find("private:")!=-1:
-            nline=""
-        elif line.find("};")!=-1:
-            nline=""
-        elif line.find("#include")!=-1:
-            tokens = line.split('"')
-            token=tokens[1]
-            tokens=token.split(".")
-            nline="from "+tokens[0]+" import "+tokens[0]+"\n"
-        elif line.find("/**")!=-1:
-            nline=""
-        elif line.find("*/")!=-1:
-            nline=""
-        elif line.find("* ")!=-1:
-            tokens = line.split("*")
-            nline ="# "
-            if len(tokens)!=1: nline=nline+tokens[1]
-            nline = nline+"\n"
-        elif line.find("class ")!=-1:
-            tokens =line.split()
-            nline="class "+tokens[1]+":"
-        elif line.find(");")!=-1:
-            start=line.find("(")
-            end=line.find(")")
-            args=line[start+1:end]
-
-            
-            tokens=line.split()
-            if len(tokens)==1:
-                funName=tokens[0][:tokens[0].find("(")]
-                if(args != ""):
-                    nline="def __init__(self, "+args+"):\n   pass\n"
-                else:
-                    nline="def __init__(self):\n   pass\n"
-            else:
-                funName=tokens[1][:tokens[1].find("(")]
-                if(args != ""):
-                    nline= "def "+funName+"(self,"+args+"):\n   pass\n"
-                else:
-                    nline= "def "+funName+"(self):\n   pass\n"
-        elif line.find("List<")!=-1:
-            tokens=line.split(">")
-            nline="list "+tokens[1]+"\n"
-        else:
-            nline=line
-
-        print "old line: "+line
-        print "new line: "+nline
-        if nline != "": ofile.write(nline)
-        
diff --git a/Tools/RunTimeTester/src/JobGroupDirectoryMaker.py b/Tools/RunTimeTester/src/JobGroupDirectoryMaker.py
deleted file mode 100755
index e0b4f2553c7994620a359b79d9ec246ae25e595c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobGroupDirectoryMaker.py
+++ /dev/null
@@ -1,207 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools        import rmdirMkdir, changePerm, mkdirIfNew
-import Tools2
-
-#===== RTT MODULES =====
-from JobGroupKits            import JobGroupKits
-from Factory_ScriptWriter    import ScriptWriterFactory
-from releaseToNum            import releaseToNum
-from getInheritanceHeirarchy import getInheritanceHeirarchy
-from formatCollection        import formatCollection
-from exc2string2             import exc2string2
-
-#===== PYTHON MODULES ==
-import os, shutil, anydbm, string
-from os.path              import join
-import shelve
-import xml.dom.ext
-import cPickle
-
-"""
-A class to make Work, Results and Run directories.
-User             calls: makeJobGroupDirs - extracts results and working
-                                           directory names (minder dependent)
-                                           
-
-makeJobGroupDirs calls: createJobGroupDir - for each of [resDir, workDir]
-                                            to actually create the directory
-
-                 calls: setupRunDir
-"""
-
-# -------------------------------------------------------------------------
-
-class JobGroupDirectoryMaker:
-    def __init__(self, descSetUpRunDirFn, logger):
-        self.logger = logger
-        self.descSetUpRunDirFn = descSetUpRunDirFn
-        
-    def createJobGroupDir(self, pathToDirectory):
-        """
-        create run and results directory trees. This function is sometimes
-        overridden by a null function in derived classes.
-        """
-
-        try:
-            rmdirMkdir(pathToDirectory)
-            self.logger.debug('making '+pathToDirectory)
-        except Exception, e:
-            errmsg = exc2string2()
-            self.logger.error(errmsg)
-
-    def makeJobGroupDirs(self, runPath, resPath):
-        """
-        makes  the run and results directories.
-        fills the run directories with necessary files before laucnching jobs 
-        """
-        
-        self.createJobGroupDir(runPath)
-        self.logger.debug('Making run directory %s', runPath)
-
-        self.logger.debug('Making results directory %s', resPath)
-        self.createJobGroupDir(resPath)
-
-    # --------------------------------------------------------------------------
-
-    def writeTestRunnerScript(self, shareArea, dbPath, runPath):
-        script = [
-            'import sys',
-            'sys.path.append("'+shareArea+'")',
-            'import shelve, RTTTestRunner',
-            'from exc2string2 import exc2string2',
-            'import RttLibraryTools',
-            'db = shelve.open("%s")' % dbPath,
-            'keys = db.keys()',
-            'print "Keys before order: %s" % str(keys)',
-            'keys.sort()',
-            'print "Keys after order: %s" % str(keys)',
-            'for k in keys:',
-            '   print "Accessing key: %s" % str(k)',
-            '   t = db[k]',
-            '   try:',
-            '      t.run()',
-            '   except:',
-            '      m = exc2string2()',
-            '      t.error=True',
-            '      t.errmsg=m',
-            '      print m',
-            ' ',
-            '   print t.testId,": ",t.result',
-            'db.close()'
-           ]
-        script = string.join(script, '\n')
-        fn = 'testRunner.py'
-        fn = os.path.join(runPath, fn)
-        of = open(fn, 'w')
-        of.write(script)
-        of.close()
-            
-            
-# --------------------------------------------------------------------------
-
-    def setupRunDir(self, jMinder):
-        
-        """
-        Creates run directory (obtained from the descriptor)
-        
-        Uses JobGroupKits to establish what is needed to be copied to it.
-        Uses ScriptWriter to write a run script to it.
-
-        This version does not write a requirements file into the run
-        directory.
-        """
-        self.logger.debug("DirectoryMakerProject: entering setupRunDir")
-        self.logger.debug("setupRunDir processing: %s" %(jMinder.identifiedName))
-
-
-        runPath = jMinder.runPath
-
-        def writeScript(fn, script):
-            ofn = os.path.join(runPath, fn)
-            of = open(ofn, 'w')
-            of.write(script)
-            of.close()
-            changePerm(ofn, 0777)
-
-        dicts = jMinder.scripts.values()
-        [writeScript(d['scriptName'], d['script']) for d in dicts]
-    
-
-        #######################################################################
-        # handle aux files (= files needed in the run dir to run the job)
-        #######################################################################
-
-
-        # if the aux file is a sub directory of the Install directory, make
-        # the corresponding directory in the work directory
-        dict = {}
-        for path in jMinder.auxFiles:
-            base, fn = os.path.split(path)
-            if base: dict[base] = fn
-
-        dirs = dict.keys()
-        dirs = [os.path.join(runPath, d) for d in dirs]
-        dirs = [mkdirIfNew(d) for d in dirs]
-        
-        auxFileDirectory = jMinder.shareArea
-
-        # explicitly remove the package XML file, we add it back later
-        auxFiles = [a for a in jMinder.auxFiles if a.strip() != os.path.basename(jMinder.confFile).strip()]
-                
-        # copy the files from the install area (sub) dir to the run (sub) dir.
-        # for f in jMinder.auxFiles:
-        for f in auxFiles:
-
-            src  = join(auxFileDirectory, f)
-            dest = join(runPath, f)
-
-            if not os.path.exists(src):
-                self.logger.error("Non-existent aux file: %s , skipping" %  src )
-                continue
-            
-            # shutil.copy(src, dest)
-            os.symlink(src, dest)
-            changePerm(dest, 0777)
-
-        # now add the package XML file
-        confFileName = os.path.basename(jMinder.confFile).strip()
-        src = jMinder.confFile.strip()
-        dst = join(runPath, confFileName)
-        os.symlink(src, dst)
-        changePerm(dst, 0777)
-                    
-
-        # write out a python script to run the RTT test wrappers
-        self.writeTestRunnerScript(jMinder.shareArea, jMinder.testDBPath, jMinder.runPath)
-
-        # Output the RTT argdict pickle file
-        pickleDict = jMinder.rttArgDictPickle
-        cPickle.dump(pickleDict['what'], open(pickleDict['where'], 'wb'))
-
-        # Now perform tasks specific to the different descriptors
-        self.descSetUpRunDirFn()
-
-        # make a note of the files in the run directory
-        # at this point - in case a rerun is needed
-        of = open(os.path.join(runPath, 'listDir.txt'), 'w')
-        [of.write(p+'\n') for p in os.listdir(runPath)]
-        of.close()
-
-        self.logger.debug("returning from setupRunDir()")
-
-        
-
-
-class JobGroupDirectoryMakerWatcher(JobGroupDirectoryMaker):
-    def __init__(self, logger):
-        JobGroupDirectoryMaker.__init__(self, logger)
-        
-    def setupRunDir(self, jMinder):
-        JobGroupDirectoryMaker.setupRunDir(self, jMinder)
-        workerRunPaths = [minder.runPath for minder in jMinder.confederation]
-        histsFile = open(join(jMinder.runPath,'histLocation.txt'),'w')
-        [histsFile.write(path+'\n') for path in workerRunPaths]
-        histsFile.close()
-
-
diff --git a/Tools/RunTimeTester/src/JobGroupKit.py b/Tools/RunTimeTester/src/JobGroupKit.py
deleted file mode 100755
index 539036a55bc1ed6a6fcafd5c053911702e72adc7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobGroupKit.py
+++ /dev/null
@@ -1,160 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection import formatCollection
-from Tools            import unique
-from Geneology        import TreeMember
-from RTTSException    import RTTCodingError
-
-import logging
-logger = logging.getLogger('rtt')
-
-class JobGroupKit(TreeMember):
-    def __init__(self, group, parent=None):
-        
-        self.group      = group
-        self.parent     = parent
-        self.logChecker = None
-
-        TreeMember.__init__(self, group, parent)
-
-        
-        self.kitItems = { 'keepFilePatterns':  [],
-                          'checkers':          [],
-                          'auxFilePatterns':   [],
-                          'actions':           [], # list of UserTestDescriptor instances
-                          # 'monActions':        [], # list of UserTestDescriptor instances
-                          'tests':             []  # list of UserTestDescriptor instances
-                          }
-
-    def __getitem__(self, what):
-        return self.getItem(what)
-    
-    def __str__(self):
-        string  ='group:       %s\n'%self.group
-        string +='parent:      %s\n'%self.parent
-        string +='log`Checker: %s\n'%str(self.logChecker)
-        string +='\n'
-        string += formatCollection(self.kitItems)
-        return string
-
-    def kitLists(self):
-        "Towards data  structure independent hiding"
-        return self.kitItems.values()
-    
-    def add(self, kit):
-        [self.kitItems[item].extend(kit.kitItems[item]) for item in self.kitItems.keys()]
-        [unique(self.kitItems[item]) for item in self.kitItems.keys()]
-        # do not overwrite the log checker if
-        if not self.logChecker: self.logChecker = kit.logChecker
-
-    def extendItem(self, item, list):
-        self.kitItems[item].extend(list)
-        unique(self.kitItems[item])
-
-    def setLogChecker(self, kits):
-        """if this kit has no log checker, set it to the one of the
-        closest ancestor"""
-
-        ancestor = self.kitFromGroup(self.parent, kits)
-        while (not self.logChecker) and ancestor:
-            self.logChecker = ancestor.logChecker
-            ancestor = self.kitFromGroup(ancestor.parent, kits)
-
-        if not self.logChecker:
-            logger.warning('No log checker set for group %s' % self.group)
-
-    def kitFromGroup(self, group, kits):
-        "Method used to find ancestors"
-        
-        for kit in kits:
-            if kit.group == group: return kit
-    
-        return None
-
-    def getItem(self, itemName):
-        if itemName in self.kitItems:
-            return self.kitItems[itemName]
-        else:
-            logger.error('Item '+itemName+' not in kit')
-            return None
-
-    def hasCheckers(self):
-        # assume that all job groups have at least one checker: the job status check
-        return len(self.kitItems['checkers'])>1 
-        
-    def equivalent(self, other):
-
-        if not hasattr(other, 'group'):      return False
-        if not hasattr(other, 'parent'):     return False
-        if not hasattr(other, 'logChecker'): return False
-        if not hasattr(other, 'kitItems'):   return False
-
-
-        if self.group  != other.group:       return False
-        if self.parent != other.parent:      return False
-
-        if not self.logChecker:
-            # either kits self and other have no log checker.... 
-            if other.logChecker:            return False
-        else:
-            if not other.logChecker:        return False
-            # or they are equivalent
-            if not self.logChecker.equivalent(other.logChecker): return False
-
-        # check the dictionary has the same keys.
-        m_keys = self.kitItems.keys()
-        o_keys = other.kitItems.keys()
-
-        if len(m_keys) != len(o_keys): return False
-
-
-        for k in m_keys:
-            if k not in o_keys: return False
-            # Each key corresponds to a list check length
-            m_list = self.kitItems[k] 
-            o_list = other.kitItems[k] 
-            if len(m_list) != len(o_list): return False
-
-
-        # the kit has lists of string and objects.
-        # the (not wonderful) check is to compare lists of strings.
-        # for lists that are already strings, use the list itself.
-        # for lists of objects use the class name of the object.
-        # this is clearly not a rigorous test (the objects can have
-        # the same classname, but different state), but it will do for now.
-        
-        stringLists = [
-            'keepFilePatterns',
-            'auxFilePatterns'
-            ]
-
-        listDict = {}
-        for k in stringLists:
-            listDict[k] = (self.kitItems[k], other.kitItems[k])
-
-        classLists = [
-            'checkers',
-            'actions',
-            # 'monActions',
-            'tests'
-            ]
-
-        for k in classLists:
-            listDict[k] = (
-                [kk.__class__.__name__ for kk in self.kitItems[k]], 
-                [kk.__class__.__name__ for kk in other.kitItems[k]]
-                )
-
-        # sanity check - have all the keys been covered?
-        if len(listDict.keys()) != len(self.kitItems.keys()):
-            raise RTTCodingError('Sanity test: bad kit equivalence test - different lengths')
-        
-        for k in listDict.keys():
-            for element in listDict[k][0]:
-                if element not in listDict[k][1]: return False
-
-        return True
-        
-if __name__ == '__main__':
-    print JobGroupKit('xxx','yyy')
-    print JobGroupKit('AthenaAtlfast')
diff --git a/Tools/RunTimeTester/src/JobGroupKits.py b/Tools/RunTimeTester/src/JobGroupKits.py
deleted file mode 100755
index 93c1169458039d34f87197e8ab377a6518097648..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobGroupKits.py
+++ /dev/null
@@ -1,65 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Geneology     import KitTree
-from RTTSException import RTTCodingError
-import copy
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class JobGroupKits:
-# class JobGroupKits:
-    """A class that contains information about all kits. Each kit
-    is represented by a dictionary. These dictionaries are stored in a
-    dictionary where they are keyed on the group name.
-
-    The class is derived from the Borg implementation of the Singleton pattern.
-    Instantiatiating the class with no argument gives a fiexed state.
-    Instantiating it with a list of kits changes the existing instances
-    everywhere.
-    """
-    def __init__(self, rttKits=None, userKits=None):
-        """If no argument is given, the common object is returned
-        Passing a list of kits will change all the existing instances
-        """
-        
-        if rttKits == None: return
-        if userKits == None: userKits = []
-        self.kitTree = KitTree(rttKits, userKits)
-
-    def kit(self, kitName):
-        kit = self.kitTree.getKit(kitName)
-        
-        if kit == None:
-            if kitName == self.kitTree.topParentLabel:
-                logger.debug('Asked for a TopParent kit, returning None')
-            else:
-                m = 'No kit available with name '+kitName
-                logger.error(m)
-                raise RTTCodingError(m)
-        else:
-                logger.debug('Found kit with name '+kitName)
-
-        return kit
-
-    def getAllKits(self):
-        # print 'getAllKits type', self.kitTree.getAllKits().__class__
-        return self.kitTree.getAllKits()
-
-    def __str__(self):
-        string = ''
-        count = 0
-        string = ''
-        for group in self.kitTree.getKitNames():
-            kkit = self.kit(group)
-            string += '\n\nKit %d \n' % (count)
-            string += str(kkit)
-            count += 1
-
-        return string
-
-    def nKits(self): return self.kitTree.nKits()
-if __name__ == '__main__':
-
-    pass
diff --git a/Tools/RunTimeTester/src/JobGroupsParser.py b/Tools/RunTimeTester/src/JobGroupsParser.py
deleted file mode 100755
index 9d1495e61fcc7ba4b1fb3269ac5dd0908c068f5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobGroupsParser.py
+++ /dev/null
@@ -1,325 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-File for parsiing the job group information in the RTT part of the unified
-test configuration file
-"""
-
-import sys,imp, types, os.path
-
-import xml.dom.minidom
-from JobGroupKit        import JobGroupKit
-from Checkers           import SimpleFileDiffer, SimpleLogChecker, ChainedJobsLogChecker
-from Checkers           import HistogramDiffer
-from formatCollection   import formatCollection
-from validateXMLFile    import Validator
-from UserTestDescriptor import UserTestDescriptor
-from Evaluate           import Evaluate
-
-import childrenToDescendants
-
-# --------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# --------------------------------------------------------
-def makeKits(element):
-    "Get the list of groups"
-    groups = Evaluate('jobGroup', element)
-    kits = [handleGroup(group) for group in groups]
-    return kits
-
-def handleGroup(group):
-    parent = group.getAttribute('parent')
-    if parent == '': parent = None
-    name = group.getAttribute('name')
-
-    jg = JobGroupKit(name, parent)
-
-    logCheckerTags    = group.getElementsByTagName('logChecker')
-    logChecker        = handleLogCheckerTags(logCheckerTags)
-
-    checkerTags       = group.getElementsByTagName('checker')
-    checkers          = [handleCheckerTag(tag) for tag in checkerTags]
-
-    keepFileTags      = group.getElementsByTagName('keepFilePattern')
-    keepFiles         = [handleKeepFileTag(tag) for tag in keepFileTags]
-
-    auxFileTags       = group.getElementsByTagName('auxFilePattern')
-    auxFiles          = [handleAuxFileTag(tag) for tag in auxFileTags]
-
-    testTags          = group.getElementsByTagName('test')
-    tests             = [handleUserTestTag(tag) for tag in testTags]
-
-    actionTags        = group.getElementsByTagName('action')
-    actions           = [handleUserTestTag(tag) for tag in actionTags]
-
-    # monActionTags     = group.getElementsByTagName('monAction')
-    # monActions        = [handleUserTestTag(tag) for tag in monActionTags]
-    
-    jg.extendItem('keepFilePatterns', keepFiles)
-    jg.extendItem('checkers', checkers)
-    jg.extendItem('auxFilePatterns', auxFiles)
-    jg.logChecker = logChecker
-
-    jg.extendItem('tests', tests)
-    jg.extendItem('actions', actions)
-    # jg.extendItem('monActions', monActions)
-    
-    return jg
-
-# ---------------------------------------------------------------
-# log checkers - need special treatment as must be the first checker!
-# ---------------------------------------------------------------
-
-def handleLogCheckerTags(tags):
-    if len(tags) > 1:
-        logger.error('Did not get a unique log checker tag')
-        return None
-    if len(tags) == 0: return
-
-    tag = tags[0]
-    checkerClass = tag.getAttribute('class')
-    if checkerClass not in logCheckerDispatcher:
-        logger.error('Log Checker not in dispatch table: '+checkerClass)
-        return None
-
-    return logCheckerDispatcher[checkerClass](tag)
-
-    
-def makeSimpleLogChecker(tag):
-    searchStringTags  = tag.getElementsByTagName('searchString')
-    searchStrings     = [getText(sst.childNodes) for sst in searchStringTags]
-
-    vetoStringTags    = tag.getElementsByTagName('vetoString')
-    vetoStrings       = [getText(vst.childNodes) for vst in vetoStringTags]
-
-    mustFindAllSearch = {'True':True, 'False':False}.get(Evaluate('mustFindAllSearch/text()', tag), True)
-    return SimpleLogChecker(searchStrings, vetoStrings, mustFindAllSearch)
-
-def makeChainedJobsLogChecker(tag):
-    searchStringTags = tag.getElementsByTagName('searchString')
-    searchStrings    = [getText(sst.childNodes) for sst in searchStringTags]
-
-    vetoStringTags   = tag.getElementsByTagName('vetoString')
-    vetoStrings      = [getText(vst.childNodes) for vst in vetoStringTags]
-
-    return ChainedJobsLogChecker(searchStrings, vetoStrings)
-
-# ---------------------------------------------------------------
-# Checker object creation
-# ---------------------------------------------------------------
-
-def handleCheckerTag(tag):
-    checkerClass = tag.getAttribute('class')
-    if checkerClass not in checkerDispatcher:
-        logger.error('Checker not in dispatch table: '+checkerClass)
-        return
-    
-    return checkerDispatcher[checkerClass](tag)
-
-
-def makeSimpleFileDiffer(tag):
-    return SimpleFileDiffer()
-
-# ---------------------------------------------------------------
-# Handle keep file pattern tag
-# ---------------------------------------------------------------
-
-def handleKeepFileTag(tag):
-    """Returns a dictionary of the keep file pattern string, info string and display colour attached."""
-    infoString   =  tag.getAttribute('info')
-    displayColor =  tag.getAttribute('displayColor')
-
-    keepString = getText(tag.childNodes)
-
-    return {'keepFileString':keepString,
-            'infoString':infoString,
-            'displayColor':displayColor}
-
-# ---------------------------------------------------------------
-# Handle aux file pattern tag
-# ---------------------------------------------------------------
-def handleAuxFileTag(tag):
-    return getText(tag.childNodes)
-    
-#---------------------------------------------------------------
-#Instantiate User Test
-#---------------------------------------------------------------
-def list2dict(keyList=[], valueList=[]):
-    dict ={}
-    iterator =0
-    while iterator < len(keyList):
-        dict[keyList[iterator]]=valueList[iterator]
-        iterator +=1
-
-    return dict
-
-#-------------------------------------------------------------------------
-#Instantiate external class
-#-------------------------------------------------------------------------
-def handleUserTestTag(test):
-    """Instanstiate an installed test or action"""
-
-    userModuleTag    =  test.getElementsByTagName('modulename')
-    userModuleName   = [getText(umt.childNodes) for umt in userModuleTag]
-
-    tags         = test.getElementsByTagName('testname')
-    userTestName = [getText(utt.childNodes) for utt in tags]
-
-    # initialise an arguments dictionary
-    testArgDict  = {}
-
-    # fill a dictionary with argument name-value pairs
-    tags         = test.getElementsByTagName('arg')
-    [testArgDict.update(handleTestArgTag(tag)) for tag in tags]
-    
-    # fill a dictionary with argument name - [list of values] pairs
-    tags         = test.getElementsByTagName('listarg')
-    [testArgDict.update(handleTestListArgTag(tag)) for tag in tags]
-
-    # get the position attribute from the tag
-    position = test.getAttribute('position').strip()
-
-    # run the action/test in the atlas environment
-    runInAtlasEnv = test.getAttribute('runInAtlasEnv').strip()
-
-    testDescriptor   = UserTestDescriptor(runInAtlasEnv,
-                                          position,
-                                          userModuleName[0],
-                                          userTestName[0],
-                                          testArgDict)
-
-    return testDescriptor
-
-#-------------------------------------------------------------------------
-
-def handleTestArgTag(tag):
-    name = Evaluate('argname/text()', tag)
-    values = [Evaluate('text()', v) for v in Evaluate('argvalue', tag)]
-    if len(values)==1: values=values[0]
-    return {name: values}
-
-#-------------------------------------------------------------------------
-
-def handleTestListArgTag(tag):
-    argNameTag   = (tag.getElementsByTagName('listargname'))[0]
-    name         = getText(argNameTag.childNodes)
-    
-    argValueTags = (tag.getElementsByTagName('listargvalue'))
-    value        = [getText(tag.childNodes) for tag in argValueTags]
-    
-    return {name: value}
-
-#---------------------------------------------------------------------------
-
-def makeHistogramDiffer(tag):
-    compTypeTag = tag.getElementsByTagName('comparisonType')
-    compType    = compTypeTag[0] # accept only the first one if we have several of these tags (should be only one)
-
-    hd = HistogramDiffer(compType)
-    return hd
-
-# ---------------------------------------------------------------
-# Extract text from tag
-# ---------------------------------------------------------------
-
-def getText(nodelist):
-    rc = ""
-    for node in nodelist:
-        if node.nodeType == node.TEXT_NODE:
-            rc = rc + str(node.data)
-    return rc.strip()
-
-# ---------------------------------------------------------------
-# Dispatcher dictionaries
-# ---------------------------------------------------------------
-
-# SimpleLogChecker can be a log checker or a normal checker
-checkerDispatcher = {
-    'SimpleFileDiffer':         makeSimpleFileDiffer,
-    'SimpleLogChecker':         makeSimpleLogChecker,
-    'HistogramDiffer':          makeHistogramDiffer
-    }
-
-logCheckerDispatcher = {
-    'SimpleLogChecker':      makeSimpleLogChecker,
-    'ChainedJobsLogChecker': makeChainedJobsLogChecker
-    }
-
-logCheckers = {}
-
-def linkKits(kits):
-    # build a dictionary where each kit has its children listed.
-    # Keys are kits. values are lists of child kits.
-    # The kit with group Top has no parent. All leaf kits (no children)
-    # are in the dictionary.
-    linkTable = {}
-    [linkTable.setdefault(parentKit(kit,kits), []).append(kit) for kit in kits]
-    [linkTable.setdefault(kit,[]) for kit in kits]
-
-    # As the 'Top' kit has parent 'None',
-    # a spurious entry for kit with 'None' appears in the table. Remove it here.
-    if None in linkTable.keys():
-        if len(linkTable[None]) == 1 and (linkTable[None][0]).group == u'Top':
-            del linkTable[None]
-        else:
-            logger.error('linkTable error: parentless  kit found in table')
-            # logger.error('None entry: '+linkTable[None])
-            children = reduce(lambda x,y:x+y,
-                              [kit.group for kit in linkTable[None]])
-            logger.error('Parent=None, Children= '+str(children))
-
-    # sanity checks:
-    [logger.error('kit for group '+kit.group+' not in link table') for kit in kits if kit not in linkTable]
-    
-    # make the link table go from
-    # {parent: [direct children]} to {parent:[all decendants]}
-    childrenToDescendants.link(linkTable)
-    # dumpLinkTable(linkTable)
-
-    # add the parents contents to the children
-    # this may duplicate material:
-    # if grandma gives to ma and granddaughter
-    # then ma gives to grandaughter, then grandaughter
-    # will have duplicate information. In actual fact
-    # we dont know the order of the linkTable keys, so
-    # we do not know the degree of duplication.
-    # take care of duplicates in the kit.add method
-    for parent in linkTable.keys():
-        if not parent: continue
-        for child in linkTable[parent]:
-            child.add(parent)
-
-    for kit in kits: kit.setLogChecker(kits)
-            
-def parentKit(kit, kits):
-    for k in kits:
-        if k.group == kit.parent: return k
-    return None
-
-def dumpLinkTable(table):
-    labelDict = {}
-    for key in table.keys():
-        childLabels = []
-        for child in table[key]:
-            childLabels.append(child.group)
-        if key == None:
-            print key, childLabels
-        else:
-            print key.group, childLabels
-        
-    formatCollection(labelDict)
-
-def makeTopLevelJobGroupKits(fileName):
-    v = Validator(fileName)
-    v.validate() # exception if invalid
-        
-    logger.info('JobGroupParser making kits - begin')
-    groups = Evaluate('rtt/jobGroups', v._dom.documentElement)
-    if not groups: return []        
-    kits = makeKits(groups[0])
-    logger.info('JobGroupParser making kits - end. Made %d kit(s).' % len(kits))
-    return kits
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/JobMaker.py b/Tools/RunTimeTester/src/JobMaker.py
deleted file mode 100755
index c37386d5309e9d60bb994fa9b38b50a20de26eba..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobMaker.py
+++ /dev/null
@@ -1,207 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection import formatCollection
-
-# -------------------------------------------------------------------------
-class JobMaker:
-    "Construct a list of worker and watcher jobs"
-    
-    def __init__(self, xmlReader, paths, descFactory, minderFactory, logger):
-
-        self.logger = logger
-        self.logger.debug("JobMaker: __init__")
-
-        
-        self.jobMinders    = []
-        self.paths         = paths
-        self.xmlReader     = xmlReader
-        self.descFactory   = descFactory
-        self.minderFactory = minderFactory
-        descs              = self.makeWorkerDescs()
-
-        self.logger.debug("No of Worker descs: "+str(len(descs)))
-        # [logger.debug(desc.dump()) for desc in descs]
-
-        [self.makeWorkerMinder(desc)  for desc in descs]
-        self.logger.debug("No of Worker jobs: "+str(len(self.jobMinders)))
-
-        descs = self.makeWatcherDescs(self.jobMinders)
-        self.logger.debug("No of Watcher descs: "+str(len(descs)))
-        [self.makeWatcherMinder(desc) for desc in descs]
-
-        self.logger.debug("No of Worker+Watcher jobs: "+str(len(self.jobMinders)))
-
-        # debug for now: will  need to append watchers to jobMinders
-        # [job.dump() for job in self.jobMinders]
-
-    def makeWorkerMinder(self, desc):
-        "use a worker descriptor to make a worker minder"
-        
-        msg = 'Creating a minder for descriptor %s' % desc.identifiedName
-        self.logger.debug(msg)
-        
-        jm = self.minderFactory.create(desc)
-        self.jobMinders.append(jm)
-
-    def makeWatcherDescs(self, workerMinders):
-        """
-        make a list of watcher descriptors.
-        !. Find confederations of worker __minders__
-        2. For each federation, construct a watcher descriptor
-
-        Keep track of the jobs associated with a watcher descriptor, and the
-        name of of the confederation in dictionaries keyed on the descriptor.
-        
-        A federation is a set of jobs with the same value of the federation
-        attribute. This is ususally set in the Jobs.xml config file
-        """
-        
-        # list all descriptors with the confedration attribute set
- 
-        confederatedMinders = [jm for jm in workerMinders if jm.confederation]
-        
-        # place in dictionary key = confederation (text), val = [minders]
-        
-        dict = {}
-        [dict.setdefault(jm.confederation, []).append(jm) for jm in confederatedMinders]
-
-        # check that all jobs in a confederation are in the same job group
-        
-        watcherDescriptors = []
-
-        
-        for key in dict.keys():
-
-            # a list of jobs...
-            confederationMembers = dict[key]
-            
-            if not self.isIsoGroup(confederationMembers):
-                # what should be done here?
-                self.logger.error('Non iso group confederationMembers ! '+key)
-
-            from getInheritanceHeirarchy import isAthenaJob
-            athenaJobs = [isAthenaJob(minder) for minder in confederationMembers]
-            
-            allAreAthenaJobs = reduce(lambda x, y: x and y, athenaJobs)
-            if allAreAthenaJobs:
-                if not self.isIsoAthenaJob(confederationMembers):
-                    # what should be done here?
-                    self.logger.error('Non isoAthenaJob confederationMembers ! '+key)
-
-
-            fJob   = confederationMembers[0]
-            descriptor = self.descFactory.create('Watcher',
-                                                 self.paths,
-                                                 {'jobGroup':fJob.jobGroup+'Watcher'})
-            
-            if allAreAthenaJobs:            
-                descriptor.datasets   = fJob.datasets
-
-            # store the list of worker minders in the watcher descriptor
-            # so the watcher can process workers' results
-            descriptor.confederation = confederationMembers
-            watcherDescriptors.append(descriptor)
-
-        return watcherDescriptors
-    
-
-    def makeWatcherMinder(self, descriptor):
-        "Minder construction"
-
-        minder = self.minderFactory.create(descriptor)        
-        self.jobMinders.append(minder)
-    
-        
-    def isIsoGroup(self, confederation):
-        "test if all jobs in a confederation are of the same jobGroup"
-        
-        fGroup = confederation[0].jobGroup
-        
-        # list the groups not equal to the jobGroup of the 0th job
-        badGroups = [job.jobGroup for job in confederation if job.jobGroup != fGroup]
-
-        if badGroups:
-            msg = 'Job federation has mixed groups, exiting '+str(badGroups)
-            self.logger.error(msg)
-            exit(0)
-            return False
-
-        return True
-
-
-    def isIsoAthenaJob(self, confederation):
-        """ check jobs have certain athena job attributes, and that these attributes
-        all have the same value"""
-        
-        # check all descs have certain attributes equal to that  of the first
-        # areSimilar is a list of bools.
-        areSimilar = [self.isAthenaSimilar(confederation[0], job)  for job in confederation]
-
-        # if any differ, return false.
-        athenaSimilar = reduce(lambda x, y: x and y, areSimilar)
-
-        return athenaSimilar
-
-    def isAthenaSimilar(self, job1, job2):
-        "checks if a number of attributes are similar for athena descriptors"
-
-        similar = True
-        
-        if job1.hasData() != job2.hasData():
-            self.logger.error('job1 hasData() : %s, job2 hasData() : %s' % (job1.hasData(), job2.hasData()))
-            similar = False
-
-
-        if len([datasetType for datasetType in job1.datasets if datasetType not in job2.datasets]) != 0:
-            self.logger.error('job1 contains data of a type not in job2!')
-            similar = False
-                
-        #if job1.isDC1data  != job2.isDC1data:
-        #    self.logger.error('job1 isDC1data : %s, job2 isDC1data : %s' % (job1.isDC1data, job2.isDC1data) )
-        #    similar = False
-        #if job1.DC1dataSet != job2.DC1dataSet:
-        #    self.logger.error('job1 DC1dataSet : %s, job2 DC1dataSet : %s' % (job1.DC1dataSet, job2.DC1dataSet) )
-        #    similar = False
-
-        if job1.jobGroup != job2.jobGroup:
-            self.logger.error('job1 jobGroup : %s, job2 jobGroup : %s' % (job1.jobGroup, job2.jobGroup) )
-            similar = False
-
-        #        if job1.isBSdata: similar =False
-
-        return similar
-
-    def makeWorkerDescs(self):
-        "Obtain the worker job descriptors from the jobs xml file"
-
-        try:
-            descList = self.xmlReader.getDescriptors()
-        except:
-            self.logger.error('JobsXMLReader object could not create descs')
-            return []
-
-        if self.paths.isNightly:
-            nBefore  = len(descList)
-            descList = [desc for desc in descList if not desc.numberedOnly]
-            nAfter   = len(descList)
-            nCulled  = nBefore - nAfter
-            msg = """numberedOnly selection: %d descriptors were removed; %d remain""" % (nCulled, nAfter)
-            self.logger.info(msg)
-
-        if self.paths.limitedNightlyVeto:
-            nBefore  = len(descList)
-            descList = [desc for desc in descList if not desc.limitedNightlies]
-            nAfter   = len(descList)
-            nCulled  = nBefore - nAfter
-            msg = 'limitedNightlies: %d descriptors were removed; %d remain' % (nCulled, nAfter)
-            self.logger.info(msg)
-
-        return descList
-    
-    def __str__(self):
-        str += self.__class__.__name__+'\n'
-        str += formatCollection(self.__dict__)+'\n'
-        return str
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/JobOptionsGenerator.py.obsolete b/Tools/RunTimeTester/src/JobOptionsGenerator.py.obsolete
deleted file mode 100755
index 4c94a0618ee2b1b2fc75937d7d9726b05f5cd27e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobOptionsGenerator.py.obsolete
+++ /dev/null
@@ -1,227 +0,0 @@
-#! /usr/bin/python
-
-"""
-Standalone script to generate job option files for Atlfast using a template
-script, and passing in parameters to sepcigy AF options.
-Author: P Sherwood, using work of J. Couchman
-"""
-
-import os,time,sys,shutil, string
-from os.path import join
-
-class JobOptionsGenerator:
-
-    def __init__(self, shareDir):
-        self.shareDir     = shareDir
-        self.joNames      = []
-        self.doAtlfast    = True
-        self.doGenerators = False
-
-    def makeAll(self):
-        if self.doAtlfast: self.joNames = self.joNames+ self.makeAtlfast()
-        if self.doGenerators: self.joNames = self.joNames+ self.makeGenerators()
-
-        longJo = 0
-        longGp = 0
-        for tokens in self.joNames:
-
-            jo=tokens[0]
-            if len(jo)>longJo: longJo = len(jo)
-
-            
-            group=tokens[1]
-            if len(group)>longGp: longGp=len(group)
-
-        longJo=longJo+3
-        longGp=longGp+3
-        
-        
-        jobList = open(join(self.shareDir, "JJJobs.cfg"),"w")
-
-        for tokens in self.joNames:
-            jon=string.ljust(tokens[0], longJo)
-            gpn=string.ljust(tokens[1], longGp)
-            jobList.write(jon+gpn+"short\n")
-        
-    
-    def makeAtlfast(self):
-        "create jobOptions files, and write out a config table for Atlfast"
-        physicsProcess = ["pythia_ele",
-                          "pythia_mu",
-                          "pythia_tau",
-                          "pythia_qcd",
-                          "pythia_ttbar",
-                          "pythia_whbb",
-                          "isajet_default",
-                          "herwig_default",
-                          "zebra_default"
-                          ]
-
-        smearing      = ["false","true"]
-        luminosity    = ["low", "high"] #1 is low, 2 is high
-        jetBuilder    = ["cone","kt","shared"]
-        clusterWith   = ["cells","tracks"]
-        # Not yet done:
-        #        for a in physicsprocess:
-        #            self.allFlags.append("atlfast_"+a+"_shower")
-        #            self.allFlags.append("atlfast_"+a+"_false_low_Cone_cells")
-        aCouple = ["pythia_default","zebra_default"]
-
-        # substitutions -
-        # 
-        # a dictionary which gives value for string that occurs
-        # in the template jobOptions files., This value is to be  are to be
-        # substituted in the template jo file to give a jo file.
-        
-
-        substitutions = {}
-        config        = []
-        joNames       = []
-        for a in aCouple:
-            for b in  smearing:
-                for c in luminosity:
-                    for d in jetBuilder:
-                        for e in clusterWith:
-                            substitutions["PROCESS"]           = a
-                            substitutions["SMEARING"]          = b
-                            if c=="low":
-                                substitutions["LUMINOSITY"]    = 1
-                            else:
-                                substitutions["LUMINOSITY"]    = 2
-                            
-                            substitutions["CLUSTERSTRATEGY"]   = d
-                            if e=="cells":
-                                substitutions["CLUSTERCELLS"]  = "true"
-                                substitutions["CLUSTERTRACKS"] = "false"
-                            elif e=="tracks":
-                                substitutions["CLUSTERCELLS"]  = "true"
-                                substitutions["CLUSTERTRACKS"] = "false"
-
-                            else:
-                                print " makeAtlfastConfigTable(): error"
-                                sys.exit()
-                                
-                            flags                   = a+"_"+b+"_"+c+"_"+d+"_"+e
-                            substitutions["FLAGS"]  = flags
-                                    
-        
-                            joName = "Atlfast_"+flags
-                            joContent = self.makeJobOptions(substitutions,
-                                                            "jobOptionsAF.tpl")
-                            open(join(self.shareDir,joName+".txt"),"w").write(
-                                joContent)
-
-                            joNames.append([joName, "Atlfast"])
-    
-        return joNames
-    
-# =============================================================================
-
-    def makeGenerators(self):
-        "create jobOptions files"
-        physicsProcess = ["pythia_ele",
-                          "pythia_mu",
-                          "pythia_tau",
-                          "pythia_qcd",
-                          "pythia_ttbar",
-                          "pythia_whbb",
-                          "isajet_default",
-                          "herwig_default",
-                          "zebra_default"
-                          ]
-        
-        substitutions = {}
-        joNames       = []
-        
-        for a in physicsProcess:
-            substitutions["FLAGS"] = a
-        
-            joName = "Generators_"+a
-            joContent = self.makeJobOptions(substitutions, "jobOptionsG.tpl")
-            open(join(self.shareDir, joName, ".txt", "w").write(joContent))
-            joNames.append([joName,"  Generators"])
-                            
-# =============================================================================
-
-    def makeJobOptions(self, subs, jobOptions):
-
-        joContent = open(join(self.shareDir, jobOptions)).read()
-
-        for key in subs.keys():
-            joContent = joContent.replace("!$"+key+"$!",str(subs[key]))
-
-
-        return joContent
-
-# ===============================
-########################################################################
-#class JobOptions_Generator code has migrated to the appropriate
-# Descriptor PS 20/9/06
-########################################################################
-
-#class JobOptions_Generator:
-#    def __init__(self):
-#        pass
-#
-#    def jobOptions(self,jobOptions,jDescriptor):
-#        if not jDescriptor.hasData(): return jobOptions
-#        
-#        fileStringList = []
-#        fileStringList.append("include(\"" + jobOptions + "\")\n")
-#        fileStringList.append(self.datasetCommand(jDescriptor)+'\n' )
-#        newJobOptionFile = join(jDescriptor.runPath,"RTTjobOptions.py")
-#        joFile = open(newJobOptionFile,"w").writelines(fileStringList)
-#
-#        return newJobOptionFile                       
-#            
-#    def datasetCommand(self,jd):
-#        import string
-#        
-#        s = ''
-#
-#        for datasetType in jd.datasets.keys():
-#            for ds in jd.datasets[datasetType]:                
-#                s += str(ds.whichAthenaList()) + ' += [\"'+string.replace(ds.name, '/castor/','rfio:/castor/')+'\"]\n'    
-#            s = string.replace(s[:-1],'+=', '=', 1) # remove final carriage return and convert first '+=' to '+'
-#            
-#        return s
-#
-#
-#
-# class JobOptions_GeneratorGrid(JobOptions_Generator):
-#     def __init__(self):
-#         JobOptions_Generator.__init__(self)
-#
-#    def datasetCommand(self,jd):
-#        import string
-#        
-#        s = ''
-#
-#        for datasetType in jd.datasets.keys():
-#            for ds in jd.datasets[datasetType]:
-#                name = os.path.basename(ds.name)
-#                s += str(ds.whichAthenaList()) +' += [\"'+string.replace(
-#                    name, '/castor/','rfio:/castor/')+'\"]\n'
-#                
-#            # remove final carriage return and convert first '+=' to '+'    
-#            s = string.replace(s[:-1],'+=', '=', 1) 
-#        
-#        return s
-#
-#    def jobOptions(self,jobOptions,jDescriptor):
-#        jobOptions = os.path.basename(jobOptions)
-#
-#        if not jDescriptor.hasData(): return jobOptions
-#        
-#        fileStringList = []
-#        fileStringList.append("include(\"" + jobOptions + "\")\n")
-#        fileStringList.append(self.datasetCommand(jDescriptor)+'\n' )
-#        newJobOptionFile = join(jDescriptor.runPath,"RTTjobOptions.py")
-#        joFile = open(newJobOptionFile,"w").writelines(fileStringList)
-#
-#        return newJobOptionFile
-#
-#if __name__ == "__main__":
-#    g = JobOptionsGenerator("../share")
-#    g.makeAll()
-
diff --git a/Tools/RunTimeTester/src/JobTransformDescriptor.py b/Tools/RunTimeTester/src/JobTransformDescriptor.py
deleted file mode 100755
index 1911114e8156014ac468500163ec81b6c9a1181f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobTransformDescriptor.py
+++ /dev/null
@@ -1,113 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""
-Class that stocks information about JobTransform jobs.
-Responsible for setting up the run directory for a job. Eventually becomes a data member
-of JobMinder.
-"""
-import os
-
-from Paths               import Paths
-from os.path             import join, exists
-from Descriptor          import Descriptor
-from formatCollection    import formatCollection
-from RTTSException       import RTTCodingError
-#from Tools2              import escapeTheCommand
-
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-class JobTransformDescriptor(Descriptor):    
-    
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-
-        self.datasets          = {}
-        self.displayClass      = []
-        self.displayProcess    = []
-        self.displayComponent  = []
-        
-        # init the base class
-        Descriptor.__init__(self, paths, logger, jobSerialNumber)
-
-        # update with values from the pacakge config file
-        # jobGroup, jobTransformCmd, jobTransformJobName
-        self.name            = argDict.pop('jobTransformJobName')
-        self.jobTransformCmd = argDict.pop('jobTransformCmd')
-        self.__dict__.update(argDict)
-
-        self.fixName()
-        
-        self.setPathExtensions()
-        self.setRunPath(self.pathExtensions)
-        # self.jobTransformCmd = escapeTheCommand(self.jobTransformCmd)
-        
-    # ----------------------------------------------------------------------
-
-    def dataForXMLNode(self):
-        names = ['displayComponent', 'displayClass', 'displayProcess']
-        dict = Descriptor.dataForXMLNode(self)
-        for n in names: dict[n] = self.__dict__[n]
-        return dict
-
-    # ----------------------------------------------------------------------
-        
-    def setPathExtensions(self):
-        self.pathExtensions = [
-            ('package',    self.paths.package),
-            ('jobGroup',   self.jobGroup),
-            ('jobName',    self.name),
-            ('SN',         str(self.jobSerialNumber))
-            ]        
-
-
-    # ----------------------------------------------------------------------
-
-    # seen by PyLint: no attribute pythonScriptName. Never called, so comment out
-    #def script(self):
-    #    script = '%s(%s)'
-    #    script = self.pythonScriptName
-
-    # ----------------------------------------------------------------------
-
-    def deferredMinderUpdate(self):
-       Descriptor.deferredDataSetUpdate(self) 
-       self.updateJobTransformCommand()
-
-    # ----------------------------------------------------------------------
-
-    def updateJobTransformCommand(self):
-        """
-        replace the logical data set names in the jobtransform command with the
-        corresponding physical data name.
-        """
-
-        m = '%s::updateTransformCmd() original job transform command \n\n%s\n' % (self.__class__.__name__, self.jobTransformCmd)
-        self.logger.debug(m)
-        
-        for dsInstances in self.datasets.values():
-            for dsInst in dsInstances:
-                physicalName = dsInst.physical
-                logicalName  = dsInst.logical
-                if not logicalName: 
-                    m = 'JobTransformDescriptor %s: logical dataset name not set for dataset %s' % (self.identifiedName, str(dsInst))
-                    raise RTTCodingError(m)
-
-                self.jobTransformCmd = self.jobTransformCmd.replace(logicalName,physicalName)
-
-        m = '%s::updateTransformCmd() amended job transform command \n\n%s\n' % (self.__class__.__name__, self.jobTransformCmd)
-        self.logger.debug(m)
-       
-    
-    def __str__(self):
-        s = '--------- JobTransformDescriptor ---------\n'
-        s += Descriptor.__str__(self) 
-        return s
-
-    # ----------------------------------------------------------------------
-    
-    def dump(self):
-        """Dumps out descriptor"""
-        self.logger.info(self.__str__)
-
diff --git a/Tools/RunTimeTester/src/JobsXMLReader.py b/Tools/RunTimeTester/src/JobsXMLReader.py
deleted file mode 100755
index 2bae71ca7fe3eb60e84cf917dac57706a05b79be..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/JobsXMLReader.py
+++ /dev/null
@@ -1,933 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""
-Part of the parser for the unfied test configuration file.
-Handles jobs for  the RTT part of the file,
-Returns a list of JobDescriptors
-
-If there is no extension to the jobOptions file name,
-'.py' is understood. Thus, to use the old '.txt' option files, supply
-the options file with the .txt extension attatched to the job options file name.
-
-Python Modules used:
-      xml.dom.minidom
-      sys
-      os.path
-      string
-
-RTT Modules used:
-      Descriptor
-      AthenaJobDescriptor
-      findPakages
-
-Usage:
-     
-"""
-
-import xml.dom.minidom, sys, os.path, string, imp, copy
-
-# from Descriptor               import Descriptor # now use factory to create
-# from AthenaJobDescriptor      import AthenaJobDescriptor # now use factory to create
-# from ReleaseJobDescriptor     import ReleaseJobDescriptor # now use factory to create
-from findPackages             import findPackages
-from RTTSException            import RTTInputError
-from RTTSException            import RTTCodingError
-from formatCollection         import formatCollection
-from exc2string2              import exc2string2
-from findRTThome              import findRTThome
-from copy                     import deepcopy
-from Factory_DataSet          import DataSetFactory, DQ2DataSet
-from XMLTools                 import getText, getElementNodes, nodeToString
-from JobGroupsParser          import handleUserTestTag,list2dict,handleKeepFileTag, handleAuxFileTag
-from Evaluate                 import Evaluate
-
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-class UseToPackage:
-    def __init__(self, releasePath):
-        self.releasePath = releasePath
-
-    def convert(self, use):
-        pack = None
-        tkns = use.split()
-        
-        # if len(tkns) == 5:
-        #     pack = os.path.join(self.releasePath, tkns[3], tkns[1], tkns[2])
-        # if len(tkns) == 4:
-        #     pack = os.path.join(self.releasePath, tkns[1], tkns[2])
-
-        # example of a use:
-        # 'use AMI AMI-02-00-07 Database/Bookkeeping (/afs/cern.ch/atlas/software/dist/10.3.0/)
-        if len(tkns) == 5:
-            pack = os.path.join(tkns[3], tkns[1], tkns[2])
-        if len(tkns) == 4:
-            pack = os.path.join(tkns[2])
-
-        return pack
-
-
-
-
-# from MethodTracer        import MetaTracer
-# __metaclass__ = MetaTracer
-
-
-class TryIt:
-    def __init__(self, job, dict):
-        self.job = job
-        
-        self.dict = dict
-        self.error = False
-        self.messages = []
-
-    def runFn(self, fn, tag):
-        try:
-            fn(self.job.getElementsByTagName(tag), self.dict)
-        except:
-            self.error = True
-            m = 'Error building descriptor in method %s\nTraceback:%s' % (fn.__name__, exc2string2())
-            self.messages.append(m)
-
-    def runXPathFn(self, fn, xpath):
-        try:
-            fn(Evaluate(xpath, self.job), self.dict)
-        except:
-            self.error = True
-            m = 'Error building descriptor in method %s\nTraceback:%s' % (fn.__name__, exc2string2())
-            self.messages.append(m)
-        
-    def handleSimpleTextTag(self, (tagName, dictKey) ):
-        nodes = self.job.getElementsByTagName(tagName)
-        if nodes: self.dict[dictKey] = string.strip(getText(nodes[0].childNodes))
-
-    def handleSimpleBoolTag(self, tagName):
-        nodes = self.job.getElementsByTagName(tagName)
-        if nodes: self.dict[tagName] = True
-
-
-class SimpleTagHandler:
-    """Add an entry to the arg dict with tagname as key, when you know
-    you want their is only one element with name = tagname"""
-
-    def __init__(self, dict, logger):
-        self.dict = dict
-
-    def doit(self, node, tagName):
-        print node.nodeName, tagName
-        try:
-            nodes = node.getElementsByTagName(tagName)
-            if not nodes: return
-            node = nodes[0]
-            self.dict[tagName] = getText(node.childNodes)
-
-        except:
-            m  = """Could not retrieve text for tagname %s, ignoring
-            Traceback:\n %s""" % (tagName, exc2string2())
-            logger.error(m)
-
-from Tracer import Tracer
-                    
-
-#def updateJobTransformCmdWithDataSets(datasetsDict, transformCmd):
-#    """grep the command for logical data set names and replace them with the
-#    physical data set path.
-#    """
-#    for dsType in datasetsDict.keys():
-#        dsInstances = datasetsDict[dsType]
-#        for dsInst in dsInstances:
-#            physicalName = dsInst.physical
-#            logicalName  = dsInst.logical
-#            if logicalName:
-#                transformCmd = transformCmd.replace(logicalName,physicalName)
-#    return transformCmd
-    
-
-class JobsXMLReader:
-    def __init__(self, parentElement, rttPilotJobs, packageName, dataSetCatalog, chainStoreMaker, logger, descFactory):
-        # parentElement = <jobList>
-        logger.debug('JobsXMLReader init')
-        self.rttPilotJobs       = rttPilotJobs
-        self.packageName        = packageName
-        self.dataSetCatalog     = dataSetCatalog
-        self.chainStoreMaker    = chainStoreMaker
-        self.chainStoreName     = None # will be filled in when processing Chain tags
-        self.descriptors        = []
-        self.classificationTags = None # the "apply to all jobs" classification
-        self.descFactory        = descFactory
-        
-        self.makeJobDescriptors(parentElement)
-        
-        
-    def getDescriptors(self): return self.descriptors
-        
-    # -------------------------------------------------------------------
-
-    def makeJobDescriptors(self, parentElement):
-        # store the list of global classification tags that will be
-        # applied to (currently) athena jobs (in theory 0 or 1 tag will be in the list)
-        # Jobs can locally provide their own tag where:
-        #      local empty classification tag ---> no CPC for that job
-        #      local non-empty classification tag ---> local CPC for that job
-        self.classificationTags = Evaluate('classification', parentElement)
-            
-        athenaTags              = Evaluate('athena',        parentElement)
-        jobTransformTags        = Evaluate('jobTransform',  parentElement)
-        chainTags               = Evaluate('chain',         parentElement)
-        pseudoTags              = Evaluate('pseudo',        parentElement) # make dummy descs for testing
-
-        nChainedAthenaJobs = 0
-        for n in chainTags:
-            nChainedAthenaJobs += len (n.getElementsByTagName('athena'))
-
-        nChainedJobTransforms = 0
-        for n in chainTags:
-            nChainedJobTransforms += len (n.getElementsByTagName('jobTransform'))
-        
-        logger.debug('no of athena jobs:              %d' % len(athenaTags))
-        logger.debug('no of job transform jobs:       %d' % len(jobTransformTags))
-        logger.debug('no of job chains:               %d' % len(chainTags))
-        logger.debug('no of chained athena jobs:      %d' % nChainedAthenaJobs)
-        logger.debug('no of chained job transforms:   %d' % nChainedJobTransforms)
-        logger.debug('no of pseudo chains:            %d' % len(pseudoTags))
-
-        descriptors = []
-        if self.rttPilotJobs:
-            descriptors  += self.handleAthenaJobs   (athenaTags,   'AthenaPilot')
-        descriptors  += self.handleAthenaJobs       (athenaTags,   'Athena')
-
-        descriptors += self.handleJobTransformJobs (jobTransformTags)
-        descriptors += self.handleChains           (chainTags)
-        descriptors += self.handlePseudoJobs       (pseudoTags)
-
-        logger.debug('no of descriptors:  %d' % len(descriptors))
-    
-        self.descriptors = descriptors
-
-    # ####################################################################################
-    #
-    # Handlers for jobs tags that are not specifics to a type of job
-    #
-    # ####################################################################################
-    
-    def handleQueues(self, queues, dict):
-        logger.debug('queues ...'+str(queues))
-        if queues:
-            dict['suggestedQueue']  =string.strip(getText(queues[0].childNodes))
-              
-
-            
-    # ####################################################################################
-    #
-    # Handlers for jobs that set attributes in Athena job descriptors
-    #
-    # ####################################################################################
-    
-
-    def handleAthenaJobs(self, jobs, factoryParameter):
-        """loops over all athena tags to make athena descriptors"""
-    
-        jobList = []
-
-        for job in jobs:
-            try:
-                jobList.append(self.handleAthenaJob(job, factoryParameter))
-            except Exception, e:
-                logger.error('Error processing athena tags for descriptor type %s, skipping job\nTraceback:\n%s' % (factoryParameter,
-                                                                                                                    exc2string2()))
-        return jobList
-
-
-    def handleAthenaJob(self, job, factoryParameter):
-        """
-        Returns a list of job descriptors and pilot job descriptors
-        """
-        
-        logger.debug("Athena job  "+str(job))
-
-        argDict = {}
-        tryIt = TryIt(job, argDict)
-
-        names = {
-            'commandLineFlags': 'commandLineFlags',
-            'commandLineString': 'commandLineString',        
-            'confederation':    'confederation',
-            'queue':            'suggestedQueue',
-            # 'mode':             'mode',
-            'doc':              'jobDocString',
-            'doclink':           'jobDocURL',
-            'displayName':      'jobDisplayName',
-            'athenaFlavour':     'athenaFlavour',
-            }
-
-        [tryIt.handleSimpleTextTag(item) for item in names.items()]
-
-        names = ['numberedOnly', 'limitedNightlies', 'athenaCommonFlags']
-        [tryIt.handleSimpleBoolTag(tag) for tag in names]
-
-        # more complicated stuff
-        tryIt.runFn(self.handleAthenaGroups,            "group")
-        tryIt.runFn(self.handleAthenaPreconfigOptions,  "preconfigOptions")
-        tryIt.runFn(self.handleAthenaOptions,           "options")
-        tryIt.runFn(self.handleFurtherAthenaOptions,    "furtherOptions")
-        tryIt.runFn(self.handleAthenaDataSets_OldStyle, "dataset")
-        tryIt.runXPathFn(self.handleAthenaDataSets,     "dataset_info")
-        tryIt.runFn(self.handleDQ2DataSets,             'dq2')
-        tryIt.runFn(self.handleAction,                  "action")
-        tryIt.runFn(self.handleTest,                    "test")
-        tryIt.runFn(self.handleClassification,          "classification")
-        tryIt.runFn(self.handleCASTORSvcClass,          "castorSvcClass")        
-        tryIt.runFn(self.handleBatchWallTime,           "batchWallTime")        
-        tryIt.runXPathFn(self.handleChainDataSets,      "chaindataset_info")
-        
-        #chaindatasets = job.getElementsByTagName('chaindataset_info')
-        #self.handleChainDataSets(chaindatasets, argDict)
-
-        keepFileEls = job.getElementsByTagName('keepFilePattern')
-        argDict['keepFilePatterns'] = [handleKeepFileTag(tag) for tag in keepFileEls]
-
-        auxFileEls = job.getElementsByTagName('auxFilePattern')
-        argDict['auxFilePatterns'] = [handleAuxFileTag(tag) for tag in auxFileEls]
-
-        # rationalise multiple job options
-        argDict['jobOptions'].extend(argDict['furtherJobOptions'])
-        del argDict['furtherJobOptions']
-        
-        argDict['hashString'] = hash(self.packageName+nodeToString(job))
-
-
-        # get the long term trend analysis (CoolDozer) id if it exists
-        argDict['trendId'] = job.getAttribute('trendId')
-        
-        jd = self.descFactory.create(factoryParameter, argDict)
-
-        if tryIt.error:
-            jd.constructionError = True
-            jd.errorMessages = tryIt.messages
-
-
-        return jd
-
-    # ####################################################################
-    # Extract the values of the athena child tags
-    # Expected tags are:
-    #          <group>       - mandatory
-    #          <package>     - mandatory
-    #          <options>     - mandatory
-    #          <queue>       - mandatory
-    #          <dataset>     - optional
-    # ###################################################################
-
-    def handleAthenaGroups(self, groups, dict):
-        # KLUDGE: Athena added to job group to make it the same as the
-        #         name of job group definition version
-        
-        logger.debug('groups '+str(groups))
-        if groups:
-            groupName = string.strip(getText(groups[0].childNodes))
-            if not groupName.startswith('Athena'): groupName = 'Athena'+groupName
-            dict['jobGroup'] = groupName
-
-    # -------------------------------------------------------------------------------------
-
-    def handleBatchWallTime(self, walltime, dict):
-        if not walltime: return
-        dict['suggestedBatchWallTime'] = getText(walltime[0].childNodes)
-
-    # -------------------------------------------------------------------------------------
-
-    def handleClassification(self, classification, dict):
-        # self.classificationTags = list of global level <classification> tags
-        # classification = list of local-to-this-job level <classification> tags            
-        def validateCPCvalue(tagName, tagValues):
-            legalValues = {'displayClass'     : ['OfflineValidation', 'OnlineValidation', 'Development'],
-                           'displayProcess'   : ['EvGen', 'Simul', 'Digi', 'Reco', 'Core'],
-                           'displayComponent' : ['Athena-Core',
-                                                 'Det-InDet', 'Det-Muon', 'Det-LAr', 'Det-TileCal', 'Det-Combined',
-                                                 'Trig-Electron', 'Trig-Photon', 'Trig-Tau', 'Trig-ETMiss',
-                                                 'Trig-Bphys', 'Trig-Btag', 'Trig-Cosmics', 'Trig-Common', 'Trig-Other',
-                                                 'Perf-Egamma', 'Perf-CombMuon', 'Perf-JetEtmiss', 'Perf-Tau',
-                                                 'Perf-TrkVtx', 'Perf-Btag', 'Phys-Top', 'Phys-Higgs', 'Phys-Susy',
-                                                 'Phys-StandardModel', 'Phys-Exotics', 'Phys-B', 'Phys-HeavyIon']
-                           }
-
-            for tagValue in tagValues: 
-                if tagValue not in legalValues[tagName]: 
-                    m  = "Invalid CPC tag name/value: %s, %s\n" % (tagName, tagValue)
-                    m += "Legal values are: %s" % str(legalValues[tagName])
-                    raise RTTInputError(m)
-
-        def extractChildren(parentNode):
-            childTags = {}
-            for tag in ['displayClass', 'displayProcess', 'displayComponent']:
-                values = [Evaluate('text()', v) for v in Evaluate(tag, parentNode)]
-                validateCPCvalue(tag, values)
-                childTags[tag] = values
-            return childTags
-
-        def addChildrenToDict(parentNode, dict):
-            dict.update(extractChildren(parentNode))
-
-        if not classification: # no local job, use global if exists
-            if not self.classificationTags or Evaluate('none', self.classificationTags[0]):
-                return
-            else:
-                addChildrenToDict(self.classificationTags[0], dict)
-        else:
-            # there exists a local CPC tag for this job
-            # empty version? ==> switch off the global CPC tag if it exists
-            if Evaluate('none', classification[0]):
-                return
-            # non-empt y version
-            else:
-                addChildrenToDict(classification[0], dict)
-    
-    # -------------------------------------------------------------------------------------
-
-    def handleAthenaPreconfigOptions(self, preconfigs, dict):
-        dict['preconfigJobOptions'] = []
-        if preconfigs:
-            dict['preconfigJobOptions'] = [string.strip(
-                self.fixAthenaJobOptions(getText(preconfigs[0].childNodes)))]
-
-    # -------------------------------------------------------------------------------------
-    
-    def handleAthenaOptions(self, options, dict):
-        logger.debug('options '+str(options))
-        dict['jobOptions'] = []
-        if options:
-            dict['jobOptions'] = [string.strip(
-                self.fixAthenaJobOptions(getText(options[0].childNodes)))]
-
-    # -------------------------------------------------------------------------------------
-    
-    def handleFurtherAthenaOptions(self, furtherOptions, dict):
-        # there can be only one joboptions tag.
-        # these tags allow chaining of athena jobs. Would have been
-        # better had jobOptions been a list from the begining.
-        
-        logger.debug("furtherOptions" +str(furtherOptions))
-        dict['furtherJobOptions']  =  []
-        if furtherOptions:
-            tags =  furtherOptions[0].getElementsByTagName("jobOptionItem")
-            dict['furtherJobOptions']  =  [getText(jot.childNodes)
-                                           for jot in tags]
-
-    # -------------------------------------------------------------------------------------
-        
-    def getPhysicalDataset(self, logicalDataSetName):
-        
-        if not self.dataSetCatalog.has_key(logicalDataSetName):
-            msg =  'logical data set name not known to  catalog %s' % logicalDataSetName
-            raise RTTInputError(msg)
-
-        resDataSetName = self.dataSetCatalog.get(logicalDataSetName)
-        if not resDataSetName:
-            msg = 'The logical data set name:\n' + str(logicalDataSetName) + '\n maps to no physical data set in the catalog!'
-            #  msg =  'physical data set name not available %s' % resDataSetName
-            raise RTTInputError(msg)
-    
-        return resDataSetName
-
-    # -------------------------------------------------------------------------------------
-
-    def handleDQ2DataSets(self, dq2Tags, dict):
-        logger.debug('handleDQ2Datasets start')
-        
-        dsObjs = []
-        type   = ''
-        for tag in dq2Tags:
-            dsName, dq2files, type = self.handleDQ2Tag(tag)
-            physicalFiles             = self.getDQ2PhysicalFiles(dsName, dq2files)
-            dsFactory  = DataSetFactory(logger)
-            for fn in physicalFiles:
-                dsInstance = dsFactory.create(type, fn, fn, 'dq2') # fn = physical = logical
-                dsObjs.append(dsInstance)
-            
-        if not dict.has_key('datasets'):
-            dict['datasets'] = {}
-        dict['datasets'].setdefault(type,[]).extend(dsObjs)
-        logger.debug('handleDQ2Datasets end')
-        
-            
-    def handleDQ2Tag(self, tag):
-        siteTag   = tag.getElementsByTagName('dq2site')
-        dq2site   = getText(siteTag[0].childNodes)
-        dsTag     = tag.getElementsByTagName('dq2dataset')
-        dq2dsName = getText(dsTag[0].childNodes)
-        typeTag   = tag.getElementsByTagName('datatype')
-        type = getText(typeTag[0].childNodes)
-        fileTags  = tag.getElementsByTagName('dq2file')
-        dq2files  = []
-        for t in fileTags:dq2files.append(getText(t.childNodes))
-        return (dq2dsName, dq2files, type)
-
-
-    def getDQ2PhysicalFiles(self, dsName, dq2files):
-
-        physFiles = []
-        for file in dq2files:
-            file = os.path.join(dsName, file)
-            if not self.dataSetCatalog.has_key(file):
-                msg =  'logical data set name not known to  catalog %s' % file
-                raise RTTInputError(msg)
-            physF = self.dataSetCatalog.get(file)
-            if not physF:
-                msg  = 'The logical data set name:\n' + str(file)
-                msg += '\n maps to no physical data set in the catalog!'
-                #  msg =  'physical data set name not available %s' % resDataSetName
-                raise RTTInputError(msg)
-            physFiles.append(physF)
-            
-        return physFiles
-    #--------------------------------------------------------------------------------------
-
-    def handleAthenaDataSets_OldStyle(self, datasets, dict):
-        """
-        Old dataset declaration: <dataset>/blah/blah</dataset>
-        Old style means dc2 data.
-        """
-
-        if not datasets: return
-        
-        logger.debug('Processing ' + str(len(datasets)) + ' old style <dataset> tags')
-
-        datasetInstanceList = []
-        
-        datasetType   = 'dc2'
-        datasetRepeat = 1
-
-        for dsElementNode in datasets:
-
-            logicalDataSetName = string.strip(getText(dsElementNode.childNodes))
-            physDataSetName = self.getPhysicalDataset(logicalDataSetName)
-
-            dsFactory = DataSetFactory(logger)            
-            datasetInstance = dsFactory.create(datasetType,physDataSetName,physDataSetName, 'standard') # physical = logical
-            datasetInstanceList.append(datasetInstance)
-
-        if not dict.has_key('datasets'):
-            dict['datasets'] = {}
-        dict['datasets'].setdefault(datasetType,[]).extend(datasetInstanceList)
-            
-    # -------------------------------------------------------------------------------------
-                
-    def handleAthenaDataSets(self, datasets, dict):
-        """New dataset declaration:
-              <dataset_info>
-                 <dc2/>
-                 <datasetName>/castor/blah/blah</dataset>
-                 <datasetRepeat>5</datasetRepeat>
-              </dataset_info>
-
-              The repeat tag is optional (defaults to 1). Known data type tags, currently: <dc2/> and <bytestream/>.
-        """
-
-        if not datasets: return
-        
-        datasetsDict = {}
-
-        logger.debug('Processing ' + str(len(datasets)) + ' <dataset_info> tags')
-        [self.handleAthenaDataSet(datasetNode, datasetsDict) for datasetNode in datasets]
-
-        if not dict.has_key('datasets'):
-            dict['datasets'] = {}
-
-        [dict['datasets'].setdefault(datasetType,[]).extend(datasetsDict[datasetType]) for datasetType in datasetsDict.keys()]
-        
-    # -------------------------------------------------------------------------------------
-    
-    def handleAthenaDataSet(self, node, dict):
-        datasetType   = getElementNodes(node.childNodes)[0].tagName
-        datasetName   = getText(node.getElementsByTagName('datasetName')[0].childNodes)                
-        datasetRepeat = self.getDatasetRepeat(node)            
-
-        physDataSetName = self.getPhysicalDataset(datasetName)
-        
-        dsFactory = DataSetFactory(logger)
-        dataSetInstance = dsFactory.create(datasetType,physDataSetName,physDataSetName,'standard') # physical = logical
-        [dict.setdefault(datasetType,[]).append(dataSetInstance) for i in range(datasetRepeat)]
-        
-        logger.info('handleAthenaDataset no of data sets %d' % (len(dict)))
-
-    # -------------------------------------------------------------------------------------
-
-    def getDatasetRepeat(self,datasetTag):
-
-        datasetRepeat = 1 # default
-        
-        repeatTags = datasetTag.getElementsByTagName('datasetRepeat')
-        if not repeatTags: return datasetRepeat
-        
-        datasetRepeat = getText(repeatTags[0].childNodes)
-            
-        try:
-            datasetRepeat = int(datasetRepeat)
-        except:
-            msg  = 'The tag <datasetRepeat> must take a non-zero positive integer as content. Instead, found: ' + str(datasetRepeat) + '\n'
-            msg += 'This Athena job will now be rejected.\n'
-            logger.error(msg)
-            raise RTTInputError(msg)
-
-        return datasetRepeat
-        
-    # -------------------------------------------------------------------------------------
-    
-    def fixAthenaJobOptions(self, jo):
-        base,ext = os.path.splitext(jo)
-    
-        if ext =='' :jo=jo+'.py'
-        return jo
-
-    # -------------------------------------------------------------------------------------
-                
-    def handleChainDataSets(self, datasetNodes, dict):
-        """New dataset declaration:
-              <chaindataset_info>
-                 <dc2/>
-                 <chaindatasetName>/castor/blah/blah</dataset>
-                 <datasetRepeat>5</datasetRepeat>
-                 <dataset_info>... </datasetinfo> <!-- fall back file -->
-              </chaindataset_info>
-
-              The repeat tag is optional (defaults to 1). Known data type tags, currently: <dc2/> and <bytestream/>.
-        """
-
-        if not datasetNodes: return
-        
-        chainDataSets = []
-
-        logger.debug('Processing ' + str(len(datasetNodes)) + ' <chaindataset_info> tags')
-
-        [chainDataSets.append(self.handleChainDataSet(node)) for node in datasetNodes]
-
-
-        dict['chainStoreFiles'] = chainDataSets
-
-    # -------------------------------------------------------------------------------------
-    
-    def handleChainDataSet(self, node):
-        """
-        Returns a tuple (datsetInstance, datasetDictionary)
-        the datasetInstancer represents the chain file to be used.
-        Its existence is not checked (this allows deferral: 
-        previous chain jobs may reate the dataset).
-        The dataset dictionary represents the fallback file(s).
-        It may have a number of entries of form 
-        'datasetType': [dataset instances]'
-        """
-        
-        def getFallBackFiles(logical):
-            """
-            PS 11/02/09
-            All fallback datasets are AthenaDatasets
-            return a dictionary of fallback datasets of the form
-            {'datasetType': [dataset objects]}
-            """
-            fallBackFileNodes = Evaluate('dataset_info', node)
-            logger.debug('%d fallback file node(s) found' % len(fallBackFileNodes))
-            tempDict = {}
-
-            # obtain a dict of form {'datasets': {dataset dictionary}}
-            self.handleAthenaDataSets(fallBackFileNodes, tempDict)
-
-            if len(tempDict)==0: logger.warning('No valid fallback files found')
-            dictionary = tempDict['datasets'] # get the inner dictionary
-            for dsInsts in dictionary.values(): # why?
-                for ds in dsInsts:
-                    logger.info('Fallback file found, name %s' % str(ds))
-                    ds.logical = logical
-                    ds.origin = 'fallback'
-            return dictionary
-
-
-        # ---- Method proper starts here ----
-
-        datasetType     = getElementNodes(node.childNodes)[0].tagName
-        datasetName     = Evaluate('chaindatasetName/text()', node) # logical name
-
-        #if not self.chainStoreArea:
-        #    msg = 'Trying to use a chain file but no self.chainStoreArea\n'
-        #    msg = 'Perhaps using a chainfile outside of a chain.'
-        #    msg = 'Chain file name = %s' % datasetName
-        #    logger.error(msg)
-        #    useFallBackFile(localDict, datasetName)
-        #    dict['datasets'] = localDict['datasets']
-        #    return
-
-        #physDataSetName = os.path.join(self.chainStoreArea, datasetName)
-        #if not (os.path.exists(physDataSetName)):
-        #    msg = 'chain file %s does not exist' % physDataSetName
-        #    logger.error(msg)
-        #    logger.error(msg)
-        #     useFallBackFile(localDict, datasetName)
-        #    dict['datasets'] = localDict['datasets']
-        #    return
-
-        physicalDatasetName = 'unknown'
-        datasetInstance = DataSetFactory(logger).create(datasetType, physicalDatasetName, datasetName, 'chainstore')
-        fallBackFileDict = getFallBackFiles(datasetName)
-        
-        return (datasetInstance, fallBackFileDict)
-        # dict.setdefault('datasets', {})
-        # dataSetInstance = DataSetFactory(logger).create(datasetType, physDataSetName, datasetName)
-        # dict['datasets'].setdefault(datasetType,[]).append(dataSetInstance)
-
-    # -------------------------------------------------------------------------------------
-    
-
-    def handleCASTORSvcClass(self, svcClass, dict):
-        if not svcClass: return        
-        dict['castorSvcClass'] = Evaluate('text()', svcClass[0]) # ignore all but first one
-        
-    def handleAction(self, actions, dict):
-        if not actions: return
-        dict['actions'] = [handleUserTestTag(action) for action in actions]
-            
-    def handleTest(self, tests, dict):
-        if not tests: return
-        dict['tests'] = [handleUserTestTag(test) for test in tests]
-
-    def handleJobTransformJobs(self, jobs):
-        jobList=[]
-        for job in jobs:
-            try:
-                jobList.append(self.handleJobTransformJob(job))
-            except:
-                m = 'Error creating a JobTransform Descriptor. Traceback\n%s' % exc2string2()
-                logger.error(m)
-                raise RTTCodingError(m)
-        return jobList
-                    
-    def handleJobTransformJob(self, job):
-
-        names = {
-            'group'               : 'jobGroup',
-            'doc'                 : 'jobDocString',
-            'queue'               : 'suggestedQueue',
-            'jobTransformCmd'     : 'jobTransformCmd',
-            'jobTransformJobName' : 'jobTransformJobName'
-            }
-        
-        argDict = {}
-
-        tryIt = TryIt(job, argDict)
-        [tryIt.handleSimpleTextTag(item) for item in names.items()]
-
-        names = ['numberedOnly', 'limitedNightlies']
-        [tryIt.handleSimpleBoolTag(tag) for tag in names]
-
-        tryIt.runFn(self.handleClassification, "classification")
-        tryIt.runFn(self.handleTest,           "test")
-        tryIt.runFn(self.handleAction,         "action")
-        tryIt.runFn(self.handleCASTORSvcClass, "castorSvcClass")        
-        tryIt.runFn(self.handleBatchWallTime,  "batchWallTime")
-
-        keepFileEls = job.getElementsByTagName('keepFilePattern')
-        argDict['keepFilePatterns'] = [handleKeepFileTag(tag) for tag in keepFileEls]
-
-        auxFileEls = job.getElementsByTagName('auxFilePattern')
-        argDict['auxFilePatterns'] = [handleAuxFileTag(tag) for tag in auxFileEls]
-
-        # grab the datasets if there are any
-        # Note: re-using the below method even though it's badly named.
-        # It should be changed to a more generic name.
-        # self.handleAthenaDataSets(job.getElementsByTagName('dataset_info'),dict)
-        tryIt.runXPathFn(self.handleAthenaDataSets,          "dataset_info")
-        argDict['hashString'] = hash(self.packageName+nodeToString(job))
-
-        # get the long term trend analysis (CoolDozer) id if it exists
-        argDict['trendId'] = job.getAttribute('trendId')
-
-        localError    = False
-        localErrorMsg = []
-        tryIt.runXPathFn(self.handleChainDataSets, "chaindataset_info")
-        
-        jd = self.descFactory.create('JobTransform', argDict)
-
-        if tryIt.error:
-            jd.constructionError = True
-            jd.errorMessages.extend(tryIt.messages)
-        if localError:
-            jd.constructionError = True
-            jd.errorMessages.extend(localErrorMsg)
-            
-        return jd
-
-    def handleChains(self, jobs):
-        jobList=[]
-        for job in jobs:
-            jobList.append(self.handleChain(job))
-        return jobList
-    
-    def handleChain(self, node):
-        # DTD guarantees there is one of <parallel>, <sequential>
-        # or chainElement
-        chainName = Evaluate('chainName/text()', node).replace(' ', '_') # protect against whitespace
-        self.chainStoreArea = self.chainStoreMaker.chainStorePath(chainName)
-
-        nodes = Evaluate('sequential', node)
-        if nodes:
-            desc = self.handleSequentialNode(nodes[0])
-            return desc
-
-        nodes = Evaluate('parallel', node)
-        if nodes:
-            desc = self.handleParallelNode(nodes[0])
-            return desc
-
-        nodes = Evaluate('ChainElement', node)
-        if nodes:
-            desc = self.handleChainElementNode(nodes[0])
-            return desc
-
-    def handleSequentialNode(self, node):
-        desc = self.handleContainerNode('Sequential', node)
-        return desc
-    
-    def handleParallelNode(self, node):
-        desc = self.handleContainerNode('Parallel', node)
-        return desc
-
-    def handleChainElementNode(self, node):
-        jd = self.descFactory.create('Chain')
-        # DTD gaurantees 1 of  athena, jobTransforms or pseudo jobs only
-
-        desc = None
-        nodes = Evaluate('athena', node)
-        if nodes:
-            desc = self.handleAthenaJob(nodes[0], 'Athena')
-            jd.addDescriptor(desc)
-
-        if not desc:
-            nodes = Evaluate('jobTransform', node)
-            if  nodes:
-                desc = self.handleJobTransformJob(nodes[0])
-                jd.addDescriptor(desc)
-
-        if not desc:
-            nodes = Evaluate('pseudo', node)
-            if  nodes:
-                desc = self.handlePseudoJob(nodes[0])
-                jd.addDescriptor(desc)
-
-        desc.chainStoreArea = self.chainStoreArea
-
-        outFileNodes = Evaluate('chainfileout', node)
-
-        jd.setOutFiles([Evaluate('text()', n) for n in outFileNodes])
-        return jd
-
-    def handleChainFileOutNodes(self, nodes):
-        return [Evaluate('text()', n) for n in nodes] 
-        
-    
-
-    def handleContainerNode(self, descName, node):
-
-        #descName is a class name starts with uppercase 
-    
-        jd = self.descFactory.create(descName)
-
-        # keys are xml node names (lower case)
-        handlers = {
-            'sequential':   self.handleSequentialNode,
-            'parallel':     self.handleParallelNode,
-            'chainElement': self.handleChainElementNode
-            }
-
-        for cnode in node.childNodes:
-            nodeName = cnode.nodeName
-            logger.info('%s Descriptor  checking child node %s' % (descName, nodeName))
-            if nodeName not in handlers.keys(): continue
-            logger.info('%s Descriptor  handling child node %s' % (descName, nodeName))
-            desc = handlers[nodeName](cnode)
-            desc.chainStoreArea = self.chainStoreArea
-            jd.addDescriptor(desc)
-            logger.info('%s Descriptor  adding child descriptor %s' % (descName, desc.__class__.__name__))
-
-        jd.chainStoreArea = self.chainStoreArea
-        return jd
-
-    def handlePseudoJobs(self, jobs):
-        jobList=[]
-        for job in jobs:
-            desc = self.handlePseudoJob(job)
-            jobList.append(desc)
-        return jobList
-    
-    def handlePseudoJob(self, node):
-
-        desc = self.handleAthenaJob(Evaluate('athena', node)[0], 'Pseudo') # pseudo has a single an athena node
-        return desc
-        #argDict = {}
-
-
-class JobsXMLReaderATN:
-    def __init__(self, ATNElement, rttPilotJobs, packageName, dataSetCatalog, chainStoreMaker, logger, descFactory):
-
-        self.descFactory = descFctory
-        self.logger      = logger
-        self.descriptors = []
-
-        self.makeJobDescriptors(ATNElement)
-        
-    def getDescriptors(self): return self.descriptors
-
-    def makeJobDescriptors(self, element):
-        # get the ATN TEST elements
-        testElements = Evaluate('TEST', element)
-        testElements = [t for t in testElements if t.getAttribute('type') == 'athena']
-        
-        if testElements:
-            self.descriptors.append(self.makeAthenaJobDescriptor(testElements))
-
-    def makeAthenaJobDescriptor(self, elements):
-
-        # The following line gets the text from the options_atn tag.
-        # This, conforming with other ATLAS usage, is in the form
-        # package/optionsfile name - if you give this to athena, it
-        # will find it. Currently, in the RTT, we find the full path
-        # to the options files ourselves - which lets up pick up
-        # options from the runpath.
-        # here, we extract the fill name part for processing in the RTT way.
-
-        def getOptionsAndScripts(elem):
-            jobOptionsEls = Evaluate('options_atn', elem)
-            prescriptEls  = Evaluate('prescript', elem)
-
-            jobOptions = [getText(e.childNodes) for e in jobOptionsEls]
-            prescripts = [getText(e.childNodes) for e in prescriptEls]
-            return (jobOptions,prescripts)
-            
-        optionsAndScripts = [getOptionsAndScripts(e) for e in elements]
-        jobOptions = []
-        [jobOptions.extend(jo[0]) for jo in optionsAndScripts]
-
-        prescripts = []
-        [prescripts.extend(ps[1]) for ps in optionsAndScripts]
-
-
-        argDict = {
-            'suggestedQueue'    : 'short',
-            'jobGroup'          : 'AthenaATN',      
-            'optionsAndScripts' : optionsAndScripts,
-            'jobOptions'        : jobOptions,
-            'prescripts'        : prescripts,
-            'rttATNJob'         : True,
-            'atnName'           : elements[0].getAttribute('name')
-            }
-
-
-        desc = self.descFactory.create('AthenaATN',
-                                       argDict)
-        logger.debug('JobsXMLReader::JobsXMLReaderATN::makeAthenaJobDescriptor %s' % desc.identifiedName)
-        # logger.debug(str(desc))
-        return desc
-        
-        
-        
-        
-if __name__ == "__main__":
-    pass
diff --git a/Tools/RunTimeTester/src/LSFBatchJobMinder.py b/Tools/RunTimeTester/src/LSFBatchJobMinder.py
deleted file mode 100755
index de8dbfee30221a0e7f1eeccdc82e7a1d9fb28382..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/LSFBatchJobMinder.py
+++ /dev/null
@@ -1,263 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from exc2string2     import exc2string2
-from BatchJobMinder  import BatchJobMinder
-from ShellCommand    import ShellCommand
-from Querier         import LSF_bjobs1, LSF_bjobs2, LSF_bjobs3, LSF_bhist
-
-import os
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-            
-class LSFBatchJobMinder(BatchJobMinder):
-    
-    def __init__ (self, argBag, jDescriptor):
-
-
-        self.setQueue(jDescriptor.suggestedQueue)
-
-        commander = argBag.commander
-
-        argBag.submitCommand      = commander.submitCommand(argBag.runPath,
-                                                            argBag.identifiedName,
-                                                            argBag.log,
-                                                            argBag.elog,
-                                                            self.queue,
-                                                            jDescriptor.suggestedBatchWallTime)
-        argBag.postProcessCommand = commander.postProcessorCommand(argBag.runPath,
-                                                                   argBag.identifiedName,
-                                                                   self.queue
-                                                                   )
-
-
-        BatchJobMinder.__init__(self, argBag, jDescriptor)
-        # total time from bjobs, spec int time from log file of samplejob
-        self.spec2000Conv = 712.0/265.1 
-        self.logger.debug('initialised LSFBatchJobMinder %s' % self.identifiedName)    
-        self.postProcessingQueue = 'short'
-
-    # ------------------------------------------------------------------------
-
-    def setQueue(self, suggestedQueue):
-        self.queue = {'short':  'atlasrttshort',
-                      'medium': 'atlasrttmedium',
-                      'long':   'atlasrttlong'}.get(suggestedQueue, 'atlasrttshort')
-
-    # ------------------------------------------------------------------------
-
-    def stopJob(self):
-        if self.done: return
-        
-        try:
-            cmd = ['bkill ' + str(self.pid)]
-            sc       = ShellCommand(cmd, self.logger, 100, None, True)
-            self.logger.debug(sc.getReply())
-        except Exception, e:
-            m  = 'Job: ' + str(self.identifiedName)
-            m += ' threw an exception while issuing cmd: %s' % cmd[0] 
-            self.logger.warning(m)
-
-
-    # -------------------------------------------------------------------------
-
-    def checkQueue(self):
-        #check if queue exists
-
-        querier = LSF_bjobs1(self.logger, self.queue)
-        goodQueue, output = querier.statusAndReply()
-
-        if not goodQueue:
-            msg  = 'Bad queue: %s' % (output)
-            self.logger.error(msg)
-            return False
-
-        return True
-
-    # -------------------------------------------------------------------------
-    
-    def checkJobSubmissionWasOK(self):
-        'Called by the StateEngine'
-
-        for l in self.subLog:
-            if l.find('is submitted to') == -1:
-                self.logger.error('Batch machine refused job submission!')
-                return False
-
-        #see if we have a reasonable jobs PID
-        try:
-            pid = int(self.pid)
-        except:
-            return False
-        
-        return True
-            
-    # -------------------------------------------------------------------------
-    
-    def batchQueryReplyFormatOK(self, reply):
-        if not reply: return False
-        for thing in reply:
-            if thing.find("batch system daemon not responding")!=-1: return False
-        return True
-    
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsInQueue(self, reply):
-        # batch queue reply to query for a running job is a header line
-        # followed by the status of each pid of the query
-        if len(reply)<2:
-            return False
-
-        tokens = reply[1].split()
-        if len(tokens)<3: return False
-        tokens = [t.strip() for t in tokens]
-        
-        self.logger.debug('tokens %s' % str(tokens))
-        if not tokens[0].startswith(str(self.pid)): return False
-        if tokens[2].startswith('RUN'): return True
-        if tokens[2].startswith('PEND'): return True
-        self.logger.debug('Job is NOT in queue')
-        return False # any other lsf state (eg DONE) - not in queue.
-
-    # -------------------------------------------------------------------------
-    
-    def batchReplyProvesJobIsAbsentFromQueue(self, reply):
-        """if the original bjobs command returned a 1 line reply
-        it might be the case that the job is part of the history.
-        check the history to see if the job is known.
-        if so, assume the job ran."""
-        
-        if not reply: return False
-        if len(reply) == 1:
-            sc = ShellCommand(['bhist %s' % str(self.pid)], self.logger)
-            if len(sc.getReply())>1:
-                return True
-            else:
-                return False
-            
-        return not self.batchReplyProvesJobIsInQueue(reply)
-
-    # -------------------------------------------------------------------------
-    
-    def batchQuery(self):
-        return 'bjobs ' + str(self.pid)
-
-    # -------------------------------------------------------------------------
-
-    def batchQueryCommand(self): return 'bjobs'
-
-    # -------------------------------------------------------------------------
-
-    def audit(self):
-        """
-        ask LSF for audit information. The availablility if the audit information varies.
-        If the job is running, cputime, mem, vmem, walltime and job status are available
-
-        For jobs wich have finished, but which do not have state DONE (eg EXIT) there was an error
-        and cputime and memory used are not available
-
-        For finished jobs in state DONE (success), memory numbers are not available. If memory numbers
-        were found durng the run, the highest value is recorded.
-        """
-
-        dict = {}
-
-        def findValues0(output, dict):
-            # self.logger.error('audit 10 '+output)
-            sstring = 'The CPU time used is'            
-            start = output.find(sstring)
-            if start < 0: return
-            output = output[start:]
-            tokens = output.split()
-
-            dict['cpuTime'] = int(round(float(tokens[5])))
-            if 'MEM:' in tokens:  dict['mem']     = tokens[8]
-            if 'SWAP:' in tokens: dict['vmem']    = tokens[11]
-
-        # status, output = commands.getstatusoutput('bjobs -l %s' % self.pid)
-        # if status:
-        #    self.logger.warning('bad status on command bjobs -l %s' % status)
-
-        querier = LSF_bjobs2(self.logger, self.pid)
-        status, reply = querier.statusAndReply()
-
-        if not status:
-            self.logger.warning('bad reply on command bjobs -l %s' % reply)
-        else:
-            self.logger.debug('good reply on command bjobs -l %s' % reply)
-            findValues0(reply, dict)
-
-
-        querier = LSF_bjobs3(self.logger, self.pid)
-        status, reply = querier.statusAndReply()
-        
-        if not status:
-            self.logger.warning('bad reply on command bjobs -a %s' % reply)
-            dict['batchStatus'] = 'unknown'
-        else:
-            dict['batchStatus'] = reply
-                
-        def findValues2(output, dict):
-            try:
-                tokens = output.split('\n')
-                tokens = tokens[2].split()
-            except:
-                m  = 'audit.findValues1 threw exception. Strange batch reply was: %s\n' % str(output)
-                m += 'Unexpected batch system reply result is being ignored. TraceBack:\n'
-                m += exc2string2()
-                self.logger.error(m)
-                dict['wallTime'] = 'Unobtainable'
-            else:   
-                dict['wallTime'] = (tokens[-1]).strip()
-
-        # status, output = commands.getstatusoutput('bhist %s' % self.pid)
-        querier = LSF_bhist(self.logger, self.pid)
-        status, reply = querier.statusAndReply()
-        if not status:
-            self.logger.warning('bad reply on  "bhist": %s' % reply)
-        else:
-            findValues2(reply, dict)
-
-        def acceptIfGreater(nstring, ostring):
-            try:
-                nf = float(nstring)
-            except:
-                return None
-
-            try:
-                of = float(ostring)
-            except:
-                return nf
-
-            if nf>of: return nf
-            return of
-
-        mem = acceptIfGreater(dict.get('mem', None), self.mem)
-        if mem != None: dict['mem'] = str(int(round(mem))) + 'MB'
-
-        mem = acceptIfGreater(dict.get('vmem', None), self.vmem)
-        if mem != None: dict['vmem'] = str(int(round(mem))) + 'MB'
-
-        try:
-            cpuTime = float(dict['cpuTime'])
-        except:
-            dict['cpuTime2000'] = 'Unknown'
-        else:
-            dict['cpuTime2000'] = int(round(self.spec2000Conv*cpuTime))
-
-        self.__dict__.update(dict)
-        return
-
-    # -------------------------------------------------------------------------
-
-    def setPid(self):
-        self.logger.debug('Calculating job pid from subLog %s' % str(self.subLog))
-        try:
-            self.pid = ((self.subLog[0].split())[1].strip())[1:-1] # submit cmd prints id to stdout
-        except:
-            self.logger.error('Could not evaluate job pid from %s' % str(self.subLog))
-        else:
-            self.logger.debug('Job pid: %s' % str(self.pid))
-
-
diff --git a/Tools/RunTimeTester/src/Launcher.py b/Tools/RunTimeTester/src/Launcher.py
deleted file mode 100755
index 2777078f7ed8fc335343ac1d1ddf9b558ddff498..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Launcher.py
+++ /dev/null
@@ -1,114 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Launchers govern the rate at which jobs are started.
-BatchLauncher,       submits all the jobs in rapid succession.
-InteractiveLauncher, waits for a job to finish before submitting the next.
-"""
-from RTTSException      import RTTCodingError
-from popen2             import Popen4
-import os
-
-from Tools2             import isType
-
-class Launcher:
-    def __init__(self, maxActiveJobs, jobs, logger):
-        if maxActiveJobs<=0:
-            msg = 'Nonsense value for maxActiveJobs: %d' % maxActiveJobs
-            logger.error(msg)
-            raise RTTCodingError(msg)            
-
-        self.logger        = logger
-        self.watchers      = []
-        self.workers       = []
-        # self.njobs         = 0
-        self.nLaunchedJobs = 0
-
-        self.maxActiveJobs = maxActiveJobs
-
-        self.workers   = [job for job in jobs if isType(job, 'WorkerJobMinder')]
-        self.watchers  = [job for job in jobs if isType(job, 'WatcherJobMinder')]        
-        # self.njobs     = len(self.watchers)
-
-        self.logger.debug('Launcher init: workers %d, watchers %d' % (len(self.workers),len(self.watchers)))
-        
-    def isWorkerJob(self,job):
-        heirarchy = [b.__name__ for b in job.__class__.__bases__]
-        heirarchy.append('WorkerJobMinder')
-        heirarchy.append('ErrorJobMinder')
-        return job.__class__.__name__ in heirarchy
-
-    def isWatcherJob(self,job):
-        heirarchy = [b.__name__ for b in job.__class__.__bases__]
-        heirarchy.append('WatcherJobMinder')
-        return job.__class__.__name__ in heirarchy
-        
-    def fire(self, job):
-        os.chdir(job.runPath)
-        job.poll()
-        self.nLaunchedJobs=self.nLaunchedJobs+1
-        
-class ThrottleLauncher(Launcher):
-
-    def __init__(self, maxActiveJobs, jobs, logger):
-        Launcher.__init__(self, maxActiveJobs, jobs, logger)
-        self.hasBeenLaunched  = []
-        self.firstBatchWaits = True
-
-    def launchNextUnlaunchedJob(self):
-        
-        if len(self.watchers)+ len(self.workers) ==0: return
-        
-        if self.firstBatchWaits:
-
-            self.firstBatchWaits = False
-            count = 0
-            for job in self.workers:
-                self.fire(job)
-                self.hasBeenLaunched.append(job)
-                count += 1
-                msg = 'first batch, fired '+job.identifiedName
-                self.logger.debug(msg)
-                if count >= self.maxActiveJobs: return
-
-        # remove finished jobs. Needs two lines: the next line
-        # gets messed up in the iteration over changing lists
-        # BAD: [self.watchers.remove(job) for  job in self.watchers if job.done]
-        doneJobs = [job  for  job in self.workers  if job.isDone()]
-        [self.workers.remove(job)  for  job in doneJobs]
-
-        doneJobs = [job  for  job in self.watchers  if job.isDone()]
-        [self.watchers.remove(job)  for  job in doneJobs]
-
-        # active jobs are jobs that have been launched, but no yet done.
-        nActiveJobs  = len( [m for m in self.workers
-                             if m in self.hasBeenLaunched])
-        
-            
-        nWatcherJobs = 0
-        for job in self.watchers:
-            if not job.isDone(): nWatcherJobs += 1
-
-    
-        
-        nToFire = min(self.maxActiveJobs-nActiveJobs, len(self.workers)-nActiveJobs)
-
-        if nToFire>0:
-            msg =  'ThrottleLauncher: worker jobs max: %d, fired: %d to fire: %d ' % (
-                self.maxActiveJobs,
-                nActiveJobs,
-                nToFire)
-        
-            self.logger.debug(msg)
-        
-        if nToFire < 0:
-            self.logger.error('Error: n active> maxActive')
-
-        if nToFire > 0:
-            for job in self.workers:
-                if job not in self.hasBeenLaunched:
-                    self.fire(job)
-                    self.hasBeenLaunched.append(job)
-                    nToFire = nToFire-1
-                    if nToFire == 0: break
-                
diff --git a/Tools/RunTimeTester/src/Legals.py b/Tools/RunTimeTester/src/Legals.py
deleted file mode 100755
index da8a4fd2683344aa06ce2bf86a3f5bd58496e667..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Legals.py
+++ /dev/null
@@ -1,90 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools               import isValidReleaseNumber
-from formatCollection    import formatCollection
-from RTTpath             import RTTpath
-
-import os, copy
-
-class LegalsBase:
-    
-    def __init__(self, hardCoded):
-
-        self.hardCoded = hardCoded
-        self.nightlies = None # overridden in subclasses as a dictionary
-        
-        self.legals = {
-            'runType':        ['build'],
-            'mode':           ['PBS', 'LinuxInteractive', 'LSFBatch',
-                               'Dummy', 'EmptyDummy', 'Grid', 'FullDummy'],
-
-            'cmtConfig':      ['i686-slc3-gcc323-dbg','i686-slc3-gcc323-opt','i686-slc4-gcc34-dbg','i686-slc4-gcc34-opt',
-                               'x86_64-slc4-gcc34-dbg', 'x86_64-slc4-gcc34-opt', 'i686-slc4-gcc43-opt', 'i686-slc4-gcc43-dbg'],
-            'branch':         ['dev', 'bugfix', 'val', 'pcache', 'devval',
-                               'mig1', 'mig2','mig3','mig4','mig5','mig6', 'mig7', 'mig8', 'mig9', 'mig10',
-                               'devmig1', 'devmig2', 'devmig3', 'devmig4','devmig5','devmig6',
-                               '13.2.0.Y', '14.0.0.Y', '14.0.10.Y', '14.1.0.Y', '14.2.0.Y',
-                               '14.2.5.Y', '14.3.0.Y', '14.2.X', '14.3.X', '14.2.10.Y', '14.2.1X', '14.2.1X.Y',
-                               '14.2.2X', '14.2.2X-VAL', '14.2.2X.Y', '14.2.2X.Y-VAL',
-                               '14.2.10.Y-VAL', '14.2.1X.Y-VAL', '14.2.20.Y', '14.2.30.Y', '14.2.OLD.Y', '14.2.OLD.Y-VAL',
-                               '14.4.X.Y','14.4.X.Y-VAL', '14.5.X-VAL', '14.5.X.Y', '14.5.X.Y-VAL','14.5.2.Y', '14.5.2.Y-VAL',
-                               '15.0.0.Y', '15.0.0.Y-VAL', '15.0.X.Y', '15.0.X.Y-VAL', '15.1.X.Y', '15.1.X.Y-VAL',
-                               '15.2.X.Y', '15.2.X.Y-VAL', '15.3.X.Y', '15.3.X.Y-VAL', '15.4.X.Y', '15.4.X.Y-VAL'],
-            
-            'site':           ['CERN', 'PetersMac','UCLslc3', 'CERNslc4'],
-            'confFileSuffix': ['_TestConfiguration', '_ReleaseTest'],
-            'releaseType':    ['project'],
-            'nightlyWord':    ['nightly', 'Nightly'],
-            'releases':       [], # filled by base class constructors
-            'nightlies':      []  # filled by base class constructors
-        }
-
-        # self.legals['webpageBranchVeto'] = copy.copy(self.legals['branch'])
-        # self.legals['webpageBranchVeto'].append('')
-
-    def validKeys(self): return self.legals.keys()
-    def vals(self, key): return self.legals[key]
-    def isLegal(self, key, val): return val in self.legals[key]
-
-    def isNightlyRelease(self, rel): return rel in self.legals['nightlies']
-    def isNightlyInRelease(self, rel):
-        return rel in self.legals['nightliesInRelease']
-
-    def convertLogicalNightly(self, release):
-        """
-        nightly release names in config files are atlrel_x
-        convert to the name appropriate to the branch and releaseType
-        if the release is of the form atlrel_x. Return unchanged otherwise
-         eg for numbered releases
-        """
-        return self.nightlies.get(release, release)
-    # ----------------------------------------------------------------------   
-
-    def __str__(self):
-        return formatCollection(self.legals)
-
-class LegalsProject(LegalsBase):
-    
-    def __init__(self, hardCoded):
-        LegalsBase.__init__(self, hardCoded)
-
-        self.nightlies =  {
-            'rel_0': 'rel_0',
-            'rel_1': 'rel_1',
-            'rel_2': 'rel_2',
-            'rel_3': 'rel_3',
-            'rel_4': 'rel_4',                      
-            'rel_5': 'rel_5',
-            'rel_6': 'rel_6'
-            }
-        
-        releases  =  copy.copy(self.nightlies.keys())
-        releases.extend(self.legals['nightlyWord'])
-        
-        self.legals['releases']    = releases
-        self.legals['nightlies']   = self.nightlies.keys()
-        self.legals['nightliesInRelease']   = self.nightlies.values()
-
-
-        # self.branch = 'N.X.0'
-        # self.releaseType = 'project'
diff --git a/Tools/RunTimeTester/src/LinuxInteractiveJobMinder.py b/Tools/RunTimeTester/src/LinuxInteractiveJobMinder.py
deleted file mode 100755
index 642df150ca88a622cd0714e64f69f3f17f05040f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/LinuxInteractiveJobMinder.py
+++ /dev/null
@@ -1,123 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from WorkerJobMinder import WorkerJobMinder
-from RTTpath         import RTTpath
-from RTTSException   import RTTCodingError
-
-from popen2 import Popen4
-
-# ==================================================================
-
-class LinuxInteractiveJobMinder(WorkerJobMinder):
-    
-    def __init__ (self, argBag, jDescriptor):
-
-        commander                 = argBag.commander
-        argBag.submitCommand      = commander.submitCommand(argBag.runPath, argBag.log)
-        argBag.postProcessCommand = commander.postProcessorCommand(argBag.runPath)
-
-
-        WorkerJobMinder.__init__(self, argBag, jDescriptor)
-
-        self.logger.debug( "initialising LinuxInteractiveJobMinder")
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('| Linux InteractiveJob Minder  dump         |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-    def fire(self, what):
-        """
-        Submits the job by spawning a subprocess.
-        The Popen objects captures std error, and std out
-        'what' is a a key in to tthe self.scripts dictionary
-        the popen object is bound to the object ot allow subsequent
-        interogation. This is a specificity of interactive running.
-        """
-
-        cmd = self.scripts[what]['cmd']
-
-        m =  'fire commandL\n: %s' % cmd
-        self.logger.debug(m)
-
-        self.popen  = Popen4(cmd)
-        self.subLog = self.popen.fromchild.readlines()
-
-
-    def jobIsRunning(self):
-        queryOK   = True
-        isRunning = False
-            
-        if self.popen.poll()==-1:
-            isRunning = True
-            print 'open.poll == -1'
-            return (queryOK, isRunning)
-            
-        #if not RTTpath(self.runPath, self.log).exists():
-        #    isRunning = True
-        #    print 'log file %s does nto exist ' % self.log
-        #    return (queryOK, isRunning)
-
-        print 'running is done %s' % str(self.popen)
-
-        return (queryOK, isRunning)
-
-    def fromQueued(self):
-        self.fire('runScript')
-        return 'running'
-
-    def fromRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-        self.diskspace('fromRunning, start')
-        if not queryOK:
-            msg = 'Unsuccesful query of a job running interactively'
-            raise RTTCodingError(msg)
-
-        if isRunning:  return 'running' # no state transition
-
-        self.fromRunningActions()
-        
-        if self.doPostProcessing():
-            self.diskspace('Launching postprocessing')
-            return 'postProcessingQueued'
-        
-        self.diskspace('Job done')
-        return 'done'
-
-    def fromRunningActions(self):
-        self.checkLog()
-        self.diskspace('Running checks')
-        self.runChecks()
-        # chainFileCopier present only if running in a chain 
-        if self.chainFileCopier:
-            self.chainFileCopier.copyFiles()
-
-        # if keep file is also a chain file, this code  will silently fail. Intended that the chain
-        # file copier is interogated to find new location
-        self.copyKeepFiles()
-        self.diskspace('After copyKeepFiles')
-
-
-    def fromPostProcessingQueued(self):
-        self.fire('postProcessorScript')
-        return 'postProcessingRunning'
-            
-    def fromPostProcessingRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-
-        if not queryOK:
-            msg = 'Unsuccesful query of a job running interactively'
-            raise RTTCodingError(msg)
-
-        if isRunning:  return 'postProcessingRunning'
-
-        self.readTestResults()
-        self.copyKeepFiles()
-        self.cleanSpace()
-
-        return 'done'
-        
diff --git a/Tools/RunTimeTester/src/LinuxInteractiveMinderStateEngine.py b/Tools/RunTimeTester/src/LinuxInteractiveMinderStateEngine.py
deleted file mode 100755
index 38f48c24aa5eda66ad42fe957404ef2124408fbd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/LinuxInteractiveMinderStateEngine.py
+++ /dev/null
@@ -1,25 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""State Engine for LinuxJobMinders: extends WorkerJobMinderStateEngine
-
-overrides: fromQueuedTest
-           fromRunningTest
-
-"""
-
-from MinderStateEngineBase  import MinderStateEngineBase
-
-class LinuxInteractiveMinderStateEngine(MinderStateEngineBase):
-    
-    def __init__(self, minder, state = 'queued'):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-        self.dispatchTable =  {
-            'queued':                minder.fromQueued,
-            'running':               minder.fromRunning,
-            'operatorCloseDown':     minder.fromOperatorCloseDown,
-            'postProcessingQueued':  minder.fromPostProcessingQueued,
-            'postProcessingRunning': minder.fromPostProcessingRunning,
-            'operatorCloseDown':     minder.fromOperatorCloseDown,
-            'done':                  minder.fromDone
-            }
-
diff --git a/Tools/RunTimeTester/src/LinuxInteractiveMinderToXML2.py b/Tools/RunTimeTester/src/LinuxInteractiveMinderToXML2.py
deleted file mode 100644
index e10b16b0636f68f0dddc45fb5c3333d4f8b5abaa..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/LinuxInteractiveMinderToXML2.py
+++ /dev/null
@@ -1,340 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an xml file summary of an RTT run
-"""
-from StringIO      import StringIO
-from makeRTTDOM    import makeRTTDOM
-from xml.dom.ext   import PrettyPrint
-from Defs          import RTTdefs
-from exc2string2   import exc2string2
-from MethodTimer   import MethodTimer
-
-import sys
-import os
-import fnmatch
-
-class LinuxInteractiveMinderToXML(MethodTimer):
-    
-    def __init__(self, minder):
-
-        MethodTimer.__init__(self)
-
-        
-        self.minder          = minder
-        self.logger          = minder.logger
-        self.elementCreator  = minder.elementCreator
-        self.textNodeCreator = minder.textNodeCreator
-        self.callBack        = None
-        self.minderXMLNode   = None
-        
-    def unlink(self):
-        del self.minder
-        if self.callBack: del self.callBack
-        
-    def update(self):
-
-        minder = self.minder # for nested methods
-        # nested scopes are used to pass parameters to these nested functions
-        def findErrorMessages():
-
-            element = self.elementCreator('errors')
-            
-            text = reduce(lambda x, y: x+y+'\n',
-                          minder.errorMessages,
-                          '')
-            if text:
-                tn = self.textNodeCreator(text)
-                element.appendChild(tn)
-
-            return element
-
-        def findDatasets():
-
-            dsElement = self.elementCreator('datasets')
-            
-            try:
-                dataSets = minder.datasets
-            except:
-                dataSets = {}
-
-            datasetInstances = []
-            [datasetInstances.extend(dataSets[k]) for k in dataSets]
-            #for ds in datasetInstances:
-            #    tn = self.textNodeCreator(os.path.basename(ds.physical.strip()))
-            #    el = self.elementCreator('dataset')
-            #    el.appendChild(tn)
-            #    dsElement.appendChild(el)
-
-            for ds in datasetInstances:
-                dsInfoEl = self.elementCreator('datasetinfo')
-
-                head, tail = os.path.split(ds.physical)
-                
-                tn = self.textNodeCreator(tail.strip())
-                el = self.elementCreator('file')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(head.strip())
-                el = self.elementCreator('directory')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.mtime).strip())
-                el = self.elementCreator('mtime')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.size).strip())
-                el = self.elementCreator('size')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.origin).strip())
-                el = self.elementCreator('origin')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-                                
-                dsElement.appendChild(dsInfoEl)
-
-            return dsElement
-
-        def findMissingDatasets():
-
-            dsElement = self.elementCreator('missingDatasets')
-            
-            try:
-                datasetInstances = minder.missingDatasets
-            except:
-                datasetInstances = []
-
-            for ds in datasetInstances:
-                tn = self.textNodeCreator(os.path.basename(ds.physical.strip()))
-                el = self.elementCreator('missingDataset')
-                el.appendChild(tn)
-                dsElement.appendChild(el)
-
-            return dsElement
-
-        def findNtestSuccess():
-            l = len([s for s in minder.finishedTests if s.result == 0])
-            tn = self.textNodeCreator(str(l))
-            el = self.elementCreator('sUserTests')
-            el.appendChild(tn)
-            return el
-
-        def findKeepFiles():
-
-            resPath = minder.resPath
-            
-            # minder.keepFilesPatterns is a list of dictionaries.
-            # kFilePatterns = [k['keepFileString']
-            #                 for k in minder.keepFilePatterns]
-            # minder.keepFiles is a dictionary of dictionaries.
-
-            results = []
-
-            for k,v in minder.keepFiles.items():
-                default = ''
-                name  = v.get('keepFileString', default)
-                info  = v.get('infoString', default)
-                color = v.get('displayColor', default)
-                md5   = v.get('md5sum', default)
-
-                lookIn  = os.path.dirname(name)
-                lookFor = os.path.basename(name)
-                if not os.path.exists(lookIn): continue
-
-                matches = fnmatch.filter(os.listdir(lookIn), lookFor)
-
-                result = [{'keepFileName' : os.path.join(lookIn, m), 'infoString' : info, 'displayColor' : color, 'md5sum' : md5} for m in matches]
-                results.extend(result)
-
-            return results
-                
-        def makeKeepFileNode():
-            # logger.debug('Going to make element deliveredkeepfiles, document: ' +str( document))
-            el = self.elementCreator('deliveredkeepfiles')
-
-            knownKeepFiles = findKeepFiles()
-            if knownKeepFiles:
-                baseEl = self.elementCreator('baseKeepFileDir')
-                dirRelToResBase = minder.resPath.split(minder.resultsBasePath)[1]
-                if dirRelToResBase.startswith('/'): dirRelToResBase = dirRelToResBase[1:]
-                if not dirRelToResBase.endswith('/'): dirRelToResBase + '/'
-                baseEl.appendChild(self.textNodeCreator(dirRelToResBase))
-                el.appendChild(baseEl)
-                
-            for k in knownKeepFiles:
-                name  = k['keepFileName']
-                info  = k['infoString']
-                color = k['displayColor']
-                md5   = k['md5sum']
-                self.logger.debug('Keepfile dict: %s' % str(k))
-
-                def keepFileNode():
-                    # def keepFileNode(style):
-                    # if style=='new_style':
-                    el2 = self.elementCreator('keepfile_newstyle')
-                    resBase = minder.resultsBasePath                    
-                    keepFileRelativeToResBase = name.strip().split(resBase)[1]
-                    if keepFileRelativeToResBase.startswith('/'): keepFileRelativeToResBase = keepFileRelativeToResBase[1:]
-                    tn = self.textNodeCreator(keepFileRelativeToResBase)
-                    el2.setAttribute('info', info)
-                    el2.setAttribute('displayColor', color)
-                    el2.setAttribute('md5sum', md5)
-                    el2.appendChild(tn)
-
-                    # if style=='old_style':
-                    #   el2 = self.elementCreator('keepfile')
-                    # tn = self.textNodeCreator(os.path.basename(str(name).strip()))
-                    #    el2.setAttribute('info',info)
-                    #   el2.setAttribute('displayColor',color)
-                    #    el2.appendChild(tn)
-                    return el2
-
-                el.appendChild(keepFileNode())
-                # el.appendChild(keepFileNode('old_style'))
-
-            return el
-            
-        def makeTextNode((tagName, text)):
-            tn = self.textNodeCreator(str(text).strip())
-            el = self.elementCreator(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(document))
-        
-        mElement = self.elementCreator('minder')
-        textInfo = {
-            'identifiedName':  self.minder.identifiedName,
-            'jobName':         self.minder.name,
-            'jobDisplayName':  str(self.minder.jobDisplayName),
-            'jobID':           self.minder.jobSerialNumber,
-            'state':           self.minder.stateEngine.state.state,
-            'status':          self.minder.status(),
-            'jobGroup':        self.minder.jobGroup,
-            'nUserTests':      self.minder.numberOfTests(),
-            'nUserActions':    self.minder.numberOfActions(),
-            'resultsPath':     self.minder.resPath,
-            'exitStatus':      self.minder.exitStatus,
-            'cpuTime':         self.minder.cpuTime,
-            'cpuTime2000':     self.minder.cpuTime2000,
-            'batchStatus':     self.minder.batchStatus,
-            'mem':             self.minder.mem,
-            'vmem':            self.minder.vmem,
-            'wallTime':        self.minder.wallTime,
-            'nRetries':        self.minder.errorStateCurRetry,
-            # 'nUpdates':        self.minder.nUpdates,
-            # 'sumTimeToUpdate': self.minder.sumTimeToUpdate,
-            'stateHistory':         str(self.minder.stateEngine.state.history),
-            'postProcessingResult': self.minder.postProcessingResult,
-            'processingResult':     self.minder.processingResult,
-            'chainLength':          str(self.minder.chainLength),
-            'chainSuccesses':       str(self.minder.chainSuccesses),
-            'jobDocString':    str(self.minder.jobDocString),
-            'jobDocURL':    str(self.minder.jobDocURL)
-            }
-
-            
-        textInfo.update(self.minder.descDataForXMLNode)
-
-        
-        for key,val in textInfo.items():
-            if val.__class__.__name__ in ['list', 'tuple']: # 1-to-many tag/value mapping
-                [mElement.appendChild(makeTextNode((key,v))) for v in val]
-            else: #1-to-1 tag/value mapping 
-                mElement.appendChild(makeTextNode((key,val)))
-            
-        #[mElement.appendChild(makeTextNode((key,val))) for key,val in textInfo.items()]
-        
-        mElement.appendChild(findDatasets())
-        mElement.appendChild(findMissingDatasets())
-        mElement.appendChild(findNtestSuccess() )
-        mElement.appendChild(makeKeepFileNode() )
-        mElement.appendChild(findErrorMessages() )
-          
-        
-        def makeFlagNode((tagName, flag)):
-            if flag: return self.elementCreator(str(tagName))
-            return None
-
-        flagInfo = [
-            ('rttPilotJob',   self.minder.rttPilotJob),
-            ('rttATNJob',     self.minder.rttATNJob)
-            ]
-
-        for f in flagInfo:
-            node = makeFlagNode(f)
-            if node: mElement.appendChild(node)
-
-        if self.minder.isDone():
-            mElement.appendChild(self.elementCreator('done'))
-
-        if self.minder.closedDownByOperator():
-            mElement.appendChild(self.elementCreator('operatorCloseDown'))
-
-
-        el = self.elementCreator('checkResults')
-        [el.appendChild(makeTextNode( ('checkResult', r) ))
-         for r in self.minder.checkResults]
-        mElement.appendChild(el)
-
-        el = self.elementCreator('testResults')
-        for r in self.minder.finishedTests:
-
-            inner = self.elementCreator('testResult')
-            n = makeTextNode( ('testName', r.testId) )
-            inner.appendChild(n)
-            if r.error:
-                n = makeTextNode( ('testStatus', 'error') )
-            else:
-                n = makeTextNode( ('testStatus', str(r.result) ) )
-            inner.appendChild(n)
-
-            el.appendChild(inner)
-
-        mElement.appendChild(el)
-        
-        self.outputToFile(mElement)
-
-        self.minderXMLNode = mElement
-
-
-        m = 'Created, minder node for %s' % self.minder.identifiedName
-        self.logger.debug(m)
-
-    def makeSummaryReport(self):
-        if not self.callBack:
-            self.logger.error('No object for Minder to call back to, abandoning update')
-            return 
-        self.update()
-        assert (self.minderXMLNode != None)
-        dict = self.minder.dataForMonitoring()
-        self.callBack.update(dict, self.minderXMLNode)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.update(), temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-
-    def outputToFile(self, minderNode):
-        space = ' '
-        f = os.path.join(self.minder.resPath, 'rttjobinfo.xml')
-        h = None
-        try:
-            h = open(f, 'w')
-            h.write(minderNode.toprettyxml(3*space, '\n'))
-            h.close()
-        except:
-            self.logger.error('Unable to output minder info XML file. Traceback: \n%s' % exc2string2())
-        if h: h.close()    
-
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/LoggerSetup.py b/Tools/RunTimeTester/src/LoggerSetup.py
deleted file mode 100755
index 4bf0f3ef928b87afd7fbc05971e9b016cc6d9f95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/LoggerSetup.py
+++ /dev/null
@@ -1,290 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-# ----------------------------------------------------------------
-# ----- IMPORTS --------------------------------------------------
-# ----------------------------------------------------------------
-
-import sys
-import os, os.path
-import logging
-
-from Tools import now, mkdirIfNew, rmdirMkdir, changePerm
-from ToVariableLoggingHandler import ToVariableLoggingHandler
-from exc2string2 import exc2string2
-
-# ----------------------------------------------------------------
-
-# pseudo private, not copied out on imports of this module
-dirToWhichLogsOutput = ''
-
-# ----------------------------------------------------------------
-
-def getDirToWhichLogsOutput():
-    return dirToWhichLogsOutput
-
-# ----------------------------------------------------------------
-# ----- SETUP FUNCTIONS ------------------------------------------
-# ----------------------------------------------------------------
-
-def setupLoggers(parentDir):
-    'Create the loggers for the RTT.'
-    
-    # ------------------------------
-    # Create the output dirs
-    # ------------------------------    
-    global dirToWhichLogsOutput
-    dirToWhichLogsOutput = createLogDirectories(parentDir)
-    
-    # ------------------------------
-    # Setup the format
-    # ------------------------------
-    formatter = createFormatter()
-
-    # -----------------------------------
-    # Create the full name of the logs
-    # -----------------------------------
-    mainLogName    = createLogFile(dirToWhichLogsOutput,'RTT.log','RTT.log\n')
-    errorLogName   = createLogFile(dirToWhichLogsOutput,'RTTerror.log','RTTerror.log\n') 
-    warningLogName = createLogFile(dirToWhichLogsOutput,'RTTwarning.log','RTTwarning.log\n') 
-    infoLogName    = createLogFile(dirToWhichLogsOutput,'RTTinfo.log','RTTinfo.log\n')
-
-    # ----------------------------------
-    # create the file handler loggers
-    # ----------------------------------
-    logFileHandler     = createFileHandler(mainLogName,logging.DEBUG,formatter)
-    errorFileHandler   = createFileHandler(errorLogName,logging.ERROR,formatter)
-    warningFileHandler = createFileHandler(warningLogName,logging.WARNING,formatter)
-    infoFileHandler    = createFileHandler(infoLogName,logging.INFO,formatter)
-
-    # ----------------------------------
-    # create the stream handler logger
-    # ----------------------------------
-    logStreamHandler = createStreamHandler(logging.INFO,formatter)
-
-    # ----------------------------------
-    # create the variable logging handler
-    # ----------------------------------
-    variableLoggingHandler = createVariableLoggingHandler(logging.INFO, formatter)
-
-    # -------------------------------------------
-    # add the handlers and set the logger level
-    # -------------------------------------------    
-    logger = addHandlers('rtt',
-                         logFileHandler,
-                         infoFileHandler,
-                         warningFileHandler,
-                         errorFileHandler,
-                         # logStreamHandler,
-                         # variableLoggingHandler)
-                         )
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    # ---------------------------------------------------------
-    # Now do all the above for a separate special case logger
-    # ---------------------------------------------------------
-
-    smallLogName     = createLogFile(dirToWhichLogsOutput,'RTTstandalone.log','RTTstandalone.log\n')
-    smallFileHandler = createFileHandler(smallLogName,logging.DEBUG,formatter)
-
-    smallLogger = addHandlers('rtt.standalone',smallFileHandler)
-    smallLogger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    return dirToWhichLogsOutput
-# ----------------------------------------------------------------
-
-def setupLoggerForDebug(level=None):
-    formatter = createFormatter()
-
-    level = loggerLevelFactory(level)
-
-    logStreamHandler = createStreamHandler(level,formatter)
-    logger = addHandlers(None,logStreamHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-# ----------------------------------------------------------------
-
-def setupLoggerForTestSuite(logDir, fixture, level=None):
-    print '-------logDir-----', logDir
-    formatter = createFormatter()    
-
-    # convert to logging variable. Sets default if necessary
-    level = loggerLevelFactory(level) 
-
-
-    debugLog    =  createLogFile(logDir,
-                           fixture+'_debug.log',
-                           fixture+'_debug.log\n')
-
-    infoLog     = createLogFile(logDir,
-                               fixture+'_info.log',
-                               fixture+'_info.log\n')
-    
-    warningLog  = createLogFile(logDir,
-                               fixture+'_warning.log',
-                               fixture+'_warning.log\n')
-
-    errorLog    = createLogFile(logDir,
-                               fixture+'_error.log',
-                               fixture+'_error.log\n')
-
-    criticalLog = createLogFile(logDir,
-                                fixture+'_critical.log',
-                                fixture+'_critical.log\n')
-
-
-    handlers = [createFileHandler(debugLog,    logging.DEBUG,   formatter),
-                createFileHandler(infoLog,     logging.INFO,    formatter),
-                createFileHandler(warningLog,  logging.WARNING, formatter),
-                createFileHandler(errorLog,    logging.ERROR,   formatter),
-                createFileHandler(criticalLog, logging.CRITICAL,formatter)
-                ]
-
-    logger = logging.getLogger('rttTestSuite')
-    [logger.addHandler(h) for h  in handlers]
-    
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    # print logger.handlers
-    return handlers
-
-# ----------------------------------------------------------------
-def resetLoggerForTestSuite(logger, handler, directory):
-    
-    level     = handler.level
-    formatter = handler.formatter
-    logger.removeHandler(handler)
-    fullPathToLog = createLogFile(directory,
-                                  'testSuite.log',
-                                  'testSuite.log\n')
-
-    fileHandler = createFileHandler(fullPathToLog, level, formatter)
-        
-    #    logger = addHandlers(None,fileHandler)
-    logger.addHandler(fileHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    print logger.handlers
-    return fileHandler # so we can reset it later
-# ----------------------------------------------------------------
-# ----- HELPER FUNCTIONS -----------------------------------------
-# ----------------------------------------------------------------
-
-def createLogDirectories(parentDir):
-    if not os.path.exists(parentDir) or not os.path.isabs(parentDir):
-        print 'Unable to setup the logger because the path: ' + str(parentDir)
-        print 'either does not exist, or is not absolute.'
-        sys.exit()
-            
-    logsDirName = 'logfiles_'+now()
-    fullLogsDirName = os.path.join(parentDir, logsDirName)
-
-    try:
-        if os.path.exists(fullLogsDirName):
-            rmdirMkdir(fullLogsDirName)
-        else:
-            mkdirIfNew(fullLogsDirName)
-    except Exception, e:
-        print 'Unable to create path: ' + str(fullLogsDirName)
-        print str(e)
-        print exc2string2()
-        sys.exit()                         
-
-    return fullLogsDirName
-
-# ----------------------------------------------------------------
-
-def createFileHandler(name,level,formatter):
-    theHandler = logging.FileHandler(name)
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-
-# ----------------------------------------------------------------
-
-def createStreamHandler(level,formatter):
-    theHandler = logging.StreamHandler()
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-
-# ----------------------------------------------------------------
-
-def createVariableLoggingHandler(level,formatter):
-    theHandler = ToVariableLoggingHandler()
-    theHandler.setFormatter(formatter)
-    theHandler.setLevel(level)
-
-    return theHandler
-    
-# ----------------------------------------------------------------
-
-def createFormatter():    
-    msgFmt    = '%(asctime)s %(filename)s, %(lineno)s, %(levelname)s %(message)s'
-    tmeFmt    = '%d%b%y %H:%M:%S'
-    formatter = logging.Formatter(msgFmt, tmeFmt)
-    return formatter
-
-# ----------------------------------------------------------------
-
-def createLogFile(parentDir,fileName,startText='A log file',permission=02775):
-    if not os.path.exists(parentDir) or not os.path.isabs(parentDir):
-        print 'The parent directory: ' + str(parentDir)
-        print 'in which to create the logger file: ' + str(fileName)
-        print 'is either inexistant, or is not absolute. It must be both.'
-        logging.shutdown()
-        sys.exit()
-
-    try:
-        fullPath = os.path.join(parentDir,fileName)
-        open(fullPath,'w').write(startText)
-    except Exception, e:
-        print 'Unable to create and write to the log file: '
-        print str(fullPath)
-        print str(e)
-        print exc2string2()
-        logging.shutdown()
-        sys.exit()
-
-    try:
-        changePerm(fullPath,permission)
-    except Exception, e:
-        print 'Unable to set permission: ' + str(permission)
-        print 'on the log file: ' + str(fullPath)
-        print str(e)
-        print exc2string2()
-        logging.shutdown()
-        sys.exit()
-    
-    return fullPath
-    
-# ----------------------------------------------------------------
-
-def addHandlers(theLoggerName=None,*handlers):
-    'Creates a logger instance and adds handlers to it. Returns the instance.'
-
-    logger = logging.getLogger(theLoggerName)
-    [logger.addHandler(handler) for handler in handlers]
-
-    return logger
-
-# ----------------------------------------------------------------
-
-def loggerLevelFactory(level=None):    
-    levels = {
-        'DEBUG':     logging.DEBUG,
-        'INFO':      logging.INFO,
-        'WARNING':   logging.WARNING,
-        'ERROR':     logging.ERROR,
-        'CRITICAL':  logging.CRITICAL
-        }
-    if level in levels:
-        level=levels[level]
-    else:
-        print 'setting logging level to DEBUG'
-        level = logging.DEBUG
-
-    return level
-
-# ----------------------------------------------------------------
diff --git a/Tools/RunTimeTester/src/MacroRunner.py.obsolete b/Tools/RunTimeTester/src/MacroRunner.py.obsolete
deleted file mode 100755
index 79e5213016f9a4b4e8e358b6694fcc841769a668..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MacroRunner.py.obsolete
+++ /dev/null
@@ -1,96 +0,0 @@
-"""
-Class used to pass commands to an application
-"""
-
-import os
-from popen2 import Popen4
-
-# from siteSpecificsFactory import siteSpecificsFactory
-from Factory_SiteSpecifics import SiteSpecificsFactory
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-#class MacroRunner(SitePostScripts):
-class MacroRunner:
-	
-	psLabel = 'site'
-
-	def __init__(self, cmd):
-		toBeEscaped=['(',')','"']
-
-		for char in toBeEscaped:
-			pieces = cmd.split(char)
-
-			cmd = pieces[0]
-			for piece in pieces[1:]:
-				cmd += '\\'+char+piece
-
-		# cmd = '\\r'+cmd
-			
-		self.cmd = cmd
-	
-	def run(self, site):
-		
-		"run some root macros"
-	       
-		logger.debug('running postscript  on ' +site)
-		logger.info('Current working directory: '+ os.getcwd())
-		#SitePostScripts.setSiteSpecifics(site)
-		self.setSiteSpecifics(site)
-
-		siteSpecicificsFact = SiteSpecificsFactory(logger).create(site)
-		rootCommand = siteSpecificsFact(site).rootCommand()
-		rootCommand = rootCommand+' '+self.cmd
-
-		logger.info( "root command: "+rootCommand)
-
-		runMacro = Popen4(rootCommand)
-		logger.info('root command output: \n'+ runMacro.fromchild.read())
-
-	def setSiteSpecifics(self, site):
-		logger.debug('setting site specific env on :' + site )
-		
-	        # siteSpec = siteSpecificsFactory(site)
-	        siteSpec = SiteSpecificsFactory(logger).create(site)
-		os.environ['LD_LIBRARY_PATH'] = siteSpec.libs()
-		os.environ['ROOTSYS']         = siteSpec.rootSys()
-	
-		
-	setSiteSpecifics=classmethod(setSiteSpecifics)
-
-
-if __name__ == "__main__":
-	import sys
-
-	from setupLoggerForDebug import setupLoggerForDebug
-	setupLoggerForDebug()
-
-	m = MacroRunner('atlfast.C("xxx", "yyy")')
-	print 'label = ', m.psLabel
-	print m.cmd
-
-	macro = 'egamma_hist.C("ntuple.root")'
-	m = MacroRunner(macro)
-	print m.cmd
-	#m.run('UCL')
-	
-	macro = 'egammaHistsToPostScript.C("cluster_validation.hist","refFile_cluster_validation.hist","cluster_validation.ps")'
-	m = MacroRunner(macro)
-	print m.cmd
-	#m.run('UCL')
-
-
-	macro = 'egammaHistsToPostScript.C("softE_validation.hist","refFile_softE_validation.hist","softE_validation.ps")'
-	m = MacroRunner(macro)
-	print m.cmd
-	#m.run('UCL')
-
-
-	macro = 'egammaHistsToPostScript.C("eGamma_validation.hist","refFile_eGamma_validation.hist","eGamma_validation.ps")'
-	m = MacroRunner(macro)
-	print m.cmd
-	#m.run('UCL')
-
diff --git a/Tools/RunTimeTester/src/MemoryUsageDiagnostic.py b/Tools/RunTimeTester/src/MemoryUsageDiagnostic.py
deleted file mode 100755
index 17910470bfaf612119866c66b2592448f76fe025..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MemoryUsageDiagnostic.py
+++ /dev/null
@@ -1,24 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from ShellCommand import ShellCommand
-class MemoryUsageDiagnostic:
-    def __init__(self, logger):
-        self.commands={
-            '/usr/bin/vmstat':    [],
-            '/usr/bin/vmstat -a': []
-            }
-        self.logger = logger
-
-    def execute(self):
-        for c in self.commands.keys():
-            sc = ShellCommand([c], self.logger)
-            reply = sc.getReply()
-            # [self.logger.debug(r) for r in reply]
-            self.commands[c].append(reply)
-            
-
-if __name__ == '__main__':
-
-    from PseudoLogger import PseudoLogger
-    md = MemoryUsageDiagnostic(PseudoLogger())
-    md.execute()
diff --git a/Tools/RunTimeTester/src/MethodTimer.py b/Tools/RunTimeTester/src/MethodTimer.py
deleted file mode 100755
index 76115f1b4ab6ed12aab7ef6f733a7cca5667a7d0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MethodTimer.py
+++ /dev/null
@@ -1,180 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"Use MethodTimer as a base class to profile an objects methods."
-
-
-import types, time
-
-def timing(f, name, self, tdict, ndict, cname):
-    def timed_f(*a, **k):
-        t0 = time.time()
-        result = f(self, *a, **k)
-        duration = time.time()-t0
-        key = '%s.%s' % (cname, name)
-        tdict[key]=tdict.get(key, 0)+duration
-        ndict[key]=ndict.get(key, 0)+1
-        # print 'inside wrapper for ',key,tdict[key],ndict[key]
-        return result
-    return timed_f
-
-def getAllBases(c, gen=0):
-    ngen = gen
-    # print c.__name__, gen
-    all       = [(gen,c)]
-    thisGener = [c]
-    parents   = [c.__bases__ for c in thisGener] # list of tuples
-
-    # print all
-    
-    if parents:
-        nextGener = []
-        ngen += 1
-        for tup in parents:
-            [all.append((ngen,e)) for e in tup]
-            [nextGener.append(e) for e in tup]
-
-        for c in nextGener:
-            all.extend(getAllBases(c, ngen))
-
-    return all
-                      
-        
-    
-    
-
-class MethodTimer:
-
-    def __init__(self):
-        """
-        calls 'timing' to obtain a function that wraps
-        the methods of this object between timing calls.
-        It also passes a dictionary tDict that stores the
-        cumulated time in each method.
-        This dictionery is an instance variable, not a class
-        variable
-        """
-        self.tDict = {}
-        self.nDict = {}
-        # print self.__class__.__dict__
-        classes = getAllBases(self.__class__)
-
-        # print classes
-        
-        ctor = [(g,c) for (g,c) in classes if c.__name__ == 'MethodTimer']
-        [classes.remove(c) for c in ctor]
-
-        temp = {}
-        for (a, b) in classes: temp[a]=b
-        classes = temp.items()
-
-        classes.sort()
-        classes.reverse()
-        # print classes
-        
-        for (g,c) in classes:
-            for f in c.__dict__.keys():
-                m= c.__dict__[f]
-                if isinstance(m, types.FunctionType):
-                    # print 'Wrapping method %s.%s' % (c.__name__, f)
-                    self.__dict__[f]=timing(m,
-                                            str(f),
-                                            self,
-                                            self.tDict,
-                                            self.nDict,
-                                            c.__name__
-                                            )
-
-    def printTimes(self): print self.tDict, self.nDict
-    def getMethodTime(self, methodName): return self.tDict.get(methodName, -1.)
-    def getTimedMethods(self): return self.tDict.keys()
-    def methodTimerReset(self):
-        for key in self.tDict.keys():
-            self.tDict[key] = 0
-
-    def formatMethodTimes(self):
-
-        # sort by time
-        temp = [(item[1], item) for item in self.tDict.items()]
-        temp.sort()
-
-        # report also the number of calls
-        text = ''
-        for i in temp:
-            item = i[1]
-            method = item[0]
-            time   = item[1]
-            ncalls = self.nDict[method]
-            method = method.ljust(50)
-            text += '%50s %15g %3d\n' % (method, time, ncalls)
-
-        return text
-
-    def printMethodTimes(self):
-        text=  self.formatMethodTimes()
-        print text
-        return text 
-
-        
-if __name__ == '__main__':
-    class TimedZ(MethodTimer):
-        def __init__(self, scale):
-            self.scale = scale
-            MethodTimer.__init__(self)
-            
-        def fz(self):
-            time.sleep(0.5*self.scale)
-        def fa(self):
-            print 'Z.a'
-        
-    class TimedY(TimedZ):
-        def __init__(self, scale):
-            TimedZ.__init__(self, scale)
-            
-        def fy(self):time.sleep(0.5*self.scale)
-
-    class TimedA(TimedY):
-        def __init__(self, scale):
-            TimedZ.__init__(self, scale)
-        def fa(self):
-            z = 1/0
-            time.sleep(1*self.scale)
-            print 'A.a'
-        def fb(self):time.sleep(2*self.scale)
-        def fc(self):time.sleep(3*self.scale)
-
-        def printSelf(self):
-            print self.printMethodTimes()
-   # print 'Execising Z'
-   # z = TimedZ(0.1)
-   # z.fz()
-   # z.printTimes()
-    
-    print 'Execising A'
-    a = TimedA(0.1)
-    a.fa()
-    a.fb()
-    a.fz()
-    a.fy()
-    a.fy()
-    a.printTimes()
-
-    a.printSelf()
-#    b= TimedA(0.3)
-#    b.fa()
-
-
-#    print 'Times for a (scale factor 0.1) for fa, fb'
-#    a.printTimes()
-#    print 'Times for b (scale factor 0.3) for fa'
-#    b.printTimes()#
-
-#    print 'second call to fa, 1st to fc'
-#    a.fa()
-#    a.fc()
-#    a.printTimes()#
-
-#    print 'zero, then fa,fc'
-#    a.zero()
-#    a.fa()
-#    a.fc()
-#    a.printTimes()
diff --git a/Tools/RunTimeTester/src/MethodTracer.py b/Tools/RunTimeTester/src/MethodTracer.py
deleted file mode 100755
index c083e4520f7ba0a002030dadc15040f40fe079d3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MethodTracer.py
+++ /dev/null
@@ -1,60 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Taken verbatim from the python cookbook (recipe 15.11)
-Monitors entering and leaving of object methods
-"""
-
-
-import types
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def tracing(f, name):
-    def traced_f(*a, **k):
-        msg = '%s(%s,%s)->'%(name, a, k)
-        result = f(*a, **k)
-        msg += str(result)
-        # logger.info(msg)
-        # print '-----------------------------------------------'
-        # print msg
-        # print '-----------------------------------------------'
-        logger.debug(msg)
-        return result
-    return traced_f
-
-class MetaTracer(type):
-    def __new__(self, classname, bases, classdict):
-        for f in classdict:
-            m = classdict[f]
-            if isinstance(m, types.FunctionType):
-                classdict[f]=tracing(m, '%s.%s'%(classname,f))
-        return type.__new__(self, classname, bases, classdict)
-
-class Tracer:
-    __metaclass__ = MetaTracer
-
-if __name__ == '__main__':
-    from setupLoggerForDebug import setupLoggerForDebug
-    
-    setupLoggerForDebug()
-    # give a __metaclass__ to all classes without one
-    __metaclass__ = MetaTracer
-
-    # a test class
-    class A:
-        def __init__(self, j): self.j=j
-        def a(self):
-            return 'hello'
-
-        def b(self, i):
-            return i+self.j
-
-    
-    a = A(23)
-    # demonstrate tracing
-    a.a()
-    a.b(10)
-
diff --git a/Tools/RunTimeTester/src/Minder.py b/Tools/RunTimeTester/src/Minder.py
deleted file mode 100755
index e9603aab6ec0454fa37c112e304396dce6083549..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Minder.py
+++ /dev/null
@@ -1,1000 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-class Minder. Top of the Minder Heirarchy.
-
-- Takes care of the state machine. This is a mixture of
-a status object and a dispatcher.
-
-The state machine  has states: queued, running, success, error,
-it nows how to proceed from one state to the next, and species the actions to be
-performed when the state changes. The way jobs are submited, and the names of
-log files are examples of items that change with mode.
-
-This class also has the following methods, which are common to all Minders,
-
-poll
-fromQueued
-fromRunning
-fromSuccess
-fromError
-setStatus
-setStatusList
-forceCloseDown
-setDone
-
-Many of these are overidden in derived classes.
-"""
-
-from MinderBase                     import MinderBase
-from formatCollection               import formatCollection
-from Defs                           import RTTdefs
-from ModuleLoader                   import ModuleLoader
-from exc2string2                    import exc2string2
-try:
-    from RTTTestRunner                  import RTTTestRunner  
-except Exception, e:
-    print str(e)
-from UserLogger                     import UserLogger
-from Tools                          import unique, timed_cmd
-from Tools2                         import toRelPath, popenCmd
-from BigFileIO                      import createBigFileIO
-from RTTdict                        import RTTdict
-from RTTSException                  import RTTCodingError
-from DDDescriptor                   import DDDescriptor
-from PPPaths                        import PPPaths
-
-import logging
-logger = logging.getLogger('rtt')
-
-import os, shutil, fnmatch, shelve, time
-
-# -------------------------------------------------------------------------
-#from Tracer        import Tracer
-# uncomment to monitor all method calls (shows args + return values)
-#__metaclass__ = Tracer
-
-
-from MethodTimer import MethodTimer
-
-
-
-class Minder(MinderBase):
-    class DiskSpaceCalculator:
-        def __init__(self, runpath, respath, logger):
-            self.runpath = runpath
-            self.respath = respath
-            self.logger  = logger # minder logger
-            self.run = []
-            self.res = []
-
-        def du(self, path):
-            try:
-                ans = timed_cmd('du -hs %s' % self.runpath)
-                return ans[0].split()[0].strip()
-            except:
-                self.logger.error('Unable to du %s' % path)
-                self.logger.error('Answer was: %s' % ans)
-                self.logger.error(exc2string2())
-                return '-1'
-            
-        def __call__(self, phase):
-            self.run.append((self.du(self.runpath), time.time(), phase))
-            self.res.append((self.du(self.respath), time.time(), phase))
-
-    def __init__(self, argBag, jDescriptor):
-        """ One function of the minder is to bring together information
-        from the job descriptor (gathered from the jobs config file),
-        the job groups file (gathered from the job groups config file),
-        and the Paths object (gathered from the RTT config file)
-
-        For files to be transferred from the run directory to the results
-        area, we use patterns (for file name matching). Lists of patterns
-        are made at in this method, with a list per use (keep files). These lists will be used to identify
-        filenames at the appropriat e moment, and converted to absolute
-        path names. Source and destination will be kept in the dictionaries
-        self.keepFiles
-        
-        """
-
-
-
-        MinderBase.__init__(self, argBag.logDir,
-                            argBag.package,
-                            jDescriptor.identifiedName,
-                            jDescriptor.jobSerialNumber,
-                            argBag.elementCreator,
-                            argBag.textNodeCreator)
-
-        self.shareArea           = argBag.shareArea
-        self.deleteNonKeepFiles  = argBag.deleteNonKeepFiles
-        self.confFile            = argBag.confFile
-        self.resultsBasePath     = argBag.resultsBasePath
-        self.isNightly           = argBag.isNightly
-        self.workDirs            = argBag.workDirs
-        self.site                = argBag.site
-
-        self.RTTLibDir           = argBag.RTTLibDir         # used for post processing
-        self.RTTSrcDir           = argBag.RTTSrcDir         # used for post processing
-        self.dCubeCfgFile        = argBag.dCubeCfgFile      # used for post processing
-        self.installArea         = argBag.installArea       # used for post processing
-        self.containerPackage    = argBag.containerPackage  # used for post processing
-        self.runType             = argBag.runType           # used for post processing
-        self.localRTTRun         = argBag.localRTTRun       # used for post processing
-        self.distArea            = argBag.distArea          # used for post processing
-        self.cmtPath             = argBag.cmtPath           # used for post processing
-        self.cmtConfig           = argBag.cmtConfig         # used for post processing
-        self.cmtLinesCmds        = argBag.cmtLinesCmds      # used for post processing
-        self.topProject          = argBag.topProject        # used for post processing
-        self.branch              = argBag.branch            # used for post processing
-        self.release             = argBag.release           # used for post processing
-        self.package             = argBag.package           # used for post processing
-        self.archivers           = [a.duplicate() for a in argBag.archivers]
-        
-        self.runPath             = jDescriptor.runPath
-        self.name                = jDescriptor.name
-        self.jobDisplayName      = jDescriptor.jobDisplayName
-        self.jobGroup            = jDescriptor.jobGroup
-        self.jobDocString        = jDescriptor.jobDocString    
-        self.jobDocURL           = jDescriptor.jobDocURL    
-        self.rttPilotJob         = jDescriptor.rttPilotJob
-        self.rttATNJob           = jDescriptor.rttATNJob
-        self.log                 = jDescriptor.log
-        self.elog                = jDescriptor.elog
-        self.errorMessages       = jDescriptor.errorMessages
-        self.resPath             = jDescriptor.resPath
-        self.descDataForXMLNode  = jDescriptor.dataForXMLNode()
-        self.keepFiles           = jDescriptor.keepFiles
-        self.datasets            = jDescriptor.datasets
-        self.missingDatasets     = jDescriptor.missingDatasets
-        self.trendId             = jDescriptor.trendId
-        self.batchWallTimeLimit  = jDescriptor.suggestedBatchWallTime
-
-        # Combine some stuff from kit and job
-        self.keepFilePatterns    = jDescriptor.keepFilePatterns + argBag.jobGroupKit['keepFilePatterns']
-        self.auxFilePatterns     = jDescriptor.auxFilePatterns + argBag.jobGroupKit['auxFilePatterns']
-        self.actionDescriptors   = jDescriptor.actions + argBag.jobGroupKit['actions']
-        self.testDescriptors     = jDescriptor.tests + argBag.jobGroupKit['tests']
-        
-        self.testDBPath          = os.path.join(self.runPath, 'RTTtests.db')
-
-        # This list holds those files that should never be copied or deleted from the run dir by this minder
-        # Currently it will hold chain files (absolute paths).
-        self.neverCopyAndNeverDelete = []
-
-        # Filled in by the archivers
-        self.keepFilesToVeto = []
-        
-        # Calculate disk space usage by the job. Stores history of usage by the job.
-        self.diskspace = Minder.DiskSpaceCalculator(self.runPath,self.resPath, self.logger)
-
-        # These jobs are NOT counted in  Launcher stats. Otherwise overide.
-        self.jobWeight = 0 
-
-        # bring together information on files to manipulate.
-        # Keep and aux files are handled as patterns - the actual file
-        # name resolution is deferred to later
-        # Reference files references are complete file names due to
-        # fact that they will go into keys before retirval - and patterns wont work
-        
-
-        self.finishedTests        = [] # a list of execcuted test objects.
-        self.checkResults         = [] # results of tests which have been shelved.
-        self.processingResult     = 'unavailable'
-        self.postProcessingResult = 'unavailable'
-        self.chainLength          = 0
-        self.chainSuccesses       = 0
-        self.chainFileCopier      = None # Might be reset by derived class
-        
-        # values for the commands are  worked out in derived class
-        self.postProcessCommand   = argBag.postProcessCommand
-        self.submitCommand        = argBag.submitCommand
-        
-        # declare dictionaries with source and destination paths
-        # for files to be transferred from the run directories to the resulsts
-        # directory.
-
-        # declare lists of files - no  wild cards, as the elements of the list
-        # are used in data base keys. Also,  no absolute paths for the same reason
-        # (would break portability)
-        
-        
-        # maximum no of times to retry a job if it enters the error state.
-        self.errorStateMaxRetry = 3
-        self.errorStateCurRetry = 0
-          
-
-        self.exitStatus  = 'Unknown'
-        self.cpuTime     = 'Unknown'
-        self.cpuTime2000 = 'Unknown'
-        self.mem         = 'Unknown'
-        self.vmem        = 'Unknown'
-        self.wallTime    = 'Unknown'
-        self.batchStatus = 'Unknown'
-        
-
-        self.scripts             = {}
-        self.timedOut            = False
-        # so complete file names go here here as well.
-
-        self.handleFileNames()
-        self.collectActionsAndTests()
-
-        self.collectScripts(argBag.scriptWriter) # after collecting tests and actions
-        # self.setupJobGroupDirs() # after making scripts
-
-   
-        jobGroupDirectoryMaker        = argBag.jobGroupDirectoryMaker
-        jobGroupDirectoryMaker.logger = self.logger
-
-        self.makeJobGroupDirs(jobGroupDirectoryMaker)
-        self.setupJobGroupDirs(jobGroupDirectoryMaker)
-
-        self.shelveTests()
-
-    
-    def makeJobGroupDirs(self, jobGroupDirectoryMaker):
-        try:
-            jobGroupDirectoryMaker.makeJobGroupDirs(self.runPath, self.resPath)
-        except:
-            msg  = 'Exception making directories by %s traceback:\n%s' % (self.identifiedName, exc2string2())
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-        
-
-    def unlink(self):
-        'break circular references to allow garbage collection'
-
-        self.stateEngine.unlink()
-        del self.stateEngine
-        self.xmlConverter.unlink()
-        del self.xmlConverter
-        MinderBase.unlink(self)
-
-        
-    def calcPostProcessingResult(self):
-        if not (self.finishedTests or self.checks):
-            self.postProcessingResult = 'no tests'
-            return
-
-        for t in self.finishedTests:
-            if t.result != 0:
-                self.postProcessingResult = 'error'
-                return
-
-        if 'error' in self.checkResults:
-            self.postProcessingResult = 'error'
-            return
-        
-        self.postProcessingResult = 'success'
-
-    def collectScripts(self, scriptWriter):
-        # post processing scripts are handled here, and so
-        # apply to Watcher and Worker minders.
-        # NOTE 26/01/09
-        # self.postProcessingCommand is minder specific.
-        # it is set in LinuxInteractive and LSFBatch, not yet set in WatcherMinder
-        
-        if not (self.tests or self.actions): return
-
-        script = scriptWriter.postProcessorScript()
-
-        if not self.postProcessCommand:
-            msg = '%s using self.postProcessing command without it having been set to something useful' % self.__class__.__name__
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-        
-        dict = {'postProcessorScript': {'script':     script,
-                                        'scriptName': 'postProcessorScript.sh',
-                                        'cmd':         self.postProcessCommand
-                                        }
-                }
-        self.scripts.update(dict)
-
-
-
-# ------------------------------------------------------------------------
-
-    def shelveTests(self):
-        # associate string identifiers with each test and action
-        # t.position is the user given order index
-        npos = 0
-
-        testIds = []
-        for runOrder, instanceList in self.tests.items():
-            for testInstance in instanceList:
-                item = (testInstance, '%s_%d' % (testInstance.__class__.__name__,npos), runOrder)
-                testIds.append(item)
-                npos +=1
-
-        actionIds = []
-        for runOrder, instanceList in self.actions.items():
-            for actionInstance in instanceList:
-                item = (actionInstance, '%s_%d' % (actionInstance.__class__.__name__,npos), runOrder)
-                actionIds.append(item)
-                npos +=1
-                
-        if npos == 0: return
-
-        # provide a database to store the tests
-        db = shelve.open(self.testDBPath, 'n')
-        db.close()
-
-        # wrap the tests to allow transportation to the computing node
-        wrappers = []
-        
-        isAction = False
-        wrappers.extend([RTTTestRunner(self.testDBPath, t, isAction) for t in testIds])
-        isAction = True
-        wrappers.extend([RTTTestRunner(self.testDBPath, t, isAction) for t in actionIds])
-
-        # Write the tests to a shelve db
-        for w in wrappers:
-            try:
-                w.autoShelve()
-            except:
-                m = 'Error shelving %s\nTraceback: %s' % (w.testId, exc2string2())
-                self.logger.error(m)
-        
-    # -------------------------------------------------------------------------
-    # 09/03/21 PS This method is never called 
-    # def hasChecks(self):
-    #    answer = False
-
-    #    if hasattr(self,'checks'):
-    #        answer = len(self.checks) > 0 or answer
-
-    #    if hasattr(self,'tests'):  
-    #        answer = len(self.tests.keys()) > 0 or answer
-
-    #    return answer
-
-# ------------------------------------------------------------------------
-
-    def collectActionsAndTests(self):
-        """
-        Test Descriptor is a list with four elements:
-        position, modulename, classname, dictionary of arguments
-        """
-
-
-        # set up a "pruned descritpor". In 01/09, there was a large refactoring of the RTT
-        # and the role of paths and descriptors was reduced. However user suppied test scripts may
-        # be expecting these variables, so they are rebuilt here.
-        # class DDDescriptor:pass
-        # class Paths: pass
-        def makePrunedDescriptor():
-            desc = DDDescriptor()
-            possibleAttrs = ['log', 'name', 'trendId', 'identifiedName', 'jobSerialNumber', 'jobDisplayName',
-            'hashString', 'runPath', 'resPath', 'jobGroup', 'jobDocString', 'jobDocURL', 'package']
-            [setattr(desc, p, getattr(self, p)) for p in possibleAttrs if hasattr(self, p)]
-
-
-            paths = PPPaths()
-            possibleAttrs =  ['installArea', 'containerPackage', 'cmtPath', 'shareArea',
-                              'release', 'cmtHomeDir', 'localRTTRun', 'workDirs', 'resultsDirs', 'runType',
-                              'build', 'originalBranch', 'branch', 'topProject', 'otherProject', 'cmtConfig',
-                              'dCubeCfgFile', 'isNightly', 'RTTSrcDir', 'RTTLibDir', 'pathSegment', 'package',
-                              'packageTag', 'fullPackageName', 'dCubeRefBaseDir', 'confFile', 'userHomeCmtDir',
-                              'testProjectCmtDir', 'distArea', 'projectCMTDirs']
-            [setattr(paths, p, getattr(self, p)) for p in possibleAttrs if hasattr(self, p)]
-
-            desc.paths = paths
-            return desc
-
-        def process(descs, tgt):
-            for descriptor in descs:
-                # Create the pruned descriptor
-                prunedDesc  = makePrunedDescriptor()
-                # prunedDesc  = self.jDescriptor.prune()
-                # prunedPaths = self.jDescriptor.paths.prune()
-                # setattr(prunedDesc, 'paths', prunedPaths)
-
-                # Add params to the dict
-                descriptor.addParameter(('JobDescriptor', prunedDesc))
-                descriptor.addParameter(('cmtLinesCmds', self.cmtLinesCmds[:]))
-                instance = self.getClassFromDescriptor(descriptor)
-                if instance != None:
-                    tgt.setdefault(descriptor.position, []).append(instance)
-
-        process(self.testDescriptors, self.tests)
-        process(self.actionDescriptors, self.actions)
-          
-#-----------------------------------------------------------------------
-
-    def getClassFromDescriptor(self, taDescriptor): # descriptor is a test or action descriptor
-        self.logger.debug("Entering getClassFromDescriptor: %s" % taDescriptor.argDict)
-        def runUserScriptInAtlasEnv():
-            return taDescriptor.runInAtlasEnv!=""
-        def createRTTdict(rtt_dict_keys, paramDict):
-            self.logger.debug('Inside createRTTdict, paramDict: %s, rtt_dict_keys: %s' % (paramDict,
-                                                                                          rtt_dict_keys))
-            rtt_dict = {}
-            for key in rtt_dict_keys:
-                rtt_dict[key] = paramDict[key]
-                del paramDict[key]
-            user_dict = paramDict
-            return RTTdict(rtt_dict, user_dict)
-        
-        # allow RTT library tools to be picked up from the local code base
-        # rather than the installed shared area if so requested in the
-        # RTT configuration file
-        sharePathToUse = {'RttLibraryTools' : self.RTTLibDir}
-        instance = None
-        self.logger.debug("share path to use is: %s" % sharePathToUse)
-
-        # RTTLibTools has been split into modules with one class per
-        # module, with the module name = the class name. Needed this
-        # for shelve to unshelve instances.
-
-        moduleName = taDescriptor.moduleName
-        sharePathIs = sharePathToUse.get(moduleName, self.shareArea)
-        m = 'dict: %s, key: %s, value: %s' % (sharePathToUse, moduleName, sharePathIs)
-        self.logger.debug(m)
-        
-        # now set module name = testname for shelving constraint
-        moduleName = taDescriptor.testName
-        className  = taDescriptor.testName
-        self.logger.debug("modulename set = classname, i.e. %s" % moduleName)
-        paramDict  = taDescriptor.argDict
-
-        self.logger.debug("paramdict: %s" % paramDict)
-        
-        rtt_dict_keys = ['JobDescriptor']
-
-        if runUserScriptInAtlasEnv():
-            sharePathIs = sharePathToUse.get('RttLibraryTools')
-            paramDict['userScript'] = moduleName
-            paramDict['userScriptLoc'] = sharePathToUse.get(taDescriptor.moduleName,
-                                                            self.shareArea)
-            moduleName = className  = 'AtlasEnvSetup'
-            
-            rtt_dict_keys.extend(['userScript', 'userScriptLoc'])
-          
-        paramDict = createRTTdict(rtt_dict_keys, paramDict)
-        
-        # pickle the dictionary to the run path: actually gets dumped in JobGroupDirectoryMaker.setupRunDir
-        self.rttArgDictPickle = {'what': paramDict, 'where': os.path.join(self.runPath,'rtt.argdict.cpickle')}
-        
-        # default is release share path if key not in dict
-        # this happens when the user supplies own module, which is
-        # picked up from the release area.
-
-        self.logger.debug('module path ' + sharePathIs)
-        self.logger.debug('moduleName  ' + moduleName)
-        self.logger.debug('className   ' + className)
-        self.logger.debug('paramDict   ' + str(paramDict))
-        self.logger.debug('logger      ' + str(self.logger))
-
-        try:            
-            mL        = ModuleLoader(moduleName, sharePathIs, self.logger)
-            instance  = mL.getClassInstance(className, paramDict)
-            self.logger.debug('Created a test instance of class %s' % (instance.__class__.__name__))
-        except Exception, e:
-            self.logger.error('Failed to create a test instance')
-            self.logger.error(exc2string2())
-            self.logger.error(str(e))
-            
-        return instance
-    
-#------------------------------------------------------------------------
-
-    def handleFileNames(self):
-        """ method to collect file names, and resolve patterns where
-        for those patterns for which it is possible at init."""
-   
-        # expand the wild cards - but do not create the full directory path
-        # as the work sub directories have yet to be created.
-        if not os.path.exists(self.shareArea):
-            m = 'Cannot set self.auxfiles due to non-existent share directory: %s' % self.shareArea
-            self.logger.fatal(m)
-            raise RTTCodingError(m)
-
-        # resolve auxFile patterns to file names
-        auxFiles = []
-        for pattern in self.auxFilePatterns:
-            base, fnpattern = os.path.split(pattern)
-            srcDir = os.path.normpath(os.path.join(self.shareArea, base))
-            filesInShare = os.listdir(srcDir)
-            auxFiles.extend([os.path.join(base,file) for file in filesInShare if fnmatch.fnmatch(file, fnpattern)])
-
-        self.auxFiles = unique(auxFiles)
-
-    # ------------------------------------------------------------------------
-    # 09/03/31 This method is never called from RTT code base
-    # def reportUserScriptError(self,message):
-    #    filePath  = os.path.join(self.runPath,'Python_Script_Output.log')
-    #    h = open(filePath,'a+')
-    #    h.write(message)
-    #    h.close()
-
-    #-----------------------------------------------
-    
-    def readTestResults(self):
-        #make UserLogger to transfer messages
-        #from UserLogger import UserLogger
-        ulogger = UserLogger().makeLogger(self.runPath, self.identifiedName)
-        self.logger.debug('made logger for tests and actions :'+str(logger))
-
-        db = shelve.open(self.testDBPath)
-
-        dbValues = []
-        try:
-            dbValues.extend(db.values())
-            self.finishedTests = db.values()
-        except Exception, e:
-            m  = 'Exception thrown trying to read values from %s\n' % str(self.testDBPath)
-            m += str(e)            
-            self.logger.error(m)
-            # now close DB to prevent too many open files
-            db.close()
-        else:
-            db.close()
-        
-        self.finishedActions = [t for t in self.finishedTests if t.isAction]
-        [self.finishedTests.remove(t) for t in self.finishedActions]
-        
-        # for t in db.values():
-        for t in dbValues:
-            if t.error:
-                self.logger.error('Error running test %s' % t.testId)
-            else:
-                self.logger.debug('No error running test %s' % t.testId)
-
-
-            # collect any messages generated by the test
-            t.transferLog(ulogger)
-
-            delimeter  = "\n***********************************\n"
-            delimeter += "           NEW ACTION/TEST         \n"
-            delimeter += "***********************************\n"
-            ulogger.debug(delimeter)
-
-        self.calcPostProcessingResult()
-       
-    #--------------------------------------------------
-    
-    #def runMoniActions(self):
-    #jobName  = self.jDescriptor.identifiedName
-    #nActions = len(self.monActions)
-    #msg      = 'Running %d monActions for job %s' % (nActions, jobName)
-    #self.logger.debug(msg)
-    #
-    #for action in self.monActions:
-    #className = str(action.__class__.__name__)
-    #try:
-    #self.logger.debug("Running monAction " + className)
-    #dataDict=action.run()
-    #self.monActionsData.append(dataDict)
-    #
-    # except Exception, e:
-    # msg = "Could not run monitoring action " + className
-    # self.logger.error(msg)
-    # self.logger.error(str(e))
-    # self.logger.error(exc2string2())
-    # msg  = '-----------------------------------------\n'
-    # msg += 'MoniAction: ' + className + ' could not be run!\n'
-    # msg += exc2string2() + '\n'
-    # msg += '-----------------------------------------\n\n'
-    # self.reportUserScriptError(msg)
-    # self.logger.debug("Running next available monAction")
-    #       
-    #
-    # self.logger.debug('Minder moniData :'+str(self.monActionsData))
-    # self.logger.debug("Finished running monActions for %s" % jobName)
-      
-    #--------------------------------------------------
-
-    def fullResultsFileName(self, filename):
-        return os.path.join(self.resPath, filename)
-
-    #-----------------------------------------------
-    
-    def makeKeepFileEntry(self, file, infoString='',displayColor='', md5=''):
-        "helper method for registerWildKeepFiles()"
-        
-        src = os.path.join(self.runPath, file)
-
-        dest = {'keepFileString': self.fullResultsFileName(file),
-                'infoString'    : infoString,
-                'displayColor'  : displayColor,
-                'md5sum'        : md5}
-        
-        self.keepFiles[src]=dest
-        return dest['keepFileString']
-    #-----------------------------------------------
-    
-    def registerWildKeepFiles(self):
-        """
-        Common implementation task.
-
-        Obtain the wild card patterns for the current job group from
-        a JobGroupKit.
-        
-        Collect all the wild carded keep files.
-        These are files that are present in the run directory at when
-        fromRunning is called, and which match a pattern in the pattern
-        list.
-
-        Give these files their full paths and add them to the keepFile
-        dictionary.
-        """
-
-        for card in self.keepFilePatterns:
-
-            keepString   = card['keepFileString'] # may contain subdirs
-            infoString   = card['infoString']
-            displayColor = card['displayColor']
-
-            keepStringTokens = keepString.strip().split('/')
-
-            if len(keepStringTokens) == 1: # current dir keepfile pattern
-                wildFiles = fnmatch.filter(os.listdir(self.runPath), keepString)
-                [self.makeKeepFileEntry(file, infoString, displayColor) for file in wildFiles]
-
-            elif len(keepStringTokens) > 1: # subdirs
-                matches = ['']
-                for pathPart in keepStringTokens:
-                    newMatches = []
-                    for match in matches:
-                        conts = os.listdir(os.path.join(self.runPath, match))
-                        newMatches.extend([os.path.join(match, f) for f in fnmatch.filter(conts, pathPart)])
-
-                    matches = newMatches
-                [self.makeKeepFileEntry(m, infoString, displayColor) for m in matches]
-
-        # now manually add the package configuration file to keep files
-        self.makeKeepFileEntry(os.path.basename(self.confFile),
-                               "Package XML test configuration file",
-                               "#cc3333")
-
-
-    #-----------------------------------------------
-
-    def cleanSpace(self):
-        self.removeNonKeepFiles()
-
-        # If runpath != respath we need to delete keepfiles in the run dir
-        # else they'll stick around.
-        if self.runPath != self.resPath:
-            allfiles = [os.path.join(self.runPath, o) for o in os.listdir(self.runPath)]
-            toDelete = [a for a in allfiles if a not in self.neverCopyAndNeverDelete]
-            self.logger.debug('Runpath!=Respath, deleting %d keepfiles from run dir' % len(toDelete))
-            self.logger.debug('Files to delete:\n%s' % str(toDelete))
-            for a in toDelete:
-                os.system('rm -rf %s' % a) 
-
-    # --------------------------------------------------------------------
-
-    def getBigFiles(self):
-        """Looks in the run/results dir for big files. Returns a list of them."""
-        toMove = {self.resPath:[], self.runPath:[]}
-
-        for tgt in toMove.keys():
-            for root, dirs, files in os.walk(tgt):
-                files = [os.path.join(root, f) for f in files]
-                def isBig(f): # only files, no links
-                    return os.path.isfile(f) and not os.path.islink(f) and os.path.getsize(f)>=self.bigFilesSize
-                toMove[tgt].extend([f for f in files if isBig(f)])
-        return toMove
-    
-    # --------------------------------------------------------------------
-
-    def moveBigFile(self, src, dst):
-        try:
-            fileSize = os.path.getsize(src)
-            shutil.move(src, dst)
-            m = 'Successfully moved big file %s [%s bytes] to %s' % (os.path.basename(src),
-                                                                     str(fileSize),
-                                                                     dst)
-            self.logger.info(m)
-        except:
-            m  = 'Unable to move big file %s to %s\n' % (os.path.basename(src), dst)
-            m += exc2string2()
-            m += str(os.listdir(os.path.dirname(src)))
-            self.logger.error(m)
-            
-    # --------------------------------------------------------------------
-
-    def makeReplacementKeepFile(self, bigFileSrc, bigFileDump):
-        """If we move a big file from run path to separate volume,
-        and that file just happened to be a keep file, then we need
-        to replace it. Create a soft link in resPath to separate volume
-        and add an expiry date."""
-
-        linkName = os.path.basename(bigFileSrc)
-        linkTgt  = os.path.join(toRelPath(os.path.dirname(bigFileSrc), bigFileDump), linkName)
-        
-        currDir = os.getcwd()
-        os.system('cd %s;ln -s %s %s;cd %s' % (os.path.dirname(bigFileSrc), linkTgt, linkName, currDir))
-
-        # add an md5sum to the keepfile to guard against mishap if the file pointed
-        # at was overwritten (the web link would work, but you'd get the
-        # overwriter, not the expected overwritee, and you would never know...
-        # If md5sum different between now calculated and future web page one,
-        # then grey-out the link.
-        md5sum = popenCmd('md5sum %s' % bigFileSrc)
-        self.logger.debug('Minder md5sum calculated is: %s' % str(md5sum))
-        md5sum = md5sum[0].split()[0].strip()
-        srcRelToRunPath = bigFileSrc.split(self.runPath)[1]
-        if srcRelToRunPath.startswith('/'): srcRelToRunPath = srcRelToRunPath[1:]
-        self.makeKeepFileEntry(srcRelToRunPath, md5=md5sum)
-        
-    # --------------------------------------------------------------------
-    
-    def moveBigFiles(self):
-        """Move all files bigger than some value onto a separate volume."""
-        if not self.bigFilesArea:
-            self.logger.info('Moving of big files to a separate volume has not been requested.')
-            return
-
-        self.logger.info('Moving of big files to a separate volume is requested. Scanning...')
-        
-        if not os.path.exists(self.bigFilesArea):
-            m = 'Cannot shift big files onto inexistent volume: %s' % self.bigFilesArea
-            self.logger.error(m)
-            return
-        
-        bigFiles = self.getBigFiles()
-
-        if not [val for val in bigFiles.values() if val]:
-            self.logger.info('No big files were found, returning.')
-            return
-        
-        placeToDump = createBigFileIO(self.site, self.bigFilesArea, self.workDirs, self.isNightly).getJobDumpLocation(self)
-        if not placeToDump:
-            m = 'Unable to retrieve location of big files volume. Not moving big files.'
-            self.logger.warning(m)
-            return
-
-        # We have files to move, let's move them
-        for bigFileBaseDir, bigFiles in bigFiles.items():
-            for bigFile in bigFiles:
-                src = bigFile # file
-                dst = placeToDump # directory
-                self.moveBigFile(src, dst)
-                # If big file origin is results path, replace with a soft link
-                # to separate big file volume.
-                if bigFileBaseDir == self.resPath:
-                    self.makeReplacementKeepFile(bigFile, placeToDump)
-            
-    # --------------------------------------------------------------------                            
-
-    def removeNonKeepFiles(self):
-        """Delete all files in the run dir not explicitly kept."""
-
-        if not self.deleteNonKeepFiles:
-            self.logger.info('Deletion of non keep files (to save space) has not been requested.')
-            return
-
-        self.logger.info('Deletion of non-keep files is requested. Scanning...')
-        
-        toDelete = []
-        for root, dirs, files in os.walk(self.runPath):
-            toDelete.extend([os.path.join(root, f) for f in files if os.path.join(root, f) not in self.keepFiles.keys()])
-            # os.walk ignores links, explicitly get them
-            toDelete.extend([os.path.join(root, f) for f in os.listdir(root) if os.path.islink(os.path.join(root,f)) and os.path.join(root,f) not in self.keepFiles.keys()])
-
-        for thing in os.listdir(self.runPath):
-            self.logger.debug('removeNonKeepFiles::before delete: %s' % str(thing))
-            
-        for thing in toDelete:
-            if thing in self.neverCopyAndNeverDelete:
-                self.logger.debug('removeNonKeepFiles::Not deleting neverCopyAndNeverDelete file: \n%s' % thing)
-                continue
-            try:
-                os.remove(thing)
-                self.logger.debug('Deleted: %s' % thing)
-            except:
-                message  = 'Unable to delete non-keepfile: %s\n' % thing
-                message += exc2string2()
-                self.logger.error(message)
-
-    # --------------------------------------------------------------------
-    
-    # copy files to be kept (log histos, ntuples...) to results dir
-    def copyKeepFiles(self):
-        
-        # find if any of the wild keep cards have a match
-        idName = str(self.identifiedName)
-        try:
-            self.registerWildKeepFiles()
-        except Exception, e:
-            msg = "Exception registering keep files for job: " + idName
-            self.logger.error(msg)
-            self.logger.error(str(e))
-            self.logger.error(exc2string2())
-
-        # Pop files in self.keepFiles dict that are in the self.keepFilesToVeto list
-        popped = [self.keepFiles.pop(k) for k in self.keepFiles.keys() if k in self.neverCopyAndNeverDelete]
-        self.logger.debug('%d neverCopyNeverDelete files were popped from the keepfile dict.' % len(popped))
-        self.logger.debug('%s' % str(popped))
-
-        popped = [self.keepFiles.pop(k) for k in self.keepFiles.keys() if k in self.keepFilesToVeto]
-        self.logger.debug('%d keepFilesToVeto files were popped from the keepfile dict.' % len(popped))
-        self.logger.debug('%s' % str(popped))
-
-        msg  = 'About to transfer %d keep files' % len(self.keepFiles.keys())
-        self.logger.debug(msg)
-
-        for file in self.keepFiles.keys():
-            if not os.path.exists(file):
-                msg = "%s does not exist! Skipping transfer." % str(file)
-                self.logger.error(msg)
-                continue
-                
-            dictValue = self.keepFiles[file]
-            desFile = dictValue['keepFileString'] # this is a filename
-
-            fileName = os.path.basename(str(file))
-            srcDir   = os.path.dirname(str(file))
-            tgtDir   = os.path.dirname(str(desFile))
-
-            if srcDir == tgtDir:
-                msg  = 'Keepfile src and tgt dirs are same.\n'
-                msg += '%s\n' % srcDir
-                msg += '===> Not transferring file %s' % fileName
-                self.logger.debug(msg)
-                continue
-
-            self.logger.debug("Moving file %s from %s to %s" % (fileName, srcDir, tgtDir))
-            if not os.path.exists(os.path.dirname(desFile)):
-                self.logger.debug('Destination dir does not exist. Making: %s' % os.path.dirname(desFile))
-                try:
-                    os.makedirs(os.path.dirname(desFile))
-                except:
-                    self.logger.error('Unable to create %s so as to copy keep file to it. Not copying it.' % os.path.dirname(desFile))
-                    self.logger.error(exc2string2())
-                    continue
-
-            try:                    
-                shutil.copy(file, os.path.dirname(desFile))
-            except:
-                # disk space problems?
-                message = '***** COPYING OF KEEP FILES PROBLEM! *****\n'
-                message += 'Unable to copy src file:\n'
-                message += '   ' + file + '\n'
-                message += 'to destination file:\n'
-                message += '   ' + desFile + '\n'
-                message += 'Doing command df on ' + desFile + ' yields the answer:\n'
-                message += str(os.popen('df ' + desFile).read())
-                message += '---------------------------------'                    
-                self.logger.error(message)
-
-        self.logger.debug("Finished copying keepfiles")
-
-    # --------------------------------------------------------------------
-        
-    def runChecks(self):
-        
-        theJob = str(self.identifiedName)
-        msg = 'Len self.checks is: %s for %s ' % (str(len(self.checks)),theJob)
-        self.logger.debug(msg)
-        
-        for check in self.checks:
-
-            check.setLogger(self.logger) # log to the minders logfile
-
-            if not callable(check):
-                msg = "Uncallable checker: %s for job %s" % (str(check),
-                                                             theJob)
-                self.logger.error(msg)
-                self.checkResults.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            self.logger.debug('Job: %s calling check: %s'%(theJob, str(check)))
-            status = ''
-            try:
-                rc = check(self) # int return code from check
-                status = RTTdefs.status(rc) # convert to string
-            except Exception, e:
-                msg = 'Exception raised while executing %s' % str(check)
-                self.logger.error(msg)
-                self.logger.error(exc2string2())
-                self.logger.error(str(e))
-                self.checkResults.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            msg = 'Job: %s  check: %s return status %s' % (theJob,
-                                                           str(check),
-                                                           status)
-            self.logger.debug(msg)
-            
-            self.checkResults.append(status)
-        
-        self.logger.debug(msg)
-
-    # --------------------------------------------------------------------       
-    """
-    success if the job terminates, subclasses will override with a
-    more precise meaning
-    """
-    def isSuccess(self): return self.isDone()
-        
-    # --------------------------------------------------------------------
-
-    def setupJobGroupDirs(self, jobGroupDirectoryMaker):
-        try:
-            jobGroupDirectoryMaker.setupRunDir(self)
-        except:
-            msg  = 'Error  setting up run directory by %s: TraceBack:\n%s' % (
-                self.identifiedName,
-                exc2string2()
-                )
-            self.logger.warning(msg)
-            raise RTTCodingError(msg)
-
-
-    def dataForMonitoring(self):
-        """
-        return a dictionay of values for monitoring. repeats data in DOM
-        document, but is faster to access.
-        """
-        dict = MinderBase.dataForMonitoring(self)
-        
-        dict['nTests']        = len(self.tests.keys())
-        dict['done']          =   self.isDone()
-        dict['nTestsSuccess'] = len([s for s in self.finishedTests if s.result == 0])
-        dict['nTestsFailure'] = len([s for s in self.finishedTests if s.result != 0])
-        dict['nRetries']      = self.errorStateCurRetry
-        dict['ppFailure']     = (self.postProcessingResult == 'error')
-        dict['ppSuccess']     = (self.postProcessingResult == 'success')
-
-        return dict
-    def __str__(self):
-        s = '----------- Minder ---------------\n'
-        s += ' done:                 '+str(self.isDone())+'\n'
-        s += ' weight:               '+str(self.jobWeight)+'\n'
-        s += ' runPath:              '+str(self.runPath)+'\n'
-        s += ' keepFilePatterns:     '+formatCollection(self.keepFilePatterns)+'\n'
-        s += ' auxFilePatterns:      '+formatCollection(self.auxFilePatterns)+'\n'
-        s += ' keepFiles:            '+formatCollection(self.keepFiles)+'\n'
-        s += ' auxFiles:             '+formatCollection(self.auxFiles)+'\n'
-        s += ' actions:              '+formatCollection(self.actions)+'\n'
-        s += ' tests:                '+formatCollection(self.tests)+'\n'
-        s += ' scripts:              '+str(self.scripts.keys())+'\n'
-        s += ' descriptor:\n'
-
-        return s
-
-    # --------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|            Minder  dump                   |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('\n'+self.__str__())
-        
-    def status(self):
-        return self.stateEngine.state.state
-
-    def printMethodTimes(self):
-        
-        m = '------ Minder  %s ---------\n%s' % (
-            self.identifiedName,
-            self.formatMethodTimes())
-
-
-        #m += '\n-----State engine ---------  \n%s' % (
-        #    self.stateEngine.formatMethodTimes())
-
-        m += '\n-----XML Converter ---------  \n%s' % (
-            self.xmlConverter.formatMethodTimes())
-        
-        self.logger.info(m)
-
-# --------------------------------------------------------------------------
-            
-    def doPostProcessing(self):
-        if self.processingResult == 'error': return False
-        if 'postProcessorScript' in self.scripts.keys(): return True
-        return False
-            
-
-    def setDone(self):
-        self.cleanSpace()
-        MinderBase.setDone(self)
diff --git a/Tools/RunTimeTester/src/Minder.py.obs b/Tools/RunTimeTester/src/Minder.py.obs
deleted file mode 100755
index c01fec2d5218087268b299f918cde7ec2958e4d6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Minder.py.obs
+++ /dev/null
@@ -1,1350 +0,0 @@
-"""
-class Minder. Top of the Minder Heirarchy.
-
-- Takes care of the state machine. This is a mixture of
-a status object and a dispatcher.
-
-The state machine  has states: queued, running, success, error,
-it nows how to proceed from one state to the next, and species the actions to be
-performed when the state changes. The way jobs are submited, and the names of
-log files are examples of items that change with mode.
-
-This class also has the following methods, which are common to all Minders,
-
-poll
-fromQueued
-fromRunning
-fromSuccess
-fromError
-setStatus
-setStatusList
-forceCloseDown
-setDone
-
-Many of these are overidden in derived classes.
-"""
-
-from Status                  import Status
-from JobGroupKits            import JobGroupKits
-from os.path                 import join
-from UserTestDescriptor      import UserTestDescriptor
-from formatCollection        import formatCollection
-from getInheritanceHeirarchy import isAthenaJob
-from Tools                   import sendMail, sortItemsInList, changePerm
-from Defs                    import RTTdefs
-from ModuleLoader            import ModuleLoader
-from exc2string2             import exc2string2
-from RTTTestRunner           import RTTTestRunner  
-from popen2                  import Popen4 
-from RTTCloseDownKeyMaker    import RTTCloseDownKeyMaker, dictToDocument
-from makeRTTDOM              import makeRTTDOM
-from MinderToXML             import MinderToXML
-from Factory_Commander       import CommanderFactory
-
-import os, shutil, shelve, fnmatch, imp, shelve, types, time
-
-import mutex
-dbmMutex     = mutex.mutex()
-# -------------------------------------------------------------------------
-#import logging
-#logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-
-#------------------------------------------------------------------------
-# Wrappers for Descriptor (and subclass) methods to provide a minder 'context'
-# -------------------------------------------------------------------------
-
-def minderContext(method):
-    """
-    Returns a method that wraps a the method passed in the argument.
-    The wrapper provides the context.
-    """
-
-    def contexed_method(self, *a, **k):
-
-        runPath = self.jDescriptor.runPath
-        if os.path.exists(runPath): os.chdir(runPath)
-
-        result = method(self, *a, **k)
-        return result
-
-    return contexed_method
-
-class MetaContexter(type):
-    """
-    A meta class that loops through all methods of a class, excudes __init__
-    and wraps the rest with minderContext
-    """
-
-    def __new__(self, classname, bases, classdict):
-
-        for mn in classdict.keys():
-            if mn == '__init__': continue
-            m = classdict[mn]
-            if isinstance(m, types.FunctionType):
-                classdict[mn] = minderContext(m)
-
-        return type.__new__(self, classname, bases, classdict)
-
-class Contexter:
-    "A class to allow method wrapping through inheritance"
-
-    __metaclass__ = MetaContexter
-            
-
-class MinderStateEngine(Contexter):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    
-    def __init__(self, name, logger):
-        self.dispatcher = {
-            "queued"              : self.fromQueued,
-            "running"             : self.fromRunning,
-            "success"             : self.fromSuccess,
-            "error"               : self.fromError,
-            "operatorCloseDown"   : self.fromOperatorCloseDown,
-            "timedOut"            : self.fromTimedOut
-            }
-        self.done       = False # ALWAYS set this through self.setDone()
-        self.name       = str(name)
-        self.status     = Status("queued")
-        self.logger     = logger
-        self.jobTimer   = None #to be set as needed. Set here for hasTimedOut()
-        
-    def setLogger(self, newLogger): self.logger = newLogger
-    
-    def poll(self, forceState=None):
-        """
-        makes status transitions by calling the fromXXX
-        methods, according to the current status.
-        returns True if job is active, False otherwise
-        """
-
-        # >= 1 checks are run if the job finishes.
-        # this returned in a list of statuses.
-        # the first one determines if the job ran
-        #
-
-        statOnEntry = self.status.primary()
-        if forceState:
-            stat = forceState
-        else:
-            stat  = self.status.primary()
-
-        msg = "Job: %s status: %s done: %s" % (self.name, stat, str(self.done))
-        self.logger.debug(msg)
-
-        if self.done: return 
-
-        try: # take care of  exceptions
-            stat = self.status.primary()
-            self.dispatcher[stat]()
-        except:
-            
-            m  = 'Exception thrown during Minder::poll()'
-            m += 'JobName: %s\n' %  str(self.jDescriptor.identifiedName)
-            m += 'Traceback:\n'
-            m +=  exc2string2()
-            self.logger.error(m)
-            self.forceCloseDown()
-            
-            return 
-
-        if stat != statOnEntry:
-
-            # make an DOM document for the minder and call back
-            # (presumably toi the testRun)
-
-            m = 'State transition detected: %s -> %s, calling back' % (
-                stat, self.status.primary())
-            self.logger.info(m)
-            startT = time.time()
-            self.xmlConverter.update()
-            diffT = time.time() - startT
-            self.sumTimeToUpdate += diffTime
-            self.nUpdates      += 1
-            m = 'Job %s completed the call back in %d seconds.' % (
-                self.jDescriptor.identifiedName,diffT)
-            self.logger.info(m)
-            # recursive call  move the job through states that take no time
-            # to complete
-            self.poll() 
-        return 
-
-    # -----------------------------------------------
-    
-    #def monitoringCallBack(self, importance):
-    #    # issue message if status has changed
-    #    self.logger.debug("Job: %s status: %s done: %s" % (
-    #        self.name,
-    #        self.status.primary(),
-    #        str(self.done))
-    #                      )
-    #    if not self.runSummaryCallBackHandler:            
-    #        m = 'Job %s: no monitoring call back handler set. Returning.' % (
-    #            self.jDescriptor.identifiedName)
-    #        self.logger.debug(m)
-    #        return
-
-    #    # allow one minder to update at a time  to stop threaded minders
-    #    # to intere with each other.
-    #    startT = time.time()#
-
-    #    try:            
-    #        acquiredMoniLock = \
-    #                         self.runSummaryCallBackHandler.update(importance)
-    #    except:
-    #        m  = """Error during monitoring callback from job: %s'
-    #        Traceback:\n%s """ % (self.jDescriptor.identifiedName,
-    #                              exc2string2())
-    #        self.logger.error(m)
-    #    else:
-    #        if acquiredMoniLock:
-    #            diffT = time.time() - startT
-    #            m = 'Job %s completed the call back in %d seconds.' % (
-    #                self.jDescriptor.identifiedName,diffT)
-    #        else:
-    #            m = 'Job %s, unable to acquire monitoring lock, continuing' % (
-    #                self.jDescriptor.identifiedName)#
-
-    #        self.logger.info(m)
-    # -----------------------------------------------
-        
-    def forceCloseDown(self):
-        if not self.done:
-            self.setStatus('error')
-            self.setDone()
-    
-    # -----------------------------------------------
-        
-    
-
-    # -----------------------------------------------
-    def stopJob(self): pass
-    
-    # =======================================================
-    #
-    # State transition calls. probably no need to override these
-    #
-    # =======================================================
-        
-    def fromQueued(self):
-        if self.fromQueuedTest(): self.fromQueuedAction()
-
-    def fromRunning(self):
-        if self.fromRunningTest(): self.fromRunningAction()
-        
-    def fromSuccess(self):
-        if self.fromSuccessTest(): self.fromSuccessAction()
-        
-    def fromError(self):
-        if self.fromErrorTest(): self.fromErrorAction()
-        
-
-    def fromOperatorCloseDown(self):
-        if self.fromOperatorCloseDownTest(): self.fromOperatorCloseDownAction()
-
-    def fromTimedOut(self):
-        if self.fromTimedOutTest():
-            self.fromTimedOutAction()
-
-
-    # =======================================================
-    #
-    # State transition calls. To be overridden by subclasses
-    #
-    # =======================================================
-        
-    def fromQueuedTest(self):
-        return True
-        
-    def fromQueuedAction(self):
-        self.setStatus("running")
-
-
-    #-----------------------------------------------
-
-    def fromRunningTest(self):
-        return True
-        
-    def fromRunningAction(self):
-        self.setStatus("success")
-
-    #-----------------------------------------------
-
-    def fromSuccessTest(self):
-        return True
-        
-    def fromSuccessAction(self):
-        self.setDone
-
-    #-----------------------------------------------
-
-    def fromErrorTest(self):
-        return True
-        
-    def fromErrorAction(self):
-        self.setDone()
-
-    #-----------------------------------------------
-    
-    def fromOperatorCloseDownTest(self):
-        return True
-        
-    def fromOperatorCloseDownAction(self):
-        self.stopJob()
-        self.setDone()
-
-
-    #-----------------------------------------------
-    
-    def fromTimedOutTest(self):
-        return True
-        
-    def fromTimedOutAction(self):
-        self.stopJob()
-        self.copyKeepFiles()
-        self.storeAllFiles()
-        self.setDone()
-        
-    #-----------------------------------------------
-    
-    def setDone(self):
-        "Sets the done flag. Sub classes may extend"
-        self.done = True
-        # importance = 'high'
-        # self.monitoringCallBack(importance)
-        self.xmlConverter.update()
-
-    #-----------------------------------------------
-    
-class Minder(MinderStateEngine):
-    def __init__(self, jDescriptor, logger, runTimer):
-        """ One function of the minder is to bring together information
-        from the job descriptor (gathered from the jobs config file),
-        the job groups file (gathered from the job groups config file),
-        and the Paths object (gathered from the RTT config file)
-
-        For files to be transferred from the run directory to the results
-        area, we use patterns (for file name matching). Lists of patterns
-        are made at in this method, with a list per use (asciiRef files,
-        binRef files, keep files. These lists will be used to identify
-        filenames at the appropriate moment, and converted to absolute
-        path names. Source and destination will be kept in the dictionaries
-        self.keepFiles self.binRefFiles etc.
-        
-        """
-
-        self.logger = logger
-        self.logger.debug('init')
-        MinderStateEngine.__init__(self, jDescriptor.identifiedName, logger)
-        
-        self.jDescriptor         = jDescriptor
-        self.jDescriptor.logger  = self.logger
-
-        self.removeMe            = False
-        self.actions             = []
-        self.monActions          = []
-        self.tests               = []
-        self.checks              = []
-        self.asciiRefFiles       = []
-        self.binRefFiles         = []
-        self.runPath             = jDescriptor.runPath
-        self.testDescriptors     = []
-        self.keepFilePatterns    = []
-        self.testDBPath          = os.path.join(self.runPath,'RTTtests.db')
-        self.testIds             = []
-        self.runTimer            = runTimer
-        self.closeDownKeyMaker   = RTTCloseDownKeyMaker(
-            self.jDescriptor.paths,
-            self.runTimer.startTime,
-            self.logger)
-        self.closeDownKeyDict = {
-            'branch':           self.jDescriptor.paths.originalBranch,
-            'build':            self.jDescriptor.paths.build,
-            'platform':         self.jDescriptor.paths.platform,
-            'package':          self.jDescriptor.paths.package,
-            'release':          self.jDescriptor.paths.release,
-            'runType':          self.jDescriptor.paths.runType,
-            'startTime':        self.runTimer.startTime,
-            'jobName':          self.jDescriptor.identifiedName,
-            'rttPilotJob':      self.jDescriptor.rttPilotJob,         
-            'atnPilotJob':      self.jDescriptor.rttATNJob,
-            'package':          self.jDescriptor.paths.package
-            }
-        if self.jDescriptor.paths.isNightly:
-            self.closeDownKeyDict['release'] = 'nightly'
-            
-
-        # For conveniencde, a copy of the key to be used to close down this
-        # job will be output to the run directory. To make the xml file
-        # start by creating a document. The output is done elsewhere.
-        self.closeDownDom = dictToDocument(self.closeDownKeyDict,
-                                           makeRTTDOM('RTTCloseDownKey'))
-        
-            
-        
-
-
-        self.jobGroupKit = None
-        try:
-            self.jobGroupKit    = JobGroupKits().kit(jDescriptor.jobGroup)
-        except:
-            pass
-
-        if not self.jobGroupKit:
-            mode = 'Athena'+jDescriptor.jobGroup
-            try:
-                self.jobGroupKit = JobGroupKits().kit(mode)
-            except:
-                pass
-
-        if not self.jobGroupKit:
-            m  = 'Fatal error: No jobgroup kit for group  '
-            m += jDescriptor.jobGroup
-            self.logger.critical(m)
-            self.forceCloseDown()
-
-        # These jobs are NOT counted in  Launcher stats. Otherwise overide.
-        self.jobWeight      = 0 
-
-        # bring together information on files to manipulate.
-        # Keep and aux files are handled as patterns - the actual file
-        # name resolution is deferred to later
-        # Reference files references are complete file names due to
-        # fact that they will go into keys before retirval - and patterns wont work
-        
-        # so complete file names go here here as well.
-        self.handleFilePatterns()
-        self.handleFileNames()
-        self.handleExternalModule()
-
-        self.makeJobGroupDirs()
-        self.shelveTests()
-
-        self.testStatusList    =[] #results returned by test tags
-        
-
-        # declare dictionaries with source and destination paths
-        # for files to be transferred from the run directories to the resulsts
-        # directory.
-
-        self.keepFiles      = {}
-
-        # declare lists of files - no  wild cards, as the elements of the list
-        # are used in data base keys. Also,  no absolute paths for the same reason
-        # (would break portability)
-        
-        self.minderpsDict   = {'log' : self.jDescriptor.log,
-                               'site': self.jDescriptor.paths.site
-                               }
-        self.refDBenable                 = jDescriptor.paths.refDBenable
-
-        self.monActionsData              = [] # monitoring information
-        self.runSummaryCallBackHandler   = None
-        
-        # maximum no of times to retry a job if it enters the error state.
-        self.errorStateMaxRetry = 3
-        self.errorStateCurRetry = 0
-          
-
-        self.exitStatus = 'Unknown'
-        self.cpuTime    = 'Unknown'
-        self.mem        = 'Unknown'
-        self.vmem       = 'Unknown'
-        self.wallTime   = 'Unknown'
-
-        self.constructionErrorMessages = jDescriptor.errorMessages
-        
-        self.xmlConverter = MinderToXML(self)
-        self.summaryCallBack = None
-
-        self.nUpdates            = 0
-        self.sumTimeToUpdate     = 0
-        
-    def setSummaryCallBack(self, newCallBack):
-        self.xmlConverter.setSummaryCallBack(newCallBack)
-    
-# ------------------------------------------------------------------------
-#    def setLogger(self, newLogger):
-#        self.logger = newLogger
-#        # to do: set the descriptor logger as well
-# ------------------------------------------------------------------------
-
-    def poll(self):
-        """
-        makes status transitions by calling the fromXXX
-        methods, according to the current status.
-        returns True if job is active, False otherwise
-        """
-
-        # >= 1 checks are run if the job finishes.
-        # this returned in a list of statuses.
-        # the first one determines if the job ran
-        #
-
-        stat = self.status.primary()
-        msg = "TestRun:poll(): Job: %s status: %s done: %s" % (self.name,
-                                                               stat,
-                                                               str(self.done))
-        self.logger.debug(msg)
-
-        
-        #
-        # perform some checks before going on to checking if there should be a
-        # change of state.
-        # - time-outs
-        forceState = None
-        if self.hasTimedOut():
-            forceState = "timedOut"
-            maxTime = str(self.jobTimer.maxTimeInHours())
-            
-            m = """Job %s run has exceeded the max run time of %s hours,
-            moving state to 'timedOut' """ % (self.name, maxTime)
-
-            self.logger.info(m)
-
-        
-        # - external closeDownFile
-        elif self.closeDownKeyMaker.laterMatch(self.closeDownKeyDict):
-            
-            m = 'Job %s moving state to operatorshutdown' % (self.name)
-            self.logger.info(m)
-
-            forceState = 'operatorCloseDown'
-
-
-        # state engine stuff done in the base class.
-        return MinderStateEngine.poll(self)
-
-# ------------------------------------------------------------------------
-
-    def shelveTests(self):
-
-        # associate string identifiers with each test and action
-        npos = 0
-        for t in self.tests:
-            self.testIds.append( (t, '%s_%d' % (t.__class__.__name__,npos)) )
-            npos +=1
-        for t in self.actions:
-            self.testIds.append( (t, '%s_%d' % (t.__class__.__name__,npos)) )
-            npos +=1
-
-        # provide a database to store the tests
-        db = shelve.open(self.testDBPath, 'n')
-        db.close()
-
-        # wrap the tests to allow transportation to the computing node
-        wrappers = [RTTTestRunner(self.testDBPath, t) for t in self.testIds]
-        
-        # write the tests to a shelf db
-        [t.autoShelve() for t in wrappers]
-        
-    def hasTimedOut(self):
-        'Returns true if job has timed out, false otherwise.'        
-        # if self.status.primary() == 'timedOut': return True
-        if self.status.primary() != 'running': return False
-
-        if self.jobTimer:
-            if self.jobTimer.timeOut():
-                return True
-        return False
-
-    # -------------------------------------------------------------------------
-
-    # def setRunSummaryCallBackHandler(self, handler):
-    # self.runSummaryCallBackHandler = handler
-
-    # -------------------------------------------------------------------------
-
-    def hasChecks(self):
-        answer = False
-
-        if hasattr(self,'checks'):
-            answer = len(self.checks) > 0 or answer
-
-        if hasattr(self,'tests'):  
-            answer = len(self.tests) > 0 or answer
-
-        return answer
-
-    # ------------------------------------------------------------------------
-
-    def handleFilePatterns (self):
-        "method to collect file patterns from the various sources"
-        
-        self.keepFilePatterns  = self.jobGroupKit.getItem('keepFilePatterns')
-        self.auxFilePatterns   = self.jobGroupKit.getItem('auxFilePatterns')
-        self.storeInDBPatterns = self.jobGroupKit.getItem('storeInDBPatterns')
-       
-        self.keepFilePatterns.extend(self.jDescriptor.keepFilePatterns)
-        self.auxFilePatterns.extend(self.jDescriptor.auxFilePatterns)
-        self.auxFilePatterns.extend(self.jDescriptor.storeInDBPatterns)
-
-# ------------------------------------------------------------------------
-
-    def handleExternalModule(self):
-        
-        """test Descriptor is a list with three elements
-           modulename, classname, dictionary of arguments
-        """
-          
-        actionDescriptors        = []
-        alist = self.jobGroupKit.getItem('actions')
-        if alist:    # might = None
-            actionDescriptors.extend(alist)
-
-        actionDescriptors.extend(self.jDescriptor.actions)
-        
-
-        tlist = self.jobGroupKit.getItem('tests')
-        if tlist:
-            self.testDescriptors.extend(tlist) # testList might = None
-        self.testDescriptors.extend(self.jDescriptor.tests)
-
-        monActionDescriptors   =self.jobGroupKit.getItem('monActions')
-        #self.actionDescriptors.extend(self.jDescriptor.actions)
-
-        msg = 'handleExternalModule -actions %d' % len(actionDescriptors)
-        self.logger.debug(msg)
-        
-        
-        for descriptor in self.testDescriptors:
-
-            descriptor.addParameter(('JobDescriptor', self.jDescriptor))
-            descriptor.addParameter(('logger', self.logger))
-            instance = self.getClassFromDescriptor(descriptor) 
-            if instance != None: self.tests.append(instance)
-        
-
-        for descriptor in actionDescriptors:
-
-            descriptor.addParameter(('JobDescriptor', self.jDescriptor))
-            descriptor.addParameter(('logger', self.logger))
-            instance = self.getClassFromDescriptor(descriptor) 
-            if instance != None: self.actions.append(instance)
-            
-
-        for descriptor in monActionDescriptors:
-
-            descriptor.addParameter(('JobDescriptor', self.jDescriptor))
-            descriptor.addParameter(('logger', self.logger))
-            instance = self.getClassFromDescriptor(descriptor) 
-            if instance != None: self.monActions.append(instance)
-          
-#-----------------------------------------------------------------------
-
-    def getClassFromDescriptor(self, testDescriptor):
-
-
-        # allow RTT library tools to be picked up from the local code base
-        # rather than the installed shared area if so requested in the
-        # RTT configuration file
-        rttLibDir = str(self.jDescriptor.paths.rttLibDir)
-        sharePathToUse = {
-            'RuleCheckerLibraryTools': rttLibDir,
-            'RttLibraryTools'        : rttLibDir
-            }
-
-        instance = None
-
-        # RttLibTools has been split into modules with one class per
-        # module, with the module name = he class name. Needed this
-        # for shelve to unshelve instances.
-        # moduleName = testDescriptor.moduleName
-        moduleName = testDescriptor.testName
-        className  = testDescriptor.testName
-        paramDict  = testDescriptor.argDict
-        
-        # default is release share path if key not in dict
-        # this happens when the user supplies own module, which is
-        # picked up from the release area.
-
-
-        # For the moment (25/9/06) only allow the module to by in rttLibDir
-        # sharePathIs = sharePathToUse.get(moduleName,
-        #                                  self.jDescriptor.paths.shareArea)
-
-        sharePathIs = rttLibDir
-        self.logger.debug('module path ' + sharePathIs)
-        self.logger.debug('moduleName  ' + moduleName)
-        self.logger.debug('className   ' + className)
-        self.logger.debug('paramDict   ' + str(paramDict))
-        self.logger.debug('logger      ' + str(self.logger))
-
-        try:            
-            mL        = ModuleLoader(moduleName, sharePathIs, self.logger)
-            instance  = mL.getClassInstance(className, paramDict)
-            self.logger.debug('Created a test instance of class %s' % (
-                instance.__class__.__name__))
-        except Exception, e:
-            self.logger.error('Failed to create a test instance')
-            self.logger.error(exc2string2())
-            self.logger.error(str(e))
-            
-        return instance
-    
-#------------------------------------------------------------------------
-
-    def handleFileNames(self):
-        """ method to collect file names, and resolve patterns where
-        for those patterns for which it is possible at init."""
-        self.asciiRefFiles.extend(self.jobGroupKit.getItem('asciiRefFiles'))
-        self.asciiRefFiles.extend(self.jDescriptor.asciiRefFiles)
-
-        self.binRefFiles.extend(self.jobGroupKit.getItem('binRefFiles'))
-        self.binRefFiles.extend(self.jDescriptor.binRefFiles)
-        
-        # resolve auxFile patterns to file names
-        auxFiles = []
-
-        auxFileDirectory = self.jDescriptor.paths.shareArea
-        if not os.path.exists(auxFileDirectory):
-            self.logger.fatal('unknown aux file directory: '+auxFileDirectory)
-        else:
-            filesInShare = os.listdir(auxFileDirectory)
-            matches = []
-            for pattern in self.auxFilePatterns:
-                matches.extend([file for file in filesInShare if fnmatch.fnmatch(file, pattern)])
-                # convert to full path names
-            self.auxFiles = [join(auxFileDirectory, file) for file in matches]
-
-            # add the test configuration xml file as an aux file
-            # self.auxFiles.append(join(auxFileDirectory,self.jDescriptor.paths.package+'_TestConfiguration.xml'))
-            self.auxFiles.append(self.jDescriptor.paths.confFile)
-
-#------------------------------------------------------------------------
-
-    def hasRefFiles(self):
-        return (len(self.asciiRefFiles)>0 or len(self.binRefFiles)>0)
-
-    # =======================================================
-    #
-    # State transition calls.Overrides to the MinderStateEngine methods
-    #
-    # =======================================================
-    def fromRunningAction(self):
-        self.logger.debug('************************ fromRunningAction ****')
-        self.storeAllFiles()
-        self.setStatus("success")
-
-    # =======================================================
-
-    def fromSuccess(self):
-        "Status transition method. To be overridden."
-        if self.fromSuccessTest(): self.fromSuccessAction()
-        
-    def fromSuccessTest(self):
-        return True
-        
-    def fromSuccessAction(self):
-        msg = 'Job: ' + str(self.jDescriptor.identifiedName) 
-        self.logger.debug(msg + ' in Minder::fromSuccessAction')
-        self.storeAllFiles()
-        self.logger.debug(msg + ' after storeAllFiles')
-        self.setDone()
-        self.logger.debug(msg + ' about to leave')
-    
-    # =======================================================
-    
-    def fromError(self):
-        "Status transition method. To be overridden."
-        if self.fromErrorTest(): self.fromErrorAction()
-        
-    def fromErrorTest(self):
-        return True
-        
-    def fromErrorAction(self):
-        self.storeAllFiles()
-        self.setDone()
- 
-    # =======================================================
-    # Do not set the status member data
-    # =======================================================
-
-    def statusToRunDir(self, statusList):
-        # write the status to a file eg for use by
-        # TransparentDummyMinders
-        statFile=join(self.jDescriptor.runPath, 'status.txt')
-        sf = open(statFile,'w')
-        sf.write(statusList[0] + ' ' + statusList[1])
-        sf.close()
-        msg = 'wrote statuses: %s and %s to %s' %(statusList[0],
-                                                  statusList[1],
-                                                  statFile)
-        self.logger.debug(msg)
-
-    #-----------------------------------------------
-    
-    def setStatus(self, status):
-        if not self.done:
-            self.status.set(status)
-            # self.statusToRunDir([self.status.primary(),self.status.secondary()])
-
-    #-----------------------------------------------
-    
-    def setStatusList(self, statusL):
-        if not self.done:
-            idName = str(self.jDescriptor.identifiedName)
-            self.logger.debug('setStatusList for Job: %s' % idName)
-            self.status.setList(statusL)
-            self.logger.debug('status to run dir for Job: %s' % idName)
-            # self.statusToRunDir(statusL)
- 
-    #-----------------------------------------------
-    
-    def forceCloseDown(self):
-
-        # make sure ALL files in run and results directory are
-        # accessible by all members of the UCL Atlas group
-        MinderStateEngine.forceCloseDown(self)
-        
-   #-----------------------------------------------
-
-    def reportUserScriptError(self,message):
-        filePath  = os.path.join(self.runPath,'Python_Script_Output.log')
-        h = open(filePath,'a+')
-        h.write(message)
-        h.close()
-
-    #-----------------------------------------------
-    
-    def runTests(self):
-        jobName = self.jDescriptor.identifiedName
-        nTests  = len(self.tests)
-        msg     = 'Running %d user defined tests for job %s' % (nTests, jobName)
-        self.logger.debug(msg)
-                    
-        for test in self.tests:
-            
-            className = str(test.__class__.__name__)
-            self.logger.debug("Running test " + className)
-            try:
-                rc     = test.run()
-                status = RTTdefs.status(int(rc))
-                self.testStatusList.append(str(status))
-            except Exception, e:
-                self.logger.error("Could not run user test " + className)
-                self.logger.error(str(e))
-                self.logger.error(exc2string2())
-                msg  = '-----------------------------------------\n'
-                msg += 'Test: ' + className + ' could not be run!\n'
-                msg += exc2string2() + '\n'
-                msg += '-----------------------------------------\n\n'
-                self.reportUserScriptError(msg)
-                
-            self.logger.debug("Running next available user test")
-
-        self.logger.debug("Finished running user tests for %s" % jobName)
-        
-
-    #-----------------------------------------------
-    
-    def runTests(self):
-
-        db = shelve.open(self.testDBPath)
-        for t in db.values():
-            if t.error:
-                self.logger.error('Error running test %s' % t.testId)
-            else:
-                self.logger.debug('No Error running test %s' % t.testId)
-
-
-            self.testStatusList.append(RTTdefs.status(t.result))
-
-
-            # collect any messages generated by the test
-            t.transferLog(self.logger) 
-
-            
-       
-    #--------------------------------------------------
-
-    def runActions(self):
-        """Running Action"""
-        return
-        os.chdir(self.jDescriptor.runPath)
-        self.logger.debug("runDir before actions :"+str(os.listdir(self.jDescriptor.runPath)))
-        jobName  = self.jDescriptor.identifiedName
-        nActions = len(self.actions)
-        msg      = 'Running %d actions for job %s' % (nActions, jobName)
-        self.logger.debug(msg)
-
-        for action in self.actions:
-            className = str(action.__class__.__name__)
-            try:
-                self.logger.debug("Running action " + className)
-                action.run()
-                
-            except Exception, e:
-                self.logger.error("Could not run action " + className)
-                self.logger.error(str(e))
-                self.logger.error(exc2string2())
-
-                msg  = '-----------------------------------------\n'
-                msg += 'Action: ' + className + ' could not be run!\n'
-                msg += exc2string2() + '\n'
-                msg += '-----------------------------------------\n\n'
-                self.reportUserScriptError(msg)
-                                                                                                
-            self.logger.debug("Running next available action")
-
-        self.logger.debug("Finished running actions for %s" % jobName)
-
-    #--------------------------------------------------
-    
-    def runMoniActions(self):
-        jobName  = self.jDescriptor.identifiedName
-        nActions = len(self.monActions)
-        msg      = 'Running %d monActions for job %s' % (nActions, jobName)
-        self.logger.debug(msg)
-
-        for action in self.monActions:
-            className = str(action.__class__.__name__)
-            try:
-                self.logger.debug("Running monAction " + className)
-                dataDict=action.run()
-                self.monActionsData.append(dataDict)
-                
-            except Exception, e:
-                msg = "Could not run monitoring action " + className
-                self.logger.error(msg)
-                self.logger.error(str(e))
-                self.logger.error(exc2string2())
-                msg  = '-----------------------------------------\n'
-                msg += 'MoniAction: ' + className + ' could not be run!\n'
-                msg += exc2string2() + '\n'
-                msg += '-----------------------------------------\n\n'
-                self.reportUserScriptError(msg)
-                                
-
-            self.logger.debug("Running next available monAction")
-             
-                
-        self.logger.debug('Minder moniData :'+str(self.monActionsData))
-        self.logger.debug("Finished running monActions for %s" % jobName)
-
-    #--------------------------------------------------
-
-    def fullResultsFileName(self, filename):
-        return join(self.jDescriptor.resPath, filename)
-
-    #-----------------------------------------------
-    
-    def makeKeepFileEntry(self, file, infoString):
-        "helper method for registerWildKeepFiles()"
-        
-        src = join(self.runPath, file)
-        # idPreFix = self.jDescriptor.identifiedName+"_"
-        dest = {'keepFileString': self.fullResultsFileName(file),
-                'infoString':     infoString}
-        
-        self.keepFiles[src]=dest
-        
-    #-----------------------------------------------
-    
-    def registerWildKeepFiles(self):
-        """
-        Common implementation task.
-
-        Obtain the wild card patterns for the current job group from
-        a JobGroupKit.
-        
-        Collect all the wild carded keep files.
-        These are files that are present in the run directory at when
-        fromRunning is called, and which match a pattern in the pattern
-        list.
-
-        Give these files their full paths and add them to the keepFile
-        dictionary.
-        """
-
-        for card in self.keepFilePatterns:
-
-            keepString = card['keepFileString']
-            infoString = card['infoString']
-            
-            wildFiles = fnmatch.filter(os.listdir(self.runPath), keepString)
-
-            [self.makeKeepFileEntry(file, infoString) for file in wildFiles]
-            
-
-        # now manually add the package configuration file to keep files
-        self.makeKeepFileEntry(
-            os.path.basename(self.jDescriptor.paths.confFile),
-            "Package XML test configuration file")
-
-        self.keepFiles.update(self.jDescriptor.keepFiles())
-
-        #    srcFiles = [join(self.runPath, file) for file in wildFiles]
-        #    desFiles = [self.fullResultsFileName(idPreFix+file)
-        #                for file in wildFiles]
-
-        #   for s, d in zip(srcFiles, desFiles):
-        #        self.keepFiles[s]=d
-
-    #-----------------------------------------------
-
-    def makeDBKey(self, release, theBranch, fileName, fileType):
-        """Ask the descriptors for the key - as different descriptors
-        do this differently.
-        fileType: can be 'bin' or 'ascii'
-        """
-        # key = [unicode(self.jDescriptor.paths.release),
-        #        unicode(self.jDescriptor.paths.build),
-        #       unicode(self.jDescriptor.name),
-        #       dataFiles2String(self.jDescriptor),
-        #       unicode(file)]
-
-        key = self.jDescriptor.makeDBKey()
-        key.append(unicode(fileName))
-        key.append(unicode(release))
-        key.append(unicode(theBranch))
-        key.append(unicode(fileType))
-        key = sortItemsInList(key)
-        
-        return key
-
-    #-----------------------------------------------
-    
-    def storeAllFiles(self):
-        """
-        find  all files specified by the xml tag 'storeInDatabasePattern'
-        Then call a function to obtain reference file names ('asciiRefFile'
-        and 'binRefFile' tags), then store the files in the db.
-        """
-
-        self.logger.debug('Store: ascii: %s' % str(self.asciiRefFiles))
-        self.logger.debug('Store: bin:   %s' % str(self.binRefFiles))
-        if not self.refDBenable:
-            self.logger.debug("DBM writing disabled")
-            return
-            
-        msg  = "DBM writing enabled for job: "
-        msg += str(self.jDescriptor.identifiedName)
-        self.logger.debug(msg)
-
-        filesToStore = {'txt': [],
-                        'bin': []
-                        }
-
-        for pattern in self.storeInDBPatterns:
-            filesToStore['txt'].extend(fnmatch.filter(os.listdir(self.runPath),
-                                                      pattern))
-
-        
-        # if isAthenaJob(self.jDescriptor): # why only Athena jobs ???
-        filesToStore['txt'].extend(self.asciiRefFiles)
-        filesToStore['bin'].extend(self.binRefFiles)
-
-        toStore = []
-        isBin = False
-        for fn in filesToStore['txt']:
-            #            key = self.makeDBKey(self.jDescriptor.paths.release, fn)
-            key = self.makeDBKey(self.jDescriptor.paths.release,
-                                 self.jDescriptor.paths.branch,
-                                 fn,
-                                 'ascii')            
-            try:
-                toStore.append(StoreItem(self.runPath, fn, key, isBin))
-                msg = 'Stored item in db using key %s' % key
-                logger.debug(msg)
-            except Exception, e:
-                msg = 'Error while storing ascii file %s, skipping file' % fn
-                self.logger.warning(msg)
-        isBin = True
-        for fn in filesToStore['bin']:
-            #            key = self.makeDBKey(self.jDescriptor.paths.release, fn)
-            key = self.makeDBKey(self.jDescriptor.paths.release,
-                                 self.jDescriptor.paths.branch,
-                                 fn,
-                                 'bin')
-            try:
-                toStore.append(StoreItem(self.runPath, fn, key, isBin))
-            except Exception:
-                msg = 'Error while storing binary file %s, skipping file' % fn
-                self.logger.warning(msg)
-
-
-        [self.logger.debug('storing in DBM:\n %s' % ts.__str__()) for ts in toStore]
-
-        msg = "about to write to the dbm: %s" % self.jDescriptor.paths.refDB
-        self.logger.debug(msg)
-
-        self.grabDBresourceAndStore(toStore)
-
-    # --------------------------------------------------------------------
-    
-    def grabDBresourceAndStore(self,storeItems):
-        # use the global dbm mutual excluder to prevent concurrent dbm access
-        try:
-            dbmMutex.lock(self.placeInDBM, storeItems)
-        except Exception, e:
-            idName = str(self.jDescriptor.identifiedName)
-            self.logger.error('Problem trying to store to DB for job: '+idName)
-            self.logger.error('Job: %s\n%s' %(idName,
-                                              exc2string2()
-                                              )
-                              )
-            self.logger.error(str(e))
-
-        # must now give back the resource
-        dbmMutex.unlock()
-
-    # --------------------------------------------------------------------
-
-    def placeInDBM(self, storeItems):
-
-
-
-        dbHandle = shelve.open(self.jDescriptor.paths.refDB, 'c')
-        for item in storeItems:
-            self.logger.debug('Storing %s length %d' % (item.fullPath, len(item.strFile)))
-            dbHandle[item.key] = item.strFile
-        dbHandle.close()                
-
-        # change mod on db if owner of db = script runner
-        changePerm(self.jDescriptor.paths.refDB,02775)
-
-        msg = 'Stored all items in the DB for job: ' + str(self.jDescriptor.identifiedName) + '...about to leave Minder::placeInDBM method'
-        self.logger.debug(msg)    
-
-    # --------------------------------------------------------------------
-                            
-    # copy files to be kept (log histos, ntuples...) to results dir
-    def copyKeepFiles(self):
-        
-        # find if any of the wild keep cards have a match
-        idName = str(self.jDescriptor.identifiedName)
-        try:
-            self.registerWildKeepFiles()
-        except Exception, e:
-            msg = "Exception registering keep files for job: " + idName
-            self.logger.error(msg)
-            self.logger.error(str(e))
-            self.logger.error(exc2string2())
-            
-
-        for file in self.keepFiles.keys():
-            
-            if not os.path.exists(file):
-                msg = "File does not exist: %s for job %s" % (str(file),
-                                                              idName
-                                                              )
-                
-                
-                self.logger.error(msg)
-
-            if os.path.exists(file):
-
-                msg  = 'Setting up to copy keep files'
-                msg += str(self.jDescriptor.identifiedName)
-                self.logger.debug(msg)
-                
-                dictValue = self.keepFiles[file]
-                desFile = dictValue['keepFileString'] # this is a filename
-
-                try:
-                    self.logger.debug("copying file %s to %s" % (str(file), str(desFile)))
-                    shutil.copy( file, desFile)
-                    changePerm(desFile, 02775)
-                except:
-                    # disk space problems?
-                    message = '***** COPYING OF KEEP FILES PROBLEM! *****\n'
-                    message += 'Unable to copy src file:\n'
-                    message += '   ' + file + '\n'
-                    message += 'to destination file:\n'
-                    message += '   ' + desFile + '\n'
-                    message += 'Doing command df on ' + desFile + ' yields the answer:\n'
-                    #                    message += os.system('df ' + desFile)
-                    message += str(os.popen('df ' + desFile).read())
-                    message += '---------------------------------'                    
-                    self.logger.error(message)
-
-        self.logger.debug("Leaving copyKeepFiles for job: " + idName)
-
-    # --------------------------------------------------------------------
-        
-    def performChecks(self):
-        
-        statusL = []
-        theJob = str(self.jDescriptor.identifiedName)
-        msg = 'Len self.checks is: %s for %s ' % (str(len(self.checks)),theJob)
-        self.logger.debug(msg)
-        
-        for check in self.checks:
-
-            check.setLogger(self.logger) # log to the minders logfile
-
-            if not callable(check):
-                msg = "Uncallable checker: %s for job %s" % (str(check),
-                                                             theJob)
-                self.logger.error(msg)
-                statusL.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            self.logger.debug('Job: %s calling check: %s'%(theJob, str(check)))
-            status = ''
-            try:
-                rc = check(self) # int return code from check
-                status = RTTdefs.status(rc) # convert to string
-            except Exception, e:
-                msg = 'Exception raised while executing %s' % str(check)
-                self.logger.error(msg)
-                self.logger.error(exc2string2())
-                self.logger.error(str(e))
-                statusL.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            msg = 'Job: %s  check: %s return status %s' % (theJob,
-                                                           str(check),
-                                                           status)
-            self.logger.debug(msg)
-            
-            statusL.append(status)
-        
-        statusL.extend(self.testStatusList)
-        msg = "Checker status List "+str(statusL) + " for job: " + theJob
-        self.logger.debug(msg)
-        return statusL    
-
-    # --------------------------------------------------------------------
-    
-    def summaryCheckStatusList(self,statusList):
-        # Reduces the potentially > 2 status list to a 2-member list
-        # of "did job complete" and "logical AND of other checks.
-        
-        summaryStatusList = [statusList[0]]
-
-        if len(statusList) == 1: # only one check for this job
-            summaryStatusList.append('n/a')
-        else:
-            if statusList[0]=='error':
-                summaryStatusList.append('error')
-            elif 'error'  in statusList[1:]:
-                summaryStatusList.append('error')
-            else:
-                summaryStatusList.append('success')
-
-        
-        return summaryStatusList
-
-    # --------------------------------------------------------------------
-    
-    # this takes the reduced status list (reduced by 'summaryCheckStatusList') as parameter    
-    def handleCheckResults(self,statusList):
-        idName = str(self.jDescriptor.identifiedName)
-        if not len(statusList)==2:
-            msg =  'handleCheckResults function requires 2 members in check '
-            msg += 'status list!!!'
-            self.logger.error(msg)
-
-    # --------------------------------------------------------------------
-           
-    def isSuccess(self):
-        if self.done and self.status.primary() == 'success':
-            return True
-        return False
-
-    # --------------------------------------------------------------------
-    
-    def queue(self): return self.jDescriptor.batchQueue()
-
-    # --------------------------------------------------------------------
-
-    def makeJobGroupDirs(self):
-        try:
-            self.jobGroupDirectoryMaker.makeJobGroupDirs(self)
-        except:
-            msg  = """
-            %s caused an exception while making directories,
-            has been scheduled for removal\n traceback:\n%s""" % (
-            self.jDescriptor.identifiedName,
-            exc2string2())
-            print msg
-            self.logger.warning(msg)
-
-            # reraise the exception. This will be caught and an error Minder
-            # will be created
-            raise
-
-    def __str__(self):
-        s = '----------- Minder ---------------\n'
-        s += ' done:                 '+str(self.done)+'\n'
-        s += ' weight:               '+str(self.jobWeight)+'\n'
-        s += ' runPath:              '+str(self.runPath)+'\n'
-        s += ' keepFilePatterns:     '+formatCollection(self.keepFilePatterns)+'\n'
-        s += ' auxFilePatterns:      '+formatCollection(self.auxFilePatterns)+'\n'
-        s += ' keepFiles:            '+formatCollection(self.keepFiles)+'\n'
-        s += ' auxFiles:             '+formatCollection(self.auxFiles)+'\n'
-        s += ' asciiRefFiles:        '+formatCollection(self.asciiRefFiles)+'\n'
-        s += ' binRefFiles:          '+formatCollection(self.binRefFiles)+'\n'
-        s += ' actions:              '+formatCollection(self.actions)+'\n'
-        s += ' tests:                '+formatCollection(self.tests)+'\n'
-        s += ' descriptor:\n'
-        s +=  self.jDescriptor.__str__()
-
-        return s
-
-    # --------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|            Minder  dump                   |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('\n'+self.__str__())
-
-# =============================================================================
-# =============================================================================
-    
-class StoreItem:
-    def __init__(self, dir, fileName, key, isBinary):
-
-        # note: only the file name is used to store the file:
-        # the directory is RTT run dependent!
-
-        fullPath = os.path.join(dir, fileName)
-        if not os.path.exists(fullPath):
-            self.logger.warning('Cannot store non existent file %s' % fullPath)
-            raise Exception
-        
-            
-        flag="r"
-        if isBinary: flag="rb"
-
-        self.isBinary = isBinary
-        self.key      = key.__repr__()
-        try:
-            file  = open(fullPath,flag)
-        except Exception, e:
-            self.logger.error('Error opening file %s' % fullPath)
-            self.logger.error(e)
-            self.logger.error(exc2string2())
-            raise e
-        
-        try:
-            self.strFile  = file.read()
-            file.close()
-        except Exception, e:
-            self.logger.error('Error reading from file %s' % fullPath)
-            self.logger.error(e)
-            self.logger.error(exc2string2())
-        self.fullPath = fullPath
-
-    def __str__(self):
-        return self.key
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/Minder2.py.obsolete b/Tools/RunTimeTester/src/Minder2.py.obsolete
deleted file mode 100755
index dc349e087ea6ea14bf5081afb70d607e6262cb66..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Minder2.py.obsolete
+++ /dev/null
@@ -1,963 +0,0 @@
-"""
-class Minder. Top of the Minder Heirarchy.
-
-- Takes care of the state machine. This is a mixture of
-a status object and a dispatcher.
-
-The state machine  has states: queued, running, success, error,
-it nows how to proceed from one state to the next, and species the actions to be
-performed when the state changes. The way jobs are submited, and the names of
-log files are examples of items that change with mode.
-
-This class also has the following methods, which are common to all Minders,
-
-poll
-fromQueued
-fromRunning
-fromSuccess
-fromError
-setStatus
-setStatusList
-forceCloseDown
-setDone
-
-Many of these are overidden in derived classes.
-"""
-
-from MinderBase                     import MinderBase
-from formatCollection               import formatCollection
-from Tools                          import changePerm
-from Defs                           import RTTdefs
-from ModuleLoader                   import ModuleLoader
-from exc2string2                    import exc2string2
-from RTTTestRunner                  import RTTTestRunner  
-from UserLogger                     import UserLogger
-from Tools                          import unique
-from Tools2                         import toRelPath, popenCmd
-from BigFileIO                      import createBigFileIO
-from RTTdict                        import RTTdict
-from RTTSException                  import RTTCodingError
-from DDDescriptor                   import DDDescriptor
-from PPPaths                        import PPPaths
-
-import logging
-logger = logging.getLogger('rtt')
-
-import os, shutil, fnmatch, shelve
-
-# -------------------------------------------------------------------------
-#from Tracer        import Tracer
-# uncomment to monitor all method calls (shows args + return values)
-#__metaclass__ = Tracer
-
-from MethodTimer import MethodTimer
-
-
-class Minder(MinderBase):
-    def __init__(self, argBag, jDescriptor):
-        """ One function of the minder is to bring together information
-        from the job descriptor (gathered from the jobs config file),
-        the job groups file (gathered from the job groups config file),
-        and the Paths object (gathered from the RTT config file)
-
-        For files to be transferred from the run directory to the results
-        area, we use patterns (for file name matching). Lists of patterns
-        are made at in this method, with a list per use (keep files). These lists will be used to identify
-        filenames at the appropriat e moment, and converted to absolute
-        path names. Source and destination will be kept in the dictionaries
-        self.keepFiles
-        
-        """
-
-
-
-        MinderBase.__init__(self, argBag.logDir,
-                            argBag.package,
-                            jDescriptor.identifiedName,
-                            jDescriptor.jobSerialNumber,
-                            argBag.elementCreator,
-                            argBag.textNodeCreator)
-
-        self.shareArea           = argBag.shareArea
-        self.deleteNonKeepFiles  = argBag.deleteNonKeepFiles
-        self.confFile            = argBag.confFile
-        self.bigFilesArea        = argBag.bigFilesArea
-        self.bigFilesSize        = argBag.bigFilesSize
-        self.resultsBasePath     = argBag.resultsBasePath
-        self.refRelease          = argBag.refRelease
-        self.isNightly           = argBag.isNightly
-        self.workDirs            = argBag.workDirs
-        self.site                = argBag.site
-        self.RTTLibDir           = argBag.RTTLibDir         # used for post processing
-        self.RTTSrcDir           = argBag.RTTSrcDir         # used for post processing
-        self.dCubeCfgFile        = argBag.dCubeCfgFile      # used for post processing
-        self.installArea         = argBag.installArea       # used for post processing
-        self.containerPackage    = argBag.containerPackage  # used for post processing
-        self.runType             = argBag.runType           # used for post processing
-        self.localRTTRun         = argBag.localRTTRun       # used for post processing
-        self.distArea            = argBag.distArea          # used for post processing
-        self.cmtPath             = argBag.cmtPath           # used for post processing
-        self.cmtConfig           = argBag.cmtConfig         # used for post processing
-        self.cmtLinesCmds        = argBag.cmtLinesCmds      # used for post processing
-        self.topProject          = argBag.topProject        # used for post processing
-        self.branch              = argBag.branch            # used for post processing
-        self.release             = argBag.release           # used for post processing
-        self.package             = argBag.package           # used for post processing
-        
-        self.runPath             = jDescriptor.runPath
-        self.name                = jDescriptor.name
-        self.jobDisplayName      = jDescriptor.jobDisplayName
-        self.jobGroup            = jDescriptor.jobGroup
-        self.jobDocString        = jDescriptor.jobDocString    
-        self.jobDocURL           = jDescriptor.jobDocURL    
-        self.rttPilotJob         = jDescriptor.rttPilotJob
-        self.rttATNJob           = jDescriptor.rttATNJob
-        self.log                 = jDescriptor.log
-        self.elog                = jDescriptor.elog
-        self.errorMessages       = jDescriptor.errorMessages
-        self.resPath             = jDescriptor.resPath
-        self.descDataForXMLNode  = jDescriptor.dataForXMLNode()
-        self.keepFiles           = jDescriptor.keepFiles
-        self.datasets            = jDescriptor.datasets
-        self.missingDatasets     = jDescriptor.missingDatasets
-        self.trendId             = jDescriptor.trendId
-
-        # Combine some stuff from kit and job
-        self.keepFilePatterns    = jDescriptor.keepFilePatterns + argBag.jobGroupKit['keepFilePatterns']
-        self.auxFilePatterns     = jDescriptor.auxFilePatterns + argBag.jobGroupKit['auxFilePatterns']
-        self.actionDescriptors   = jDescriptor.actions + argBag.jobGroupKit['actions']
-        self.testDescriptors     = jDescriptor.tests + argBag.jobGroupKit['tests']
-        
-        self.testDBPath          = os.path.join(self.runPath, 'RTTtests.db')
-
-        # These jobs are NOT counted in  Launcher stats. Otherwise overide.
-        self.jobWeight      = 0 
-
-        # bring together information on files to manipulate.
-        # Keep and aux files are handled as patterns - the actual file
-        # name resolution is deferred to later
-        # Reference files references are complete file names due to
-        # fact that they will go into keys before retirval - and patterns wont work
-        
-
-        self.finishedTests        = [] # a list of execcuted test objects.
-        self.checkResults         = [] # results of tests which have been shelved.
-        self.processingResult     = 'unavailable'
-        self.postProcessingResult = 'unavailable'
-        self.chainLength          = 0
-        self.chainSuccesses       = 0
-        # declare dictionaries with source and destination paths
-        # for files to be transferred from the run directories to the resulsts
-        # directory.
-
-        # declare lists of files - no  wild cards, as the elements of the list
-        # are used in data base keys. Also,  no absolute paths for the same reason
-        # (would break portability)
-        
-        
-        # maximum no of times to retry a job if it enters the error state.
-        self.errorStateMaxRetry = 3
-        self.errorStateCurRetry = 0
-          
-
-        self.exitStatus  = 'Unknown'
-        self.cpuTime     = 'Unknown'
-        self.cpuTime2000 = 'Unknown'
-        self.mem         = 'Unknown'
-        self.vmem        = 'Unknown'
-        self.wallTime    = 'Unknown'
-        self.batchStatus = 'Unknown'
-        
-
-        self.scripts             = {}
-        self.timedOut            = False
-        # so complete file names go here here as well.
-
-        self.handleFileNames()
-        self.collectActionsAndTests()
-
-        self.collectScripts(argBag.scriptWriter) # after collecting tests and actions
-        # self.setupJobGroupDirs() # after making scripts
-
-   
-        jobGroupDirectoryMaker        = argBag.jobGroupDirectoryMaker
-        jobGroupDirectoryMaker.logger = self.logger
-
-        self.makeJobGroupDirs(jobGroupDirectoryMaker)
-        self.setupJobGroupDirs(jobGroupDirectoryMaker)
-
-        self.shelveTests()
-
-    
-    def makeJobGroupDirs(self, jobGroupDirectoryMaker):
-        try:
-            jobGroupDirectoryMaker.makeJobGroupDirs(self.runPath, self.resPath)
-        except:
-            msg  = 'Exception making directories by %s traceback:\n%s' % (self.identifiedName, exc2string2())
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-        
-
-    def unlink(self):
-        'break circular references to allow garbage collection'
-
-        self.stateEngine.unlink()
-        del self.stateEngine
-        self.xmlConverter.unlink()
-        del self.xmlConverter
-        MinderBase.unlink(self)
-
-        
-    def calcPostProcessingResult(self):
-        if not (self.finishedTests or self.checks):
-            self.postProcessingResult = 'no tests'
-            return
-
-        for t in self.finishedTests:
-            if t.result != 0:
-                self.postProcessingResult = 'error'
-                return
-
-        if 'error' in self.checkResults:
-            self.postProcessingResult = 'error'
-            return
-        
-        self.postProcessingResult = 'success'
-
-    def collectScripts(self, scriptWriter):
-        # post processing scripts are handled here, and so
-        # apply to Watcher and Worker minders.
-        # NOTE 26/01/09
-        # self.postProcessingCommand is minder specific.
-        # it is set in LinuxInteractive and LSFBatch, not yet set in WatcherMinder
-        
-        if not (self.tests or self.actions): return
-
-        script = scriptWriter.postProcessorScript()
-
-        dict = {'postProcessorScript': {'script':     script,
-                                        'scriptName': 'postProcessorScript.sh',
-                                        'cmd':         self.postProcessCommand
-                                        }
-                }
-        self.scripts.update(dict)
-
-
-
-# ------------------------------------------------------------------------
-
-    def shelveTests(self):
-        # associate string identifiers with each test and action
-        # t.position is the user given order index
-        npos = 0
-
-        testIds = []
-        for runOrder, instanceList in self.tests.items():
-            for testInstance in instanceList:
-                item = (testInstance, '%s_%d' % (testInstance.__class__.__name__,npos), runOrder)
-                testIds.append(item)
-                npos +=1
-
-        actionIds = []
-        for runOrder, instanceList in self.actions.items():
-            for actionInstance in instanceList:
-                item = (actionInstance, '%s_%d' % (actionInstance.__class__.__name__,npos), runOrder)
-                actionIds.append(item)
-                npos +=1
-                
-        if npos == 0: return
-
-        # provide a database to store the tests
-        db = shelve.open(self.testDBPath, 'n')
-        db.close()
-
-        # wrap the tests to allow transportation to the computing node
-        wrappers = []
-        
-        isAction = False
-        wrappers.extend([RTTTestRunner(self.testDBPath, t, isAction) for t in testIds])
-        isAction = True
-        wrappers.extend([RTTTestRunner(self.testDBPath, t, isAction) for t in actionIds])
-
-        # Write the tests to a shelve db
-        for w in wrappers:
-            try:
-                w.autoShelve()
-            except:
-                m = 'Error shelving %s\nTraceback: %s' % (w.testId, exc2string2())
-                self.logger.error(m)
-                print m
-                
-    # -------------------------------------------------------------------------
-
-    def hasChecks(self):
-        answer = False
-
-        if hasattr(self,'checks'):
-            answer = len(self.checks) > 0 or answer
-
-        if hasattr(self,'tests'):  
-            answer = len(self.tests.keys()) > 0 or answer
-
-        return answer
-
-# ------------------------------------------------------------------------
-
-    def collectActionsAndTests(self):
-        """
-        Test Descriptor is a list with four elements:
-        position, modulename, classname, dictionary of arguments
-        """
-
-
-        # set up a "pruned descritpor". In 01/09, there was a large refactoring of the RTT
-        # and the role of paths and descriptors was reduced. However user suppied test scripts may
-        # be expecting these variables, so they are rebuilt here.
-        # class DDDescriptor:pass
-        # class Paths: pass
-        def makePrunedDescriptor():
-            desc = DDDescriptor()
-            possibleAttrs = ['log', 'name', 'trendId', 'identifiedName', 'jobSerialNumber', 'jobDisplayName',
-            'hashString', 'runPath', 'resPath', 'jobGroup', 'jobDocString', 'jobDocURL', 'package']
-            [setattr(desc, p, getattr(self, p)) for p in possibleAttrs if hasattr(self, p)]
-
-
-            paths = PPPaths()
-            possibleAttrs =  ['installArea', 'containerPackage', 'cmtPath', 'shareArea',
-                              'release', 'cmtHomeDir', 'localRTTRun', 'workDirs', 'resultsDirs', 'runType',
-                              'build', 'originalBranch', 'branch', 'topProject', 'otherProject', 'cmtConfig',
-                              'dCubeCfgFile', 'isNightly', 'RTTSrcDir', 'RTTLibDir', 'pathSegment', 'package',
-                              'packageTag', 'fullPackageName', 'dCubeRefBaseDir', 'confFile', 'userHomeCmtDir',
-                              'testProjectCmtDir', 'distArea', 'projectCMTDirs']
-            [setattr(paths, p, getattr(self, p)) for p in possibleAttrs if hasattr(self, p)]
-
-            desc.paths = paths
-            return desc
-
-        def process(descs, tgt):
-            for descriptor in descs:
-                # Create the pruned descriptor
-                prunedDesc  = makePrunedDescriptor()
-                # prunedDesc  = self.jDescriptor.prune()
-                # prunedPaths = self.jDescriptor.paths.prune()
-                # setattr(prunedDesc, 'paths', prunedPaths)
-
-                # Add params to the dict
-                descriptor.addParameter(('JobDescriptor', prunedDesc))
-                descriptor.addParameter(('cmtLinesCmds', self.cmtLinesCmds[:]))
-                instance = self.getClassFromDescriptor(descriptor)
-                if instance != None:
-                    tgt.setdefault(descriptor.position, []).append(instance)
-
-        process(self.testDescriptors, self.tests)
-        process(self.actionDescriptors, self.actions)
-          
-#-----------------------------------------------------------------------
-
-    def getClassFromDescriptor(self, taDescriptor): # descriptor is a test or action descriptor
-        def runUserScriptInAtlasEnv():
-            return taDescriptor.runInAtlasEnv!=""
-        def createRTTdict(rtt_dict_keys, paramDict):
-            rtt_dict = {}
-            for key in rtt_dict_keys:
-                rtt_dict[key] = paramDict[key]
-                del paramDict[key]
-            user_dict = paramDict
-            return RTTdict(rtt_dict, user_dict)
-        
-        # allow RTT library tools to be picked up from the local code base
-        # rather than the installed shared area if so requested in the
-        # RTT configuration file
-        sharePathToUse = {'RttLibraryTools' : self.RTTLibDir}
-        instance = None
-
-        # RTTLibTools has been split into modules with one class per
-        # module, with the module name = the class name. Needed this
-        # for shelve to unshelve instances.
-
-        moduleName = taDescriptor.moduleName
-        sharePathIs = sharePathToUse.get(moduleName, self.shareArea)
-
-        # now set module name = testname for shelving constraint
-        moduleName = taDescriptor.testName
-        className  = taDescriptor.testName
-        self.logger.debug("modulename set = classname, i.e. %s" % moduleName)
-        paramDict  = taDescriptor.argDict
-
-        rtt_dict_keys = ['JobDescriptor']
-
-        if runUserScriptInAtlasEnv():
-            sharePathIs = sharePathToUse.get('RttLibraryTools')
-            paramDict['userScript'] = moduleName
-            paramDict['userScriptLoc'] = sharePathToUse.get(taDescriptor.moduleName,
-                                                            self.shareArea)
-            moduleName = className  = 'AtlasEnvSetup'
-            
-            rtt_dict_keys.extend(['userScript', 'userScriptLoc'])
-
-        paramDict = createRTTdict(rtt_dict_keys, paramDict)
-        
-        # pickle the dictionary to the run path: actually gets dumped in JobGroupDirectoryMaker.setupRunDir
-        self.rttArgDictPickle = {'what': paramDict, 'where': os.path.join(self.runPath,'rtt.argdict.cpickle')}
-        
-        # default is release share path if key not in dict
-        # this happens when the user supplies own module, which is
-        # picked up from the release area.
-
-        self.logger.debug('module path ' + sharePathIs)
-        self.logger.debug('moduleName  ' + moduleName)
-        self.logger.debug('className   ' + className)
-        self.logger.debug('paramDict   ' + str(paramDict))
-        self.logger.debug('logger      ' + str(self.logger))
-
-        try:            
-            mL        = ModuleLoader(moduleName, sharePathIs, self.logger)
-            instance  = mL.getClassInstance(className, paramDict)
-            self.logger.debug('Created a test instance of class %s' % (instance.__class__.__name__))
-        except Exception, e:
-            self.logger.error('Failed to create a test instance')
-            self.logger.error(exc2string2())
-            self.logger.error(str(e))
-            
-        return instance
-    
-#------------------------------------------------------------------------
-
-    def handleFileNames(self):
-        """ method to collect file names, and resolve patterns where
-        for those patterns for which it is possible at init."""
-   
-        # expand the wild cards - but do not create the full directory path
-        # as the work sub directories have yet to be created.
-        if not os.path.exists(self.shareArea):
-            m = 'Cannot set self.auxfiles due to non-existent share directory: %s' % self.shareArea
-            self.logger.fatal(m)
-            raise RTTCodingError(m)
-
-        # resolve auxFile patterns to file names
-        auxFiles = []
-        for pattern in self.auxFilePatterns:
-            base, fnpattern = os.path.split(pattern)
-            srcDir = os.path.normpath(os.path.join(self.shareArea, base))
-            filesInShare = os.listdir(srcDir)
-            auxFiles.extend([os.path.join(base,file) for file in filesInShare if fnmatch.fnmatch(file, fnpattern)])
-
-        self.auxFiles = unique(auxFiles)
-
-    # ------------------------------------------------------------------------
-
-    def reportUserScriptError(self,message):
-        filePath  = os.path.join(self.runPath,'Python_Script_Output.log')
-        h = open(filePath,'a+')
-        h.write(message)
-        h.close()
-
-    #-----------------------------------------------
-    
-    def readTestResults(self):
-        #make UserLogger to transfer messages
-        #from UserLogger import UserLogger
-        ulogger = UserLogger().makeLogger(self.runPath, self.identifiedName)
-        self.logger.debug('made logger for tests and actions :'+str(logger))
-
-        db = shelve.open(self.testDBPath)
-
-        dbValues = []
-        try:
-            dbValues.extend(db.values())
-            self.finishedTests = db.values()
-        except Exception, e:
-            m  = 'Exception thrown trying to read values from %s\n' % str(self.testDBPath)
-            m += str(e)            
-            self.logger.error(m)
-            # now close DB to prevent too many open files
-            db.close()
-        else:
-            db.close()
-        
-        self.finishedActions = [t for t in self.finishedTests if t.isAction]
-        [self.finishedTests.remove(t) for t in self.finishedActions]
-        
-        # for t in db.values():
-        for t in dbValues:
-            if t.error:
-                self.logger.error('Error running test %s' % t.testId)
-            else:
-                self.logger.debug('No error running test %s' % t.testId)
-
-
-            # collect any messages generated by the test
-            #t.transferLog(self.logger)
-            t.transferLog(ulogger)
-            ulogger.debug('-----------new test-------------')
-            # db.close()
-        self.calcPostProcessingResult()
-       
-    #--------------------------------------------------
-    
-    #def runMoniActions(self):
-    #jobName  = self.jDescriptor.identifiedName
-    #nActions = len(self.monActions)
-    #msg      = 'Running %d monActions for job %s' % (nActions, jobName)
-    #self.logger.debug(msg)
-    #
-    #for action in self.monActions:
-    #className = str(action.__class__.__name__)
-    #try:
-    #self.logger.debug("Running monAction " + className)
-    #dataDict=action.run()
-    #self.monActionsData.append(dataDict)
-    #
-    # except Exception, e:
-    # msg = "Could not run monitoring action " + className
-    # self.logger.error(msg)
-    # self.logger.error(str(e))
-    # self.logger.error(exc2string2())
-    # msg  = '-----------------------------------------\n'
-    # msg += 'MoniAction: ' + className + ' could not be run!\n'
-    # msg += exc2string2() + '\n'
-    # msg += '-----------------------------------------\n\n'
-    # self.reportUserScriptError(msg)
-    # self.logger.debug("Running next available monAction")
-    #       
-    #
-    # self.logger.debug('Minder moniData :'+str(self.monActionsData))
-    # self.logger.debug("Finished running monActions for %s" % jobName)
-      
-    #--------------------------------------------------
-
-    def fullResultsFileName(self, filename):
-        return os.path.join(self.resPath, filename)
-
-    #-----------------------------------------------
-    
-    def makeKeepFileEntry(self, file, infoString='',displayColor='', md5=''):
-        "helper method for registerWildKeepFiles()"
-        
-        src = os.path.join(self.runPath, file)
-
-        dest = {'keepFileString': self.fullResultsFileName(file),
-                'infoString'    : infoString,
-                'displayColor'  : displayColor,
-                'md5sum'        : md5}
-        
-        self.keepFiles[src]=dest
-        return dest['keepFileString']
-    #-----------------------------------------------
-    
-    def registerWildKeepFiles(self):
-        """
-        Common implementation task.
-
-        Obtain the wild card patterns for the current job group from
-        a JobGroupKit.
-        
-        Collect all the wild carded keep files.
-        These are files that are present in the run directory at when
-        fromRunning is called, and which match a pattern in the pattern
-        list.
-
-        Give these files their full paths and add them to the keepFile
-        dictionary.
-        """
-
-        for card in self.keepFilePatterns:
-
-            keepString   = card['keepFileString'] # may contain subdirs
-            infoString   = card['infoString']
-            displayColor = card['displayColor']
-
-            keepStringTokens = keepString.strip().split('/')
-
-            if len(keepStringTokens) == 1: # current dir keepfile pattern
-                wildFiles = fnmatch.filter(os.listdir(self.runPath), keepString)
-                [self.makeKeepFileEntry(file, infoString, displayColor) for file in wildFiles]
-
-            elif len(keepStringTokens) > 1: # subdirs
-                matches = ['']
-                for pathPart in keepStringTokens:
-                    newMatches = []
-                    for match in matches:
-                        conts = os.listdir(os.path.join(self.runPath, match))
-                        newMatches.extend([os.path.join(match, f) for f in fnmatch.filter(conts, pathPart)])
-
-                    matches = newMatches
-                [self.makeKeepFileEntry(m, infoString, displayColor) for m in matches]
-
-        # now manually add the package configuration file to keep files
-        self.makeKeepFileEntry(os.path.basename(self.confFile),
-                               "Package XML test configuration file",
-                               "#cc3333")
-
-
-    #-----------------------------------------------
-
-    def cleanSpace(self):
-        # important that this be called before move big files otherwise
-        # we may move files that would have been deleted i.e. waste space
-        self.removeNonKeepFiles()        
-        self.moveBigFiles()    
-    
-    # --------------------------------------------------------------------
-
-    def getBigFiles(self):
-        """Looks in the run/results dir for big files. Returns a list of them."""
-        toMove = {self.resPath:[], self.runPath:[]}
-
-        for tgt in toMove.keys():
-            for root, dirs, files in os.walk(tgt):
-                files = [os.path.join(root, f) for f in files]
-                def isBig(f): # only files, no links
-                    return os.path.isfile(f) and not os.path.islink(f) and os.path.getsize(f)>=self.bigFilesSize
-                toMove[tgt].extend([f for f in files if isBig(f)])
-        return toMove
-    
-    # --------------------------------------------------------------------
-
-    def moveBigFile(self, src, dst):
-        try:
-            fileSize = os.path.getsize(src)
-            shutil.move(src, dst)
-            m = 'Successfully moved big file %s [%s bytes] to %s' % (os.path.basename(src),
-                                                                     str(fileSize),
-                                                                     dst)
-            self.logger.info(m)
-        except:
-            m  = 'Unable to move big file %s to %s\n' % (os.path.basename(src), dst)
-            m += exc2string2()
-            m += str(os.listdir(os.path.dirname(src)))
-            self.logger.error(m)
-            
-    # --------------------------------------------------------------------
-
-    def makeReplacementKeepFile(self, bigFileSrc, bigFileDump):
-        """If we move a big file from run path to separate volume,
-        and that file just happened to be a keep file, then we need
-        to replace it. Create a soft link in resPath to separate volume
-        and add an expiry date."""
-
-        linkName = os.path.basename(bigFileSrc)
-        linkTgt  = os.path.join(toRelPath(os.path.dirname(bigFileSrc), bigFileDump), linkName)
-        
-        currDir = os.getcwd()
-        os.system('cd %s;ln -s %s %s;cd %s' % (os.path.dirname(bigFileSrc), linkTgt, linkName, currDir))
-
-        # add an md5sum to the keepfile to guard against mishap if the file pointed
-        # at was overwritten (the web link would work, but you'd get the
-        # overwriter, not the expected overwritee, and you would never know...
-        # If md5sum different between now calculated and future web page one,
-        # then grey-out the link.
-        md5sum = popenCmd('md5sum %s' % bigFileSrc)
-        self.logger.debug('Minder md5sum calculated is: %s' % str(md5sum))
-        md5sum = md5sum[0].split()[0].strip()
-        srcRelToRunPath = bigFileSrc.split(self.runPath)[1]
-        if srcRelToRunPath.startswith('/'): srcRelToRunPath = srcRelToRunPath[1:]
-        self.makeKeepFileEntry(srcRelToRunPath, md5=md5sum)
-        
-    # --------------------------------------------------------------------                                
-
-    def moveBigFiles(self):
-        """Move all files bigger than some value onto a separate volume."""
-        if not self.bigFilesArea:
-            self.logger.info('Moving of big files to a separate volume has not been requested.')
-            return
-
-        self.logger.info('Moving of big files to a separate volume is requested. Scanning...')
-        
-        if not os.path.exists(self.bigFilesArea):
-            m = 'Cannot shift big files onto inexistent volume: %s' % self.bigFilesArea
-            self.logger.error(m)
-            return
-        
-        bigFiles = self.getBigFiles()
-
-        if not [val for val in bigFiles.values() if val]:
-            self.logger.info('No big files were found, returning.')
-            return
-        
-        placeToDump = createBigFileIO(self.site, self.bigFilesArea, self.workDirs, self.isNightly).getJobDumpLocation(self)
-        if not placeToDump:
-            m = 'Unable to retrieve location of big files volume. Not moving big files.'
-            self.logger.warning(m)
-            return
-
-        # We have files to move, let's move them
-        for bigFileBaseDir, bigFiles in bigFiles.items():
-            for bigFile in bigFiles:
-                src = bigFile # file
-                dst = placeToDump # directory
-                self.moveBigFile(src, dst)
-                # If big file origin is results path, replace with a soft link
-                # to separate big file volume.
-                if bigFileBaseDir == self.resPath:
-                    self.makeReplacementKeepFile(bigFile, placeToDump)
-            
-    # --------------------------------------------------------------------                            
-
-    def removeNonKeepFiles(self):
-        """Delete all files in the run dir not explicitly kept."""
-
-        if not self.deleteNonKeepFiles:
-            self.logger.info('Deletion of non keep files (to save space) has not been requested.')
-            return
-
-        self.logger.info('Deletion of non-keep files is requested. Scanning...')
-        
-        toDelete = []
-        for root, dirs, files in os.walk(self.runPath):
-            toDelete.extend([os.path.join(root, f) for f in files if os.path.join(root, f) not in self.keepFiles.keys()])
-
-        for thing in os.listdir(self.runPath):
-            self.logger.debug('Before delete: %s' % str(thing))
-            
-        for thing in toDelete:
-            try:
-                os.remove(thing)
-                self.logger.debug('Deleted: %s' % thing)
-            except:
-                message  = 'Unable to delete non-keepfile: %s\n' % thing
-                message += exc2string2()
-                self.logger.error(message)
-
-        for k in self.keepFiles.keys():
-            self.logger.debug('Keepfile: %s' % str(k))
-        
-    # --------------------------------------------------------------------
-    
-    # copy files to be kept (log histos, ntuples...) to results dir
-    def copyKeepFiles(self):
-        
-        # find if any of the wild keep cards have a match
-        idName = str(self.identifiedName)
-        try:
-            self.registerWildKeepFiles()
-        except Exception, e:
-            msg = "Exception registering keep files for job: " + idName
-            self.logger.error(msg)
-            self.logger.error(str(e))
-            self.logger.error(exc2string2())
-
-        msg  = 'About to transfer %d keep files' % len(self.keepFiles.keys())
-        self.logger.debug(msg)
-
-        for file in self.keepFiles.keys():
-            
-            if not os.path.exists(file):
-                msg = "%s does not exist! Skipping transfer." % str(file)
-                self.logger.error(msg)
-                continue
-                
-            dictValue = self.keepFiles[file]
-            desFile = dictValue['keepFileString'] # this is a filename
-
-            try:
-                fileName = os.path.basename(str(file))
-                srcDir   = os.path.dirname(str(file))
-                tgtDir   = os.path.dirname(str(desFile))
-
-                if srcDir == tgtDir:
-                    msg  = 'Keepfile src and tgt dirs are same.\n'
-                    msg += '%s\n' % srcDir
-                    msg += '===> Not transferring file %s' % fileName
-                    self.logger.debug(msg)
-                    continue
-                self.logger.debug("Moving file %s from %s to %s" % (fileName, srcDir, tgtDir))
-                if not os.path.exists(os.path.dirname(desFile)):
-                    self.logger.debug('Destination dir does not exist. Making: %s' % os.path.dirname(desFile))
-                    try:
-                        os.makedirs(os.path.dirname(desFile))
-                    except:
-                        self.logger.error('Unable to create %s so as to copy keep file to it.' % os.path.dirname(desFile))
-                        self.logger.error(exc2string2())
-                        continue
-                    
-                shutil.move(file, os.path.dirname(desFile))
-                changePerm(desFile, 02775)
-            except:
-                # disk space problems?
-                message = '***** COPYING OF KEEP FILES PROBLEM! *****\n'
-                message += 'Unable to copy src file:\n'
-                message += '   ' + file + '\n'
-                message += 'to destination file:\n'
-                message += '   ' + desFile + '\n'
-                message += 'Doing command df on ' + desFile + ' yields the answer:\n'
-                message += str(os.popen('df ' + desFile).read())
-                message += '---------------------------------'                    
-                self.logger.error(message)
-
-        self.logger.debug("Finished copying keepfiles")
-
-    # --------------------------------------------------------------------
-        
-    def runChecks(self):
-        
-        theJob = str(self.identifiedName)
-        msg = 'Len self.checks is: %s for %s ' % (str(len(self.checks)),theJob)
-        self.logger.debug(msg)
-        
-        for check in self.checks:
-
-            check.setLogger(self.logger) # log to the minders logfile
-
-            if not callable(check):
-                msg = "Uncallable checker: %s for job %s" % (str(check),
-                                                             theJob)
-                self.logger.error(msg)
-                self.checkResults.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            self.logger.debug('Job: %s calling check: %s'%(theJob, str(check)))
-            status = ''
-            try:
-                rc = check(self) # int return code from check
-                status = RTTdefs.status(rc) # convert to string
-            except Exception, e:
-                msg = 'Exception raised while executing %s' % str(check)
-                self.logger.error(msg)
-                self.logger.error(exc2string2())
-                self.logger.error(str(e))
-                self.checkResults.append('error')
-                self.logger.error('Proceding to the next check')
-                continue
-            
-            msg = 'Job: %s  check: %s return status %s' % (theJob,
-                                                           str(check),
-                                                           status)
-            self.logger.debug(msg)
-            
-            self.checkResults.append(status)
-        
-        self.logger.debug(msg)
-
-    # --------------------------------------------------------------------       
-    """
-    success if the job terminates, subclasses will override with a
-    more precise meaning
-    """
-    def isSuccess(self): return self.isDone()
-        
-    # --------------------------------------------------------------------
-
-    def setupJobGroupDirs(self, jobGroupDirectoryMaker):
-        try:
-            jobGroupDirectoryMaker.setupRunDir(self)
-        except:
-            msg  = 'Error  setting up run directory by %s: TraceBack:\n%s' % (
-                self.identifiedName,
-                exc2string2()
-                )
-            self.logger.warning(msg)
-            raise RTTCodingError(msg)
-
-
-    def dataForMonitoring(self):
-        """
-        return a dictionay of values for monitoring. repeats data in DOM
-        document, but is faster to access.
-        """
-        dict = MinderBase.dataForMonitoring(self)
-        
-        dict['nTests']        = len(self.tests.keys())
-        dict['done']          =   self.isDone()
-        dict['nTestsSuccess'] = len([s for s in self.finishedTests if s.result == 0])
-        dict['nTestsFailure'] = len([s for s in self.finishedTests if s.result != 0])
-        dict['nRetries']      = self.errorStateCurRetry
-        dict['ppFailure']     = (self.postProcessingResult == 'error')
-        dict['ppSuccess']     = (self.postProcessingResult == 'success')
-
-        return dict
-    def __str__(self):
-        s = '----------- Minder ---------------\n'
-        s += ' done:                 '+str(self.isDone())+'\n'
-        s += ' weight:               '+str(self.jobWeight)+'\n'
-        s += ' runPath:              '+str(self.runPath)+'\n'
-        s += ' keepFilePatterns:     '+formatCollection(self.keepFilePatterns)+'\n'
-        s += ' auxFilePatterns:      '+formatCollection(self.auxFilePatterns)+'\n'
-        s += ' keepFiles:            '+formatCollection(self.keepFiles)+'\n'
-        s += ' auxFiles:             '+formatCollection(self.auxFiles)+'\n'
-        s += ' actions:              '+formatCollection(self.actions)+'\n'
-        s += ' tests:                '+formatCollection(self.tests)+'\n'
-        s += ' scripts:              '+str(self.scripts.keys())+'\n'
-        s += ' descriptor:\n'
-
-        return s
-
-    # --------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|            Minder  dump                   |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('\n'+self.__str__())
-        
-    def status(self):
-        if self.operatorCloseDown: return 'operatorCloseDown'
-        return self.stateEngine.state.state
-
-    def printMethodTimes(self):
-        
-        m = '------ Minder  %s ---------\n%s' % (
-            self.identifiedName,
-            self.formatMethodTimes())
-
-
-        m += '\n-----State engine ---------  \n%s' % (
-            self.stateEngine.formatMethodTimes())
-
-        m += '\n-----XML Converter ---------  \n%s' % (
-            self.xmlConverter.formatMethodTimes())
-        
-        self.logger.info(m)
-
-# --------------------------------------------------------------------------
-            
-    def doPostProcessing(self):
-        if 'postProcessorScript' in self.scripts.keys(): return True
-        return False
-            
-# =============================================================================
-    
-class StoreItem:
-    def __init__(self, dir, fileName, key, isBinary):
-
-        # note: only the file name is used to store the file:
-        # the directory is RTT run dependent!
-
-        fullPath = os.path.join(dir, fileName)
-        if not os.path.exists(fullPath):
-            m = 'Cannot store non existent file %s' % fullPath
-            self.logger.warning(m)
-            raise RTTCodingError(m)
-        
-            
-        flag="r"
-        if isBinary: flag="rb"
-
-        self.isBinary = isBinary
-        self.key      = key.__repr__()
-        try:
-            file  = open(fullPath,flag)
-        except Exception, e:
-            m = 'Error opening file %s\nException:\n%s\nTraceback:\n%s' % (fullPath, str(e), exc2string2())
-            self.logger.error(m)
-            raise RTTCodingError(m)
-        
-        try:
-            self.strFile  = file.read()
-            file.close()
-        except Exception, e:
-            self.logger.error('Error reading from file %s' % fullPath)
-            self.logger.error(e)
-            self.logger.error(exc2string2())
-        self.fullPath = fullPath
-
-
-    def __str__(self):
-        return self.key
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/MinderBase.py b/Tools/RunTimeTester/src/MinderBase.py
deleted file mode 100644
index 703348b4026854964a44cd239a8193df6491915e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderBase.py
+++ /dev/null
@@ -1,237 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-class MinderBase. Top of the Minder Heirarchy.
-
-- Takes care of the state machine. This is a mixture of
-a status object and a dispatcher.
-
-The state machine  has states: queued, running, success, error,
-it nows how to proceed from one state to the next, and species the actions to be
-performed when the state changes. The way jobs are submited, and the names of
-log files are examples of items that change with mode.
-
-This class also has the following methods, which are common to all Minders,
-
-poll
-fromQueued
-fromRunning
-fromSuccess
-fromError
-setStatus
-setStatusList
-forceCloseDown
-setDone
-
-Many of these are overidden in derived classes.
-"""
-
-from Factory_MinderStateEngine   import MinderStateEngineFactory
-from Factory_XMLConverter        import XMLConverterFactory
-from exc2string2                 import exc2string2
-from RTTSException               import RTTCodingError
-from Tools                       import nowSecs
-import RTTFileLogger
-
-import logging
-logger = logging.getLogger('rtt')
-
-import shelve
-
-import os.path
-#import cPickle
-
-
-
-from MethodTimer import MethodTimer
-#from Tracer      import Tracer
-
-class MinderBase(MethodTimer):
-    """
-    A robust (ie simple) base class for all minders.
-    Some, such as ErrorMinder and Container minders,
-    inherit only form this class. Because Error minders
-    are one of the notification mechanisms for problems,
-    this base class should always be instantiable (no exceptions
-    are to thrown)
-
-    The "done" flag is handled in this class as this is a fundemental
-    control flag for all minders. is minder.done = True, the minder
-    is dicarded form active minder lists.
-
-    The class provides:
-    - a logger
-    - a state engine
-    - an xmlconverter
-    - the done flag and simple methods for handling it
-    - identified name. This is a unique string for the minder. Used, e.g. for XML reporting
-    """
-
-    def __init__(self, logDir, package, identifiedName, serialNumber, elementCreator, textNodeCreator):
-        MethodTimer.__init__(self)
-
-        self.done            = False
-        self.logDir          = logDir
-        self.package         = package
-        self.identifiedName  = identifiedName
-        self.jobSerialNumber = serialNumber
-        self.elementCreator  = elementCreator
-        self.textNodeCreator = textNodeCreator
-        self.jobGroup        = None
-        self.rttPilotJob     = False
-        self.rttATNJob       = False
-
-        self.logger          = None
-        self.makeLog()
-        self.logger.debug('MinderBase init')
-        self.stateEngine     = None
-        try:
-            self.stateEngine = MinderStateEngineFactory(self.logger).create(self)
-        except Exception, e:
-            msg = 'Error instantiating  state engine:\n%s, Traceback:\n%s' %  (str(e), exc2string2())
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-            
-        msg =  'MinderBase instantiated a state engine of type %s' % self.stateEngine.__class__.__name__
-        self.logger.info(msg)
-
-        self.xmlConverter        = XMLConverterFactory(self.logger).create(self)
-        msg =  'MinderBase instantiated an xml converter of type %s' % self.xmlConverter.__class__.__name__
-        self.logger.info(msg)
-        
-
-        self.removeMe            = False
-        # self.monActions          = {} # key = run position, value = instances
-        self.actions             = {} # key = run position, value = instances
-        self.tests               = {} # key = run position, value = instances
-        self.checks              = []
-
-
-
-
-    # --------------------------------------------------------------------
-    def numberOfActions(self):
-        tot = 0
-        for k,v in self.actions.items():
-            tot += len(v)
-        return tot
-    
-    def numberOfTests(self):
-        tot = 0
-        for k,v in self.tests.items():
-            tot += len(v)
-        return tot
-        
-    def setSummaryCallBack(self, newCallBack):
-        self.xmlConverter.setSummaryCallBack(newCallBack)
-        self.logger.debug('Set summary call back%s' % str(newCallBack))
-        
-
-    def isDone(self): return self.done
-    
-    def setDone(self):
-
-        self.done = True
-        self.xmlConverter.makeSummaryReport()
-
-        # self.xmlConverter.makeSummaryReport() 
-        
-        m = 'done set in minder %s' % self.identifiedName
-        self.logger.info(m)
-        return 'terminate'
-
-    def setOperatorCloseDown(self):
-        self.stateEngine.setOperatorCloseDown()
-        self.poll()
-
-    def status(self):
-        return self.stateEngine.state.state
-
-    #def dataForXMLNode(self):
-    #    return self.dataForXMLNode
-
-    def dataForMonitoring(self):
-        "Added to by subclasses"
-
-        dict =   {
-            'name':              self.identifiedName,
-            'nTests':            0,
-            'done':              self.isDone(),
-            'operatorCloseDown': self.closedDownByOperator(),
-            'nTestsSuccess':     0,
-            'nTestsFailure':     0,
-            'nRetries':          0,
-            'jobSuccess':        False,
-            'jobFailure':        False,
-            'ppSuccess':         False,
-            'ppFailure':         False
-            }
-    
-    
-        return dict
-
-    
-
-# =============================================================================
-
-    def makeLog(self, loggerNameExt=''):
-        # set up a log file for this TestRun instance
-        # all logs for a pacakge go in a sub dir
-
-        logfile = os.path.join(self.logDir, self.package, self.identifiedName)
-
-        self.logger = RTTFileLogger.getLogger(self.identifiedName, logfile)
-
-    def __del__(self):
-        self.logger.info('Destructor for Minder %s' % self.identifiedName)
-
-    def unlink(self): pass
-
-    def poll(self):
-        """
-        polls the state engine. If the state engine
-        changes state, report and re- poll
-        """
-        m = '%s Start (Base) %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        self.logger.debug(m)
-        
-        transition = self.stateEngine.poll()
-        
-        if transition:
-
-            m = 'State transition detected (%s, %s): %s -> %s, calling back' % (self.__class__.__name__,
-                                                                                self.identifiedName,
-                                                                                self.stateEngine.pollEntryState,
-                                                                                self.stateEngine.pollExitState)
-            self.logger.info(m)
-
-            try:
-                self.xmlConverter.makeSummaryReport()
-            except:
-                m='Error on calling back to TestRun:\n%s' % exc2string2()
-                raise RTTCodingError(m)
-                
-            # self.poll()
-        m = '%s End (Base) %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        self.logger.debug(m)
-
-    def fromOperatorCloseDown(self):
-        self.setDone()
-        return 'done'
-
-    def closedDownByOperator(self): return self.stateEngine.closedDownByOperator()
-    
-    def fromDone(self):
-        self.setDone()
-        assert(self.done)
-        return 'done'
-
-    def currentState(self):
-        return self.stateEngine.state.state
-
-    def __str__(self):
-        s = '----------- %s ---------------\n' % self.__class__.__name__
-        for k in self.__dict__.keys():
-            s += ' 40%s:      %s\n' % (str(k), str(self.__dict__[k]))
-            return s
-
diff --git a/Tools/RunTimeTester/src/MinderBase.py.obsolete b/Tools/RunTimeTester/src/MinderBase.py.obsolete
deleted file mode 100644
index a84de7dbd4a317d48c4b2e7f6114f7bbb0a1d362..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderBase.py.obsolete
+++ /dev/null
@@ -1,200 +0,0 @@
-"""
-class Minder. Top of the Minder Heirarchy.
-
-- Takes care of the state machine. This is a mixture of
-a status object and a dispatcher.
-
-The state machine  has states: queued, running, success, error,
-it nows how to proceed from one state to the next, and species the actions to be
-performed when the state changes. The way jobs are submited, and the names of
-log files are examples of items that change with mode.
-
-This class also has the following methods, which are common to all Minders,
-
-poll
-fromQueued
-fromRunning
-fromSuccess
-fromError
-setStatus
-setStatusList
-forceCloseDown
-setDone
-
-Many of these are overidden in derived classes.
-"""
-
-from Factory_MinderStateEngine      import MinderStateEngineFactory
-from Factory_XMLConverter           import XMLConverterFactory
-from simpleDataForMonitoring        import simpleDataForMonitoring
-from RTTSException                  import RTTCodingError
-import RTTFileLogger
-
-import logging
-logger = logging.getLogger('rtt')
-
-import shelve
-
-import os.path
-#import cPickle
-
-# -------------------------------------------------------------------------
-#from Tracer        import Tracer
-# uncomment to monitor all method calls (shows args + return values)
-#__metaclass__ = Tracer
-
-
-from MethodTimer import MethodTimer
-
-
-class MinderBase(MethodTimer):
-    """
-    A robust (ie simple) base class for all minders.
-    Some, such as ErrorMinder and Container minders,
-    inherit only form this class. Because Error minders
-    are one of the notification mechanisms for problems,
-    this base class should always be instantiable (no exceptions
-    are to thrown)
-
-    The "done" flag is handled in this class as this is a fundemental
-    control flag for all minders. is minder.done = True, the minder
-    is dicarded form active minder lists.
-
-    The class provides:
-    - a logger
-    - a state engine
-    - an xmlconverter
-    - the done flag and simple methods for handling it
-    - identified name. This is a unique string for the minder. Used, e.g. for XML reporting
-    """
-
-    def __init__(self, logDir, package, identifiedName, serialNumber, elementCreator, textNodeCreator):
-        MethodTimer.__init__(self)
-
-        self.done            = False
-        self.logDir          = logDir
-        self.package         = package
-        self.identifiedName  = identifiedName
-        self.jobSerialNumber = serialNumber
-        self.elementCreator  = elementCreator
-        self.textNodeCreator = textNodeCreator
-        self.jobGroup        = None
-        self.rttPilotJob     = False
-        self.rttATNJob       = False
-
-        self.makeLog()
-        self.logger.debug('MinderBase init')
-
-        try:
-            self.stateEngine = MinderStateEngineFactory(self.logger).create(self)
-        except Exception, e:
-            msg = 'MinderStateEngine requires methods not furnished by Minder %s', str(e)
-            self.logger.error(msg)
-            raise RTTCodingError(msg)
-            
-        msg =  'MinderBase instantiated a state engine of type %s' % self.stateEngine.__class__.__name__
-        self.logger.info(msg)
-
-        self.xmlConverter        = XMLConverterFactory(self.logger).create(self)
-        msg =  'MinderBase instantiated an xml converter of type %s' % self.xmlConverter.__class__.__name__
-        self.logger.info(msg)
-        
-
-        self.removeMe            = False
-        # self.monActions          = {} # key = run position, value = instances
-        self.actions             = {} # key = run position, value = instances
-        self.tests               = {} # key = run position, value = instances
-        self.checks              = []
-
-
-
-        self.operatorCloseDown   = False
-
-    # --------------------------------------------------------------------
-    def numberOfActions(self):
-        tot = 0
-        for k,v in self.actions.items():
-            tot += len(v)
-        return tot
-    
-    def numberOfTests(self):
-        tot = 0
-        for k,v in self.tests.items():
-            tot += len(v)
-        return tot
-        
-    def setSummaryCallBack(self, newCallBack):
-        self.xmlConverter.setSummaryCallBack(newCallBack)
-        self.logger.debug('Set summary call back%s' % str(newCallBack))
-        
-    def poll(self):
-        self.stateEngine.poll()
-
-    def isDone(self): return self.stateEngine.state.isDone()
-    
-    def setDone(self):
-        self.done = True
-
-        # self.xmlConverter.makeSummaryReport() 
-        
-        m = 'done set in minder %s' % self.identifiedName
-        self.logger.info(m)
-
-    def setOperatorCloseDown(self):
-        self.operatorCloseDown = True
-
-    def status(self):
-        s = 'operatorCloseDown'
-        if s in self.stateEngine.state.history: return s
-        
-        return self.stateEngine.state.state
-
-    def dataForXMLNode(self):
-        return self.dataForXMLNode
-
-    def dataForMonitoring(self):
-        return simpleDataForMonitoring(self)
-    
-    def queueOperatorCloseDown(self): self.stateEngine.queueOperatorCloseDown()
-
-
-    def autoShelve(self):
-
-        def loggerSwapper( (a,b) ):
-            blogger = b.logger
-            a.logger = b.logger
-            b.logger = a.logger
-
-        class A:
-            def __init__(self):
-                self.logger = None
-
-        tmpList = [(self.xmlConverter, A()),
-                   (self.stateEngine, A()),
-                   (self, A())
-                   ]
-
-        # remove loggers prior to shelving
-        [swapLogger(t) for t in tmpList]
-
-        of = shelve.open(self.runPath, 'minder.db')
-        of[self.identifiedName] = None
-        of.close()
-        # reinstate  loggers after shelving
-        [swapLogger(t) for t in tmpList]
-
-# =============================================================================
-
-    def makeLog(self, loggerNameExt=''):
-        # set up a log file for this TestRun instance
-        # all logs for a pacakge go in a sub dir
-
-        logfile = os.path.join(self.logDir, self.package, self.identifiedName)
-
-        self.logger = RTTFileLogger.getLogger(self.identifiedName, logfile)
-
-    def __del__(self):
-        self.logger.info('Destructor for Minder %s' % self.identifiedName)
-
-    def unlink(self): pass
-
diff --git a/Tools/RunTimeTester/src/MinderDebugger.py.obsolete b/Tools/RunTimeTester/src/MinderDebugger.py.obsolete
deleted file mode 100755
index 8e747fac1806f71b518c544eeb188049bc8148a5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderDebugger.py.obsolete
+++ /dev/null
@@ -1,46 +0,0 @@
-import types
-    
-def wrapMethod(f, list, name):
-    def wrapped_f(object, *a, **k):
-        print '%s: self.asciiRefFiles %s' % (name, str(list))
-        result = f(object, *a, **k)
-        print '%s: self.asciiRefFiles %s' % (name, str(list))
-        return result
-    return wrapped_f
-
-def wrapMethods(dict, classname, list):
-    # print 'wrapMethods starting for %s' % classname
-    for f in dict:
-        # print f, object.__class__.__dict__[f]
-        m = dict[f]
-        if isinstance(m, types.FunctionType):
-            # print 'wrapping ', m
-            dict[f] = wrapMethod(m, list, '%s.%s'%(classname,f))
-    print 'wrapMethods ending for %s' % classname
-        
-
-
-def wrapObjectMethods(object):
-    print 'wrapObjectMethods start'
-    # print 'wrapObjectMethods 10'
-    name = object.__class__.__name__
-    # print 'wrapObjectMethods 12'
-    list = object.__dict__.get('asciiRefFiles', None)
-    # print 'wrapObjectMethods 20'
-    
-    if list:
-        print 'wrapObjectMethods: %s object has ref files' % name
-    else:
-        print 'wrapObjectMethods: %s object has no ref files' % name
-
-    # print 'wrapObjectMethods 30'
-    # object
-    dict = object.__class__.__dict__
-    # print 'object class dictionary', dict
-    wrapMethods(dict, name, list)
-
-    # and its bases (should be recursive)
-
-    for base in object.__class__.__bases__:
-        wrapMethods(base.__dict__, base.__name__, list)
-        
diff --git a/Tools/RunTimeTester/src/MinderStateEngine.py.obsolete b/Tools/RunTimeTester/src/MinderStateEngine.py.obsolete
deleted file mode 100755
index e0629201e7bcd0e82b9918f0962bf41e83cb3152..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderStateEngine.py.obsolete
+++ /dev/null
@@ -1,141 +0,0 @@
-"""State Engine for WorkerJobMinders.
-
-Transitions:
-
-queued                operatorCloseDown
-queued                running
-
-noKitError            done                  (we get a record of the problem)
-
-running               operatorCloseDown
-running               postProcessingQueued
-
-postProcessingQueued  operatorCloseDown
-postProcessingQueued  postProcessingRunning
-
-postProcessingRunning operatorCloseDown
-postProcessingRunning done
-
-
-"""
-
-from exc2string2            import exc2string2
-from MinderStateEngineBase  import MinderStateEngineBase
-
-import time
-
-class MinderStateEngine(MinderStateEngineBase):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    
-    def __init__(self, minder, state='queued'):
-        MinderStateEngineBase.__init__(self, minder, state)
-
-        self.dispatcher = {
-            'queued'               :  self.fromQueued,
-            'running'              :  self.fromRunning,
-            'operatorCloseDown'    :  self.fromOperatorCloseDown,
-            'postProcessingQueued' :  self.fromPostProcessingQueued,
-            'postProcessingRunning':  self.fromPostProcessingRunning,
-            'noKitError'           :  self.fromNoKitError,
-            'done'                 :  self.fromDone
-            }
-
-    # -----------------------------------------------
-    def functionsMinderMustHave(self):
-        return ['jobIsRunning', 'doPostProcessing', 'cleanSpace', 'runChecks', 'copyKeepFiles', 'calcPostProcessingResult',
-                'fire', 'readTestResults']
-
-    # -----------------------------------------------
-    #
-    def unlink(self):
-        ' break circular references'
-        
-        del self.minder
-    # -----------------------------------------------
-    
-    def fromQueued(self):
-        if self.fromQueuedTest():
-            self.state.set('running')
-            return
-        # otherwise remain in 'queued'
-            
-    # -----------------------------------------------
-    
-    def fromQueuedTest(self): # derived classes might override
-        return True 
-            
-    # -----------------------------------------------
-    
-    def fromRunning(self):
-        queryOK, isRunning =  self.minder.jobIsRunning()
-        
-        if isRunning:  return # no state transition
-
-        #self.fromRunningActions()
-
-        if self.minder.doPostProcessing():
-            self.state.set('postProcessingQueued')
-            return
-        
-        self.fromRunningActions() #introduced
-        self.minder.cleanSpace()
-        self.state.set('done')
-
-
-    # -----------------------------------------------    
-
-    def fromRunningActions(self):
-        'in a method to allow derived class extensions'
-
-        # self.minder.runMoniActions()
-        self.minder.runChecks()
-
-        # chainFileCopier present only if running in a chain 
-        if hasattr(self.minder, 'chainFileCopier'):
-            self.minder.chainFileCopier.copyFiles()
-
-        # if keep file is also a chain file, this code
-        # will silently fail. Intended that the chain
-        # file copier is interogated to find new location
-        self.minder.copyKeepFiles()
-        # make sure ALL files in run and results directory are
-        # accessible by all members of the UCL Atlas group
-        self.minder.calcPostProcessingResult()
-
-    # -----------------------------------------------
-    
-    def fromPostProcessingQueued(self):        
-        self.minder.fire('postProcessorScript')
-        self.state.set('postProcessingRunning')
-            
-    # -----------------------------------------------
-    
-    def fromPostProcessingRunning(self):
-        queryOK, isRunning =  self.minder.jobIsRunning()
-        if isRunning:  return # no state transition
-
-        self.minder.readTestResults()
-        self.minder.copyKeepFiles() # grab any new keep files
-        self.minder.cleanSpace()
-
-        self.state.set('done')
-        
-            
-    # -----------------------------------------------
-    
-    def fromNoKitError(self):
-        self.state.set('done')
-
-    # -----------------------------------------------
-
-    def fromDone(self):
-        #del self.minder.popen
-        #self.minder.popen = None
-        MinderStateEngineBase.fromDone(self)
-        # leave in 'done' state
-    # -----------------------------------------------
diff --git a/Tools/RunTimeTester/src/MinderStateEngineBase.py b/Tools/RunTimeTester/src/MinderStateEngineBase.py
deleted file mode 100755
index 3a56bb2e844c41c58f850bc93c82295fd260d45f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderStateEngineBase.py
+++ /dev/null
@@ -1,66 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Holds state 
-"""
-
-
-from State          import State
-from exc2string2    import exc2string2
-from RTTSException  import RTTCodingError
-from Tools          import nowSecs
-
-
-#from Tracer      import Tracer
-
-class MinderStateEngineBase:
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    def __init__(self, state, identifiedName):
-
-        self.state          = State(state)
-
-        self.pollEntryState = state
-        self.pollExitState  = state
-        self.identifiedName = identifiedName
-        self.dispatchTable  = {} # filled in by derived classes
-        
-    def poll(self):
-        # m = '%s Start %s::poll() for %s' % (nowSecs(), self.__class__.__name__, self.identifiedName)
-        # print m
-
-        self.pollEntryState  = self.state.state
-
-        try:
-            newState = self.dispatchTable[self.pollEntryState]()
-        except KeyError:
-            msg = 'State %s  not in dispatch table:\n%s' % (self.pollEntryState, str(self.dispatchTable))
-            print msg
-            raise RTTCodingError(msg)
-        except:
-            msg = 'Error in poll(): Traceback:\n%s' % exc2string2()
-            print msg
-            raise RTTCodingError(msg)
-                                                            
-
-        self.state.set(newState)
-        self.pollExitState   = self.state.state
-        transition = self.pollEntryState != self.pollExitState
-
-        # m = '%s End %s::poll() for %s %s->%s' % (nowSecs(), self.__class__.__name__, self.identifiedName, self.pollEntryState, self.pollExitState)
-        # print m
-        return  transition
-
-        
-    def closedDownByOperator(self): return self.state.state == 'operatorCloseDown'
-
-    def unlink(self): self.dispatchTable.clear()
-
-    # def isDone(self):
-    #    isDone = self.state.isDone()
-    #    # print '   %s %s::isDone %s, state %s' % (self.identifiedName, self.__class__.__name__,  str(isDone), self.state.state)
-    #    return isDone
diff --git a/Tools/RunTimeTester/src/MinderStateEngineBase.py.obsolete b/Tools/RunTimeTester/src/MinderStateEngineBase.py.obsolete
deleted file mode 100755
index e2730cc5641ecddf12a4ae662df726da8a0cf56e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderStateEngineBase.py.obsolete
+++ /dev/null
@@ -1,130 +0,0 @@
-"""
-Common poll method for all state engines
-"""
-
-
-import time
-from State          import State
-from exc2string2    import exc2string2         
-from RTTSException  import RTTCodingError
-
-from MethodTimer import MethodTimer
-
-class MinderStateEngineBase(MethodTimer):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    def __init__(self, minder, state='queued'):
-        MethodTimer.__init__(self)
-        self.state  = State(state)
-        self.minder = minder
-        self.logger = minder.logger
-
-        # operator close down is handled externally due high
-        # disk io traffic if done by the minder
-        self.operatorCloseDownFlag = False
-        self.checkMinderFunctions()
-        
-    # -----------------------------------------------
-        
-    def checkMinderFunctions(self):
-        mustHave = self.functionsMinderMustHave()
-        doesntHave = [f for f in mustHave if not hasattr(self.minder, f)]
-        if doesntHave:
-            msg = '%s does not have the following required functions %s' % (self.minder.__class__.__name__, str(doesntHave))
-            raise RTTCodingError(msg)
-        
-    # -----------------------------------------------
-        
-    def functionsMinderMustHave(self):
-        return ['setOperatorCloseDown', 'setDone']
-
-    def unlink(self):
-        del self.minder
-        
-    def poll(self):
-        """
-        makes status transitions by calling the fromXXX
-        methods, according to the current status.
-        returns True if job is active, False otherwise
-        """
-
-        # >= 1 checks are run if the job finishes.
-        # this returned in a list of statuses.
-        # the first one determines if the job ran
-        #
-        
-
-        stateOnEntry = self.state.state
-        jobName = self.minder.identifiedName
-        m = 'start poll(): Minder %s for job %s found in  state %s' % (self.__class__.__name__,
-                                                                       jobName,
-                                                                       stateOnEntry)
-        self.logger.debug(m)
-        
-        #if stateOnEntry == 'done':
-        #    return
-        
-        if self.operatorCloseDownFlag:
-            self.operatorCloseDownFlag = False
-            self.state.set('operatorCloseDown')            
-            jobName = self.minder.identifiedName
-            m = 'Forcing state transition (%s, %s): %s -> operatorCloseDown, calling back' % (self.__class__.__name__,
-                                                                                              jobName,
-                                                                                              stateOnEntry)
-            stateOnEntry = self.state.state
-            self.logger.info(m)
-
-        self.dispatcher[stateOnEntry]()
-        currentState = self.state.state
-
-        if stateOnEntry != currentState:
-
-            # make an DOM document for the minder and call back
-            # (presumably to the testRun)
-
-            jobName = self.minder.identifiedName
-            m = 'State transition detected (%s, %s): %s -> %s, calling back' % (self.__class__.__name__,
-                                                                                jobName,
-                                                                                stateOnEntry,
-                                                                                currentState)
-            self.logger.info(m)
-            startT = time.time()
-
-            try:
-                # self.minder.xmlConverter.update()
-                self.minder.xmlConverter.makeSummaryReport()
-            except:
-                m='Error on calling back to TestRun:\n%s' % exc2string2()
-                self.logger.error(m)
-                
-            diffT = time.time() - startT
-            # self.minder.sumTimeToUpdate += diffT
-            # self.minder.nUpdates      += 1
-            # m = 'Job %s completed the call back in %d seconds.' % (
-            #    self.minder.identifiedName,diffT)
-            # self.logger.info(m)
-            # recursive call  move the job through states that take no time
-            # to complete
-            self.poll()
-
-
-    def fromOperatorCloseDown(self):
-        # m = '%s: operatorCloseDown.' % self.minder.identifiedName
-        # self.logger.info(m)
-        self.minder.setOperatorCloseDown()
-        self.state.set('done')
-
-    def fromDone(self):
-        # self.minder.xmlConverter.update()
-        self.minder.xmlConverter.makeSummaryReport()
-        self.minder.setDone()
-
-        
-        
-
-    def queueOperatorCloseDown(self):
-        self.operatorCloseDownFlag = True
diff --git a/Tools/RunTimeTester/src/MinderStateEngineDummy.py.obsolete b/Tools/RunTimeTester/src/MinderStateEngineDummy.py.obsolete
deleted file mode 100644
index dace980cfdea55e4444064ab99e352c20b6d6564..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderStateEngineDummy.py.obsolete
+++ /dev/null
@@ -1,45 +0,0 @@
-
-from exc2string2            import exc2string2
-from MinderStateEngineBase  import MinderStateEngineBase
-import time
-
-class MinderStateEngineDummy(MinderStateEngineBase):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    
-    def __init__(self, minder, state='queued'):
-        MinderStateEngineBase.__init__(self, minder, state)
-
-        self.dispatcher = {
-            'queued'               :  self.fromQueued,
-            'running'              :  self.fromRunning,
-            'done'                 :  self.fromDone
-            }
-        self.nPollInRunning   = 0 
-        self.maxPollInRunning = 2 
-
-    # -----------------------------------------------
-
-    def functionsMinderMustHave(self): return []
-    # -----------------------------------------------
-    
-    def fromQueued(self):
-        self.state.set('running')
-        return
-            
-    # -----------------------------------------------
-    
-    def fromRunning(self):
-        if self.nPollInRunning >= self.maxPollInRunning:
-            self.minder.processingResult = 'success = set by dummy state engine'
-            self.state.set('done')
-            return
-
-        self.nPollInRunning += 1
-
-    def fromDone(self):
-        MinderStateEngineBase.fromDone(self)
diff --git a/Tools/RunTimeTester/src/MinderToString.py.obsolete b/Tools/RunTimeTester/src/MinderToString.py.obsolete
deleted file mode 100644
index 003bd19e3a66d8e1dd78dff06d79b0c810105d8f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderToString.py.obsolete
+++ /dev/null
@@ -1,74 +0,0 @@
-class MinderToStr:
-    def __init__(self, offset=''):
-        self.offset = offset
-
-    def toStr(self, m):
-        mname =  m.__class__.__name__
-
-        s  = '\n------------------------------\n'
-        try:
-            s += '%s%s:%s \n' % (self.offset, mname, m.identifiedName)
-        except:
-            s += '%s%s \n' % (self.offset,mname)
-            self.offset += '   '
-        mts = MinderToStr(self.offset+'   ')
-        if mname == "ParallelMinder":
-            s+=self.offset+'Minders in minder list: %s' % str(m.minders)
-            mts = MinderToStr(self.offset+'   ')
-            s += self.offset+'Listing the %d minders in to process list\n' % len(m.minders)
-            for sm in m.minders:
-                s += self.offset+('Minders in to process list\n')
-                s+=self.offset+mts.toStr(sm)
-        elif mname == 'SequentialMinder':
-            s += self.offset+'currentMinder: %s\n' % str(m.currentMinder.__class__.__name__)
-            mts = MinderToStr(self.offset+'   ')
-            s += mts.toStr(m.currentMinder)+'\n'
-            s += self.offset+'Listing the %d minders in to process list\n' % len(m.minders)
-            for sm in m.minders:
-                s += self.offset+('Minder in to process list\n')
-                s+=self.offset+mts.toStr(sm)
-        elif mname == 'ChainJobMinder':
-            mts = MinderToStr(self.offset+'   ')
-            s += self.offset+'Contained Minder:'
-            s += mts.toStr(m.minder)+'\n'
-            if hasattr(m,'outChainFiles'):
-                for k in m.outChainFiles:
-                    s += self.offset + 'Chain file out\n'
-                    s += self.offset + '%s\n' % k
-
-            else: 
-                s += self.offset+'No outout chain files\n'
-                if m.fileCopier:
-                    s += self.offset + str(m.fileCopier)
-                else:
-                    s += self.offset + 'NotYetKnown'
-        elif mname == 'PseudoMinder':
-            if hasattr(m, 'datasets'):
-                datasets = m.datasets
-                for k in datasets.keys():
-                    s += self.offset + 'Dataset type %s\n' % k
-                    for d in datasets[k]:
-                        s += self.offset + 'Dataset %s\n' % str(d)
-            else:
-                s += self.offset+'No datasets\n'
-        else:
-            s += self.offset+m.__class__.__name__
-        return s
-
-def dumpMinders(minders):
-    mTos = MinderToStr()
-    for m in minders:
-        print '\n=====================================================\n'
-        print 'Minder %s:%s' % (m.__class__.__name__, m.identifiedName)
-        print mTos.toStr(m)
-        print '\n=====================================================\n'
-
-def dumpMindersToString(minders):
-    mTos = MinderToStr()
-    s = ''
-    for m in minders:
-        s += '\n=====================================================\n'
-        s += 'Minder %s:%s' % (m.__class__.__name__, m.identifiedName)
-        s +=  mTos.toStr(m)
-        s += '\n=====================================================\n'
-        return s
diff --git a/Tools/RunTimeTester/src/MinderToXML2.py b/Tools/RunTimeTester/src/MinderToXML2.py
deleted file mode 100755
index 18894dd8b26b87100d29c1d00c8cff711b7cd055..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinderToXML2.py
+++ /dev/null
@@ -1,393 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an xml file summary of an RTT run
-"""
-from StringIO      import StringIO
-from makeRTTDOM    import makeRTTDOM
-from xml.dom.ext   import PrettyPrint
-from Defs          import RTTdefs
-from exc2string2   import exc2string2
-from MethodTimer   import MethodTimer
-
-import sys
-import os
-import fnmatch
-
-class MinderToXML(MethodTimer):
-    
-    def __init__(self, minder):
-
-        MethodTimer.__init__(self)
-        
-        self.minder          = minder
-        self.logger          = minder.logger
-        self.elementCreator  = minder.elementCreator
-        self.textNodeCreator = minder.textNodeCreator
-        self.callBack        = None
-        self.minderXMLNode   = None
-        
-    def unlink(self):
-        del self.minder
-        if self.callBack: del self.callBack
-        
-    def update(self):
-
-        minder = self.minder # for nested methods
-        # nested scopes are used to pass parameters to these nested functions
-        def findErrorMessages():
-
-            element = self.elementCreator('errors')
-            
-            text = reduce(lambda x, y: x+y+'\n',
-                          minder.errorMessages,
-                          '')
-            if text:
-                tn = self.textNodeCreator(text)
-                element.appendChild(tn)
-
-            return element
-
-        def findDatasets():
-
-            dsElement = self.elementCreator('datasets')
-            
-            try:
-                dataSets = minder.datasets
-            except:
-                dataSets = {}
-
-            datasetInstances = []
-            [datasetInstances.extend(dataSets[k]) for k in dataSets]
-            #for ds in datasetInstances:
-            #    tn = self.textNodeCreator(os.path.basename(ds.physical.strip()))
-            #    el = self.elementCreator('dataset')
-            #    el.appendChild(tn)
-            #    dsElement.appendChild(el)
-
-            for ds in datasetInstances:
-                dsInfoEl = self.elementCreator('datasetinfo')
-
-                head, tail = os.path.split(ds.physical)
-                
-                tn = self.textNodeCreator(tail.strip())
-                el = self.elementCreator('file')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(head.strip())
-                el = self.elementCreator('directory')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.mtime).strip())
-                el = self.elementCreator('mtime')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.size).strip())
-                el = self.elementCreator('size')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-
-                tn = self.textNodeCreator(str(ds.origin).strip())
-                el = self.elementCreator('origin')
-                el.appendChild(tn)
-                dsInfoEl.appendChild(el)
-                                
-                dsElement.appendChild(dsInfoEl)
-
-            return dsElement
-
-        def findMissingDatasets():
-
-            dsElement = self.elementCreator('missingDatasets')
-            
-            try:
-                datasetInstances = minder.missingDatasets
-            except:
-                datasetInstances = []
-
-            for ds in datasetInstances:
-                tn = self.textNodeCreator(os.path.basename(ds.physical.strip()))
-                el = self.elementCreator('missingDataset')
-                el.appendChild(tn)
-                dsElement.appendChild(el)
-
-            return dsElement
-
-        def getArchiveInfo():
-            archiversEl = self.elementCreator('archivers')
-            for archiver in minder.archivers:
-                archiverEl = self.elementCreator('archiver')
-                typeEl = self.elementCreator('archivertype')
-                typeEl.appendChild(self.textNodeCreator(archiver.type))
-                archiverEl.appendChild(typeEl)
-
-                # Now the files
-                files = archiver.copiedOK + archiver.notCopiedOK
-                if not files: continue # no files matched for archiving
-                    
-                archivedFilesEl = self.elementCreator('archivedfiles')
-                for (copiedOK, from_, to_) in files:
-                    archivedFileEl = self.elementCreator('archivedfile')
-                    
-                    okEl   = self.elementCreator('archivedOK')
-                    okEl.appendChild(self.textNodeCreator(str(copiedOK)))
-                    fromEl = self.elementCreator('archiveSrc')
-                    fromEl.appendChild(self.textNodeCreator(os.path.basename(from_)))
-                    toEl   = self.elementCreator('archiveDst')
-                    toEl.appendChild(self.textNodeCreator(to_))
-
-                    archivedFileEl.appendChild(okEl)
-                    archivedFileEl.appendChild(fromEl)
-                    archivedFileEl.appendChild(toEl)
-                    archivedFilesEl.appendChild(archivedFileEl)
-
-                archiverEl.appendChild(archivedFilesEl)
-                archiversEl.appendChild(archiverEl)
-            return archiversEl    
-            
-        def findNtestSuccess():
-            l = len([s for s in minder.finishedTests if s.result == 0])
-            tn = self.textNodeCreator(str(l))
-            el = self.elementCreator('sUserTests')
-            el.appendChild(tn)
-            return el
-
-        def diskSpaceUsedByJob():
-            dsEl = self.elementCreator('diskspace')
-
-            runpEl = self.elementCreator('runpath')
-            for value, when, phase in minder.diskspace.run:
-                datumEl = self.elementCreator('datum')
-                datumEl.appendChild(makeTextNode(('time', when)))
-                datumEl.appendChild(makeTextNode(('value', value)))
-                datumEl.appendChild(makeTextNode(('phase', phase)))
-                runpEl.appendChild(datumEl)
-                
-            respEl = self.elementCreator('respath')
-            for value, when, phase in minder.diskspace.res:
-                datumEl = self.elementCreator('datum')
-                datumEl.appendChild(makeTextNode(('time', when)))
-                datumEl.appendChild(makeTextNode(('value', value)))
-                datumEl.appendChild(makeTextNode(('phase', phase)))
-                respEl.appendChild(datumEl)
-
-            dsEl.appendChild(runpEl)
-            dsEl.appendChild(respEl)
-            return dsEl
-
-        def findKeepFiles():
-
-            resPath = minder.resPath
-            
-            # minder.keepFilesPatterns is a list of dictionaries.
-            # kFilePatterns = [k['keepFileString']
-            #                 for k in minder.keepFilePatterns]
-            # minder.keepFiles is a dictionary of dictionaries.
-
-            results = []
-
-            for k,v in minder.keepFiles.items():
-                default = ''
-                name  = v.get('keepFileString', default)
-                info  = v.get('infoString', default)
-                color = v.get('displayColor', default)
-                md5   = v.get('md5sum', default)
-
-                lookIn  = os.path.dirname(name)
-                lookFor = os.path.basename(name)
-                if not os.path.exists(lookIn): continue
-
-                matches = fnmatch.filter(os.listdir(lookIn), lookFor)
-
-                result = [{'keepFileName' : os.path.join(lookIn, m), 'infoString' : info, 'displayColor' : color, 'md5sum' : md5} for m in matches]
-                results.extend(result)
-
-            return results
-                
-        def makeKeepFileNode():
-            # logger.debug('Going to make element deliveredkeepfiles, document: ' +str( document))
-            el = self.elementCreator('deliveredkeepfiles')
-
-            knownKeepFiles = findKeepFiles()
-            if knownKeepFiles:
-                baseEl = self.elementCreator('baseKeepFileDir')
-                dirRelToResBase = minder.resPath.split(minder.resultsBasePath)[1]
-                if dirRelToResBase.startswith('/'): dirRelToResBase = dirRelToResBase[1:]
-                if not dirRelToResBase.endswith('/'): dirRelToResBase + '/'
-                baseEl.appendChild(self.textNodeCreator(dirRelToResBase))
-                el.appendChild(baseEl)
-                
-            for k in knownKeepFiles:
-                name  = k['keepFileName']
-                info  = k['infoString']
-                color = k['displayColor']
-                md5   = k['md5sum']
-                self.logger.debug('Keepfile dict: %s' % str(k))
-
-                def keepFileNode():
-                    # def keepFileNode(style):
-                    # if style=='new_style':
-                    el2 = self.elementCreator('keepfile_newstyle')
-                    resBase = minder.resultsBasePath                    
-                    keepFileRelativeToResBase = name.strip().split(resBase)[1]
-                    if keepFileRelativeToResBase.startswith('/'): keepFileRelativeToResBase = keepFileRelativeToResBase[1:]
-                    tn = self.textNodeCreator(keepFileRelativeToResBase)
-                    el2.setAttribute('info', info)
-                    el2.setAttribute('displayColor', color)
-                    el2.setAttribute('md5sum', md5)
-                    el2.appendChild(tn)
-
-                    # if style=='old_style':
-                    #   el2 = self.elementCreator('keepfile')
-                    # tn = self.textNodeCreator(os.path.basename(str(name).strip()))
-                    #    el2.setAttribute('info',info)
-                    #   el2.setAttribute('displayColor',color)
-                    #    el2.appendChild(tn)
-                    return el2
-
-                el.appendChild(keepFileNode())
-                # el.appendChild(keepFileNode('old_style'))
-
-            return el
-            
-        def makeTextNode((tagName, text)):
-            tn = self.textNodeCreator(str(text).strip())
-            el = self.elementCreator(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(document))
-        
-        mElement = self.elementCreator('minder')
-        textInfo = {
-            'identifiedName':  self.minder.identifiedName,
-            'jobName':         self.minder.name,
-            'jobDisplayName':  str(self.minder.jobDisplayName),
-            'jobID':           self.minder.jobSerialNumber,
-            'state':           self.minder.stateEngine.state.state,
-            'status':          self.minder.status(),
-            'jobGroup':        self.minder.jobGroup,
-            'nUserTests':      self.minder.numberOfTests(),
-            'nUserActions':    self.minder.numberOfActions(),
-            'resultsPath':     self.minder.resPath,
-            'exitStatus':      self.minder.exitStatus,
-            'cpuTime':         self.minder.cpuTime,
-            'cpuTime2000':     self.minder.cpuTime2000,
-            'batchStatus':     self.minder.batchStatus,
-            'mem':             self.minder.mem,
-            'vmem':            self.minder.vmem,
-            'wallTime':        self.minder.wallTime,
-            'nRetries':        self.minder.errorStateCurRetry,
-            # 'nUpdates':        self.minder.nUpdates,
-            # 'sumTimeToUpdate': self.minder.sumTimeToUpdate,
-            'stateHistory':         str(self.minder.stateEngine.state.history),
-            'postProcessingResult': self.minder.postProcessingResult,
-            'processingResult':     self.minder.processingResult,
-            'chainLength':          str(self.minder.chainLength),
-            'chainSuccesses':       str(self.minder.chainSuccesses),
-            'jobDocString':    str(self.minder.jobDocString),
-            'jobDocURL':    str(self.minder.jobDocURL)
-            }
-
-            
-        textInfo.update(self.minder.descDataForXMLNode)
-
-        
-        for key,val in textInfo.items():
-            if val.__class__.__name__ in ['list', 'tuple']: # 1-to-many tag/value mapping
-                [mElement.appendChild(makeTextNode((key,v))) for v in val]
-            else: #1-to-1 tag/value mapping 
-                mElement.appendChild(makeTextNode((key,val)))
-            
-        mElement.appendChild(findDatasets())        
-        mElement.appendChild(findMissingDatasets())
-        mElement.appendChild(getArchiveInfo())
-        mElement.appendChild(diskSpaceUsedByJob())
-        mElement.appendChild(findNtestSuccess())
-        mElement.appendChild(makeKeepFileNode())
-        mElement.appendChild(findErrorMessages())          
-        
-        def makeFlagNode((tagName, flag)):
-            if flag: return self.elementCreator(str(tagName))
-            return None
-
-        flagInfo = [
-            ('rttPilotJob',   self.minder.rttPilotJob),
-            ('rttATNJob',     self.minder.rttATNJob)
-            ]
-
-        for f in flagInfo:
-            node = makeFlagNode(f)
-            if node: mElement.appendChild(node)
-
-        if self.minder.isDone():
-            mElement.appendChild(self.elementCreator('done'))
-
-        if self.minder.closedDownByOperator():
-            mElement.appendChild(self.elementCreator('operatorCloseDown'))
-
-
-        el = self.elementCreator('checkResults')
-        [el.appendChild(makeTextNode( ('checkResult', r) ))
-         for r in self.minder.checkResults]
-        mElement.appendChild(el)
-
-        el = self.elementCreator('testResults')
-        for r in self.minder.finishedTests:
-
-            inner = self.elementCreator('testResult')
-            n = makeTextNode( ('testName', r.testId) )
-            inner.appendChild(n)
-            if r.error:
-                n = makeTextNode( ('testStatus', 'error') )
-            else:
-                n = makeTextNode( ('testStatus', str(r.result) ) )
-            inner.appendChild(n)
-
-            el.appendChild(inner)
-
-        mElement.appendChild(el)
-        
-        self.outputToFile(mElement)
-
-        self.minderXMLNode = mElement
-
-
-        m = 'Created, minder node for %s' % self.minder.identifiedName
-        self.logger.debug(m)
-
-    def makeSummaryReport(self):
-        if not self.callBack:
-            self.logger.error('No object for Minder to call back to, abandoning update')
-            return 
-        self.update()
-        assert (self.minderXMLNode != None)
-        dict = self.minder.dataForMonitoring()
-        self.callBack.update(dict, self.minderXMLNode)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.update(), temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-
-    def outputToFile(self, minderNode):
-        space = ' '
-        f = os.path.join(self.minder.resPath, 'rttjobinfo.xml')
-        h = None
-        try:
-            h = open(f, 'w')
-            h.write(minderNode.toprettyxml(3*space, '\n'))
-            h.close()
-        except:
-            self.logger.error('Unable to output minder info XML file. Traceback: \n%s' % exc2string2())
-        if h: h.close()    
-
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/MinimalXMLConverter.py.obsolete b/Tools/RunTimeTester/src/MinimalXMLConverter.py.obsolete
deleted file mode 100644
index bc4d3844da120998200f17b5d961d73b8a49abf0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/MinimalXMLConverter.py.obsolete
+++ /dev/null
@@ -1,105 +0,0 @@
-"""
-A class to provide an xml file summary of an RTT run
-"""
-import sys, os, fnmatch, os.path
-from StringIO      import StringIO
-from makeRTTDOM    import makeRTTDOM
-from xml.dom.ext   import PrettyPrint
-from Defs          import RTTdefs
-from exc2string2   import exc2string2
-from MethodTimer   import MethodTimer
-
-class MinimalXMLConverter(MethodTimer):
-    
-    def __init__(self, minder):
-
-        MethodTimer.__init__(self)
-        
-        self.minder          = minder
-        self.logger          = minder.logger
-        self.elementCreator  = minder.elementCreator
-        self.textNodeCreator = minder.textNodeCreator
-        self.callBack        = None
-
-    def unlink(self):
-        del self.minder
-        if self.callBack: del self.callBack
-        
-    def update(self):
-
-        if not self.callBack:
-            self.logger.error('No object for Minder to call back to, abandoning update')
-            return 
-        minder = self.minder # for nested methods
-        # nested scopes are used to pass parameters to these nested functions
-        def findErrorMessages():
-
-            element = self.elementCreator('errors')
-            
-            text = reduce(lambda x, y: x+y+'\n',
-                          minder.errorMessages,
-                          '')
-            if text:
-                tn = self.textNodeCreator(text)
-                element.appendChild(tn)
-
-            return element
-
-        def makeTextNode((tagName, text)):
-            tn = self.textNodeCreator(str(text).strip())
-            el = self.elementCreator(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(document))
-        
-        mElement = self.elementCreator('minder')
-        textInfo = {
-            'identifiedName':  self.minder.identifiedName,
-            'jobID':           self.minder.jobSerialNumber,
-            'state':           self.minder.stateEngine.state.state,
-            'status':          self.minder.status(),
-            'jobGroup':        self.minder.jobGroup,
-            'stateHistory':    str(self.minder.stateEngine.state.history),
-            }
-
-            
-        textInfo.update(self.minder.descDataForXMLNode)
-
-        
-        for key,val in textInfo.items():
-            if val.__class__.__name__ in ['list', 'tuple']: # 1-to-many tag/value mapping
-                [mElement.appendChild(makeTextNode((key,v))) for v in val]
-            else: #1-to-1 tag/value mapping 
-                mElement.appendChild(makeTextNode((key,val)))
-            
-        self.outputToFile(mElement)
-
-        m = 'Created, minder node for %s' % self.minder.identifiedName
-        self.logger.debug(m)
-        dict = self.minder.dataForMonitoring()
-        self.callBack.update(dict, mElement)
-
-    def __str__(self):
-        temp = StringIO() 
-        PrettyPrint(self.update(), temp)
-        str = temp.getvalue()
-        temp.close()
-        return str
-
-    def outputToFile(self, minderNode):
-        space = ' '
-        f = os.path.join(self.minder.resPath, 'rttjobinfo.xml')
-        h = None
-        try:
-            h = open(f, 'w')
-            h.write(minderNode.toprettyxml(3*space, '\n'))
-            h.close()
-        except:
-            self.logger.error('Unable to output minder info XML file. Traceback: \n%s' % exc2string2())
-        if h: h.close()    
-
-    def setSummaryCallBack(self, newCallBack):
-        self.logger.debug('setting a callback of type %s' % (
-            newCallBack.__class__.__name__))
-        self.callBack = newCallBack
diff --git a/Tools/RunTimeTester/src/ModuleLoader.py b/Tools/RunTimeTester/src/ModuleLoader.py
deleted file mode 100755
index f032e37d030cd9d92cafaa74c256e2536cdfdb6d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ModuleLoader.py
+++ /dev/null
@@ -1,92 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Class that loads an external module A.
-
-Calling
-A.getClassInstance(classname, parameters)
-will return
-classname(parameters)
-for classname in A
-
-
-Module name is without the .py extension.
-
-"""
-#------------------------------------------------
-from exc2string2              import exc2string2
-import imp, sys, os.path
-
-import logging
-logger = logging.getLogger('rtt')
-
-#------------------------------------------------
-class ModuleLoader:
-
-    def __init__(self, modulename, pathToModule, logger):
-        
-        self.modulename   = modulename
-        self.pathToModule = pathToModule
-        self.logger       = logger
-
-    def getClassInstance(self, classname, paramDict={}):
-        """paramDict is a dictionary of argname-argvalue pairs
-        that will be passed to the class constructor.
-        See __main__ example below."""
-
-        
-        try:
-            file,pathname,description = imp.find_module(self.modulename,
-                                                        [self.pathToModule])
-        except Exception, e:
-            m = 'Failed to find module: \n%s %s' %(self.modulename,
-                                                  self.pathToModule)
-            self.logger.warning (m)
-            self.logger.warning(exc2string2())
-            self.logger.warning(str(e))
-            return None
-
-        try:
-            klass                     = imp.load_module(classname,
-                                                        file,
-                                                        pathname,
-                                                        description)
-        except Exception, e:
-            m  = 'Failed to load module:'
-            m += '\n  class %s '         % classname
-            m += '\n  file %s '          % file.name
-            m += '\n  pathname %s '      % pathname
-            if not os.path.exists(os.path.join(pathname, file.name)):
-                m += '\npath to module does not exist'
-            else:
-                m += '\npath to module ok. Ensure class is present in module'
-                                  
-                                  
-            self.logger.warning(m)
-            self.logger.warning(exc2string2())
-            self.logger.warning(str(e))
-            return None
-
-        if classname not in vars(klass).keys():
-            m  = 'Failed to instantiate class:'
-            m += '\n  class %s '         % classname
-            m += '\n  from pathname %s \n'      % pathname
-
-            self.logger.warning(m)
-            return None
-        
-        return((vars(klass)[classname])(paramDict))  #instantiate class with paramDict
-
-
-if __name__=="__main__":
-    class Logger:
-        def debug(self, s): print s
-        def warning(self, s): print s
-        def error(self, s): print s
-
-    paramDict={'SimpleTest_key': 'SimpleTest_value'}
-    ml = ModuleLoader('AtlasEnvSetup', '../share', Logger())
-    pp = ml.getClassInstance('AtlasEnvSetup', paramDict)
-
-    
-        
diff --git a/Tools/RunTimeTester/src/NICOS.py b/Tools/RunTimeTester/src/NICOS.py
deleted file mode 100755
index 3c3ab1f0e7dfbc8f1b3823c2b695c4c9d259f840..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/NICOS.py
+++ /dev/null
@@ -1,18 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-from Tools2                  import getFileOrDirDateStamp
-from RTTSException           import RTTCodingError
-
-#----------------------------------------------------
-class InValidNICOSDateError(Exception): pass
-def getReleaseDate(pathNameBuilder):
-    """For a given release, get the stamp date from the directory on AFS,
-    or, for nightlies, from the link date if it is today's release."""
-    
-    fileWithDate = str(pathNameBuilder.pathToFindReleaseDate())
-    date = getFileOrDirDateStamp(fileWithDate)
-    if not date.isValid():
-        m = 'InValidNICOSDateError %s' % str(date)
-        raise RTTCodingError(m)
-    return date
diff --git a/Tools/RunTimeTester/src/NameAndStatus.py.obsolete b/Tools/RunTimeTester/src/NameAndStatus.py.obsolete
deleted file mode 100755
index 897df8582b132b00580b6cd728749bc55a54b8cd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/NameAndStatus.py.obsolete
+++ /dev/null
@@ -1,25 +0,0 @@
-#!/usr/local/bin/python
-class NameAndStatus:
-    def __init__(self, name, status):
-        self.name = name
-        self.status = status
-    def __getinit__(self):
-        return (self.name, self.status)
-    def dump(self):
-        print self.dumps()
-
-    def dumps(self):
-        s=''
-        s+='-------------------------\n'
-        s+='name   = '+str(self.name)+'\n'
-        s+='status = '+self.status+'\n'
-
-        return s
-
-class MonitorFood:
-    def __init__(self, msg, jobs):
-        self.msg  = msg
-        self.jobs = jobs
-
-    def __getinit__(self):
-        return(self.msg, self.jobs)
diff --git a/Tools/RunTimeTester/src/Null.py.obsolete b/Tools/RunTimeTester/src/Null.py.obsolete
deleted file mode 100755
index 6930d57247c923e7b34b3927c0d7041cc2312251..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Null.py.obsolete
+++ /dev/null
@@ -1,37 +0,0 @@
-class Null:
-    "Null object pattern See Python Cookbook - A class version of None"
-
-    def __init__(self, *args, **kwargs):pass
-    def __call__(self, *args, **kwargs): return self
-    def __repr__(self): return 'Null()'
-    def __str__(self): return 'Null()'
-    def __nonzero__(self): return 0
-
-    def __getattr__(self, name):
-        print '__getattr__(), '+name
-        return self
-    def __setattr__(self, name, value): return self
-    def __delattr__(self, name): return self
-
-    def __eq__(self, other):
-        if other == None: return True
-        if other.__class__ == self.__class__: return True
-        return False
-
-    
-    
-if __name__=='__main__':
-    n1 = Null()
-    n2 = Null()
-
-    print n1==n2
-
-    class A:
-        def __init__(self): self.i=1
-
-    a=A()
-    print n1 == a
-
-    print n1 == 1
-
-    print n1.group
diff --git a/Tools/RunTimeTester/src/NullXMLConverter.py b/Tools/RunTimeTester/src/NullXMLConverter.py
deleted file mode 100644
index 6f57d08fc60cd1f99a1e85bbdef967eeb63885d4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/NullXMLConverter.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-class NullXMLConverter:
-    def __init__(self, minder):
-        self.logger = minder.logger
-
-    def update(self):
-        self.logger.debug("NullXMLConverter::update")
-
-    def setSummaryCallBack(self, callBackFn):
-        self.callBack = callBackFn
diff --git a/Tools/RunTimeTester/src/PBSMinderStateEngine.py.obsolete b/Tools/RunTimeTester/src/PBSMinderStateEngine.py.obsolete
deleted file mode 100755
index 0af3fc491eac445d9c7e902045343adbee21f71c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PBSMinderStateEngine.py.obsolete
+++ /dev/null
@@ -1,123 +0,0 @@
-"""State Engine for PBSJobMinders: extends WorkerJobMinderStateEngine
-
-overrides: fromQueued
-           fromQueuedTest
-           fromRunning
-
-removes transitions
-
-
-adds state transitions
-
-queued   -> pbsError
-running  -> retry
-retry    -> queued
-pbsError -> done
-"""
-
-from WorkerMinderStateEngine  import WorkerMinderStateEngine
-
-class PBSMinderStateEngine(WorkerMinderStateEngine):
-    'Overrides for PBS'
-    
-    def __init__(self, pbsMinder, state='queued'):
-        WorkerMinderStateEngine.__init__(self, pbsMinder, state)
-        newDispatcher = {
-            'retry':      self.fromRetry,
-            'pbsError':   self.fromPBSError,
-            'timedOut':   self.fromTimedOut
-            }
-        self.dispatcher.update(newDispatcher)
-    # -----------------------------------------------
-
-    def fromQueuedTest(self): assert(False) # base method refused
-            
-    # -----------------------------------------------
-    
-    def fromQueued(self):
-
-
-        self.minder.fire('runScript')
-                    
-        if not self.minder.checkPBSSubmissionWasOK():
-            self.state.set('pbsError')
-            return
-
-
-        self.state.set('running')
-        # othewise stay in 'queued'
-
-    # -----------------------------------------------
-
-    def fromRunning(self):
-        "Running is complete if the log files have returned"
-        
-        # query if job is in batch queue
-        queryOK, inBatch = self.minder.jobIsRunning()
-
-        # give up if the query failed
-        if not queryOK:
-            self.state.set('pbsError')
-            return
-
-        # if the job is still running, no state transition.
-        if inBatch: return
-
-        # are the log files visible?
-        if not self.minder.areLogFilesInRunDir():
-            if self.minder.checkForTimeOut():
-                self.state.set('timedOut')
-                return
-            # waiting for log files to return: no state change
-            return
-
-        self.fromRunningActions() # must perform log checks before retry test.
-
-        if self.minder.retryTest():
-            self.state.set('retry')
-            return
-
-        if self.minder.doPostProcessing():
-            self.state.set('postProcessingQueued')
-            return
-
-        self.state.set('done')
-        
-    # -------------------------------------------------------------------------
-
-    def fromRunningActions(self):
-        self.minder.audit()
-        WorkerMinderStateEngine.fromRunningActions(self)
-        
-        
-    # -------------------------------------------------------------------------
-    
-    def fromRetry(self):
-
-        # the presence of the log files signals the end of the running
-        # phase - remove the files,and ensure they are copied to the
-        # results web pages.
-
-        self.minder.prepareForRetry()
-        self.state.set('queued')
-        
-    # -----------------------------------------------
-        
-    def fromTimedOut(self):
-        'A marker state'
-
-        if self.minder.doPostProcessing():
-            self.state.set('postProcessingQueued')
-            return
-
-        self.state.set('done')
-        
-        
-    # -----------------------------------------------
-        
-    def fromPBSError(self):
-        self.state.set('done')
-        
-    def fromOperatorCloseDown(self):
-        self.minder.stopJob()
-        WorkerMinderStateEngine.fromOperatorCloseDown(self)
diff --git a/Tools/RunTimeTester/src/PPPaths.py b/Tools/RunTimeTester/src/PPPaths.py
deleted file mode 100644
index 2c6eb7237338f23835efca3ba627cfb1b5c46921..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PPPaths.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-class PPPaths: pass
diff --git a/Tools/RunTimeTester/src/PackageMailer.py b/Tools/RunTimeTester/src/PackageMailer.py
deleted file mode 100644
index 87c2d3bdef952d0dbc600e6ccdb63fd90afa2ad6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PackageMailer.py
+++ /dev/null
@@ -1,180 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools                    import sendMail
-from Tools2                   import deSpamifyEmailAddress
-import os.path
-
-class PackageMailer:
-
-    def __init__(self, packageName, mailTo, release, runType, originalBranch,
-                 cmtConfig, pathSegment, packageTag, resultsDirs,
-                 resultsBaseURL, confFile, disabled, logger):
-        self.name            = packageName
-        self.mailTo          = mailTo
-        self.release         = release
-        self.runType         = runType
-        self.originalBranch  = originalBranch
-        self.cmtConfig       = cmtConfig
-        self.pathSegment     = pathSegment
-        self.packageTag      = packageTag
-        self.resultsDirs     = resultsDirs
-        self.resultsBaseURL  = resultsBaseURL
-        self.confFile        = confFile
-        self.disabled        = disabled
-
-        self.logger          = logger
-        self.emailStats      = {}
-        
-    def sendMailToPackageDevelopers(self):
-
-        if self.disabled: return
-
-        if not self.packageRequestsMailOnFailure():
-            self.logger.debug('Package did not request mail via <mailto> tag. None sent.')
-            return
-
-        if not self.emailStats.keys(): return
-
-        emailAddresses = deSpamifyEmailAddress(self.mailTo)
-        
-        self.logger.debug('Package requests mail be sent to: %s' % str(emailAddresses))
-
-        release  = self.release
-        runType  = self.runType
-        branch   = self.originalBranch
-        platform = self.cmtConfig
-        majMinProj  = self.pathSegment
-        packageTag = self.packageTag
-        
-        packageResDir = os.path.join(self.resultsDirs['packages'], self.name)
-        baseResURL    = self.resultsBaseURL
-
-        failedJobs = [k for k,v in self.emailStats.items() if not v]
-        okJobsFailedTests = [(k,v) for k,v in self.emailStats.items() if v]
-
-        if not okJobsFailedTests:
-            text = '%d job(s) failed' % len(failedJobs)
-        elif not failedJobs:
-            text = '%d successful job(s) had one or more failed test' % len(okJobsFailedTests)
-        else:
-            text = '%d job(s) failed and %d successful jobs had failed tests' % (len(failedJobs), len(okJobsFailedTests))
-        
-        message  = '**************************************************************\n'
-        message += 'PLEASE DO NOT REPLY TO THIS EMAIL ADDRESS.\n'
-        message += '**************************************************************\n\n'
-        message += 'Dear RTT user,\n\n'
-        message += 'The RTT is sending you this automated email because %s\n' % text
-        message += 'in package %s and, in such cases, you requested to be email alerted.\n' % self.name
-        message +=  '--- you did this using the <mailto> XML tag in your package configuration file:\n\n'
-        message += str(self.confFile) + '\n\n'
-        message += 'To stop future emails, simply remove the tag.\n\n'
-        message += 'Questions? Contact: rtt@hep.ucl.ac.uk\n\n'
-        message += '----------------\n'
-        message += 'Run      : %s/%s/%s\n' % (release, branch, runType)
-        message += 'Platform : %s\n' % platform
-        message += 'Package  : %s\n\n' % self.name
-        message += 'CVS tag  : %s\n\n' % packageTag
-        message += 'Number of jobs that failed: %d\n' % len(failedJobs)
-        message += 'Number of successful jobs but with >=1 failed tests: %d\n' % len(okJobsFailedTests)
-        message += '----------------\n\n'
-
-        index = 1
-
-
-        if failedJobs:
-            message += '**** JOBS THAT FAILED ****\n\n'
-            for (jobName, jobID, jobGroup) in failedJobs:
-                m  = 'Job #%s\n' % str(index)
-                m += '   - Name  : %s\n' % jobName
-                m += '   - Group : %s\n' %  jobGroup
-                m += '   - AFS   : %s\n' % os.path.join(packageResDir, jobGroup, jobName, jobID)
-                jobURL = os.path.join(baseResURL, 'page2.php')
-                jobURL += '?xml=%s/%s/%s/%s/%s/RTTSummary.xml&package=%s&job=%s&id=%s' % (release,
-                                                                                          branch,
-                                                                                          runType,
-                                                                                          platform,
-                                                                                          majMinProj,
-                                                                                          self.name,
-                                                                                          jobName,
-                                                                                          jobID)
-                m += '   - URL   : %s\n\n' % jobURL
-                message += m
-                index   += 1
-
-        if okJobsFailedTests:    
-            message += '**** JOBS THAT SUCCEEDED BUT THAT HAD FAILED TESTS ****\n\n'
-            for (jobName, jobID, jobGroup), failedTests in okJobsFailedTests:
-                m  = 'Job #%s\n' % str(index)
-                m += '   - Name  : %s\n' % jobName
-                m += '   - Group : %s\n' %  jobGroup
-                m += '   - AFS   : %s\n' % os.path.join(packageResDir, jobGroup, jobName, jobID)
-                jobURL = os.path.join(baseResURL, 'page2.php')
-                jobURL += '?xml=%s/%s/%s/%s/%s/RTTSummary.xml&package=%s&job=%s&id=%s' % (release,
-                                                                                          branch,
-                                                                                          runType,
-                                                                                          platform,
-                                                                                          majMinProj,
-                                                                                          self.name,
-                                                                                          jobName,
-                                                                                          jobID)
-                m += '   - URL   : %s\n\n' % jobURL
-
-                m += '   - Tests that failed in this job:\n'
-                for ft in failedTests: # ft = RTTTestRunner instance
-                    m += '\t%s\n' % ft
-                m += '\n'
-                message += m
-                index   += 1
-        message += '--------------------------\n'    
-        message += 'AFS base results dir for your entire package:\n'
-        message += '   %s\n' % packageResDir
-        message += 'URL for your entire package:\n'
-        message += '   %s?xml=%s/%s/%s/%s/%s/RTTSummary.xml&package=%s' % (os.path.join(baseResURL, 'page1.php'),
-                                                                           release,
-                                                                           branch,
-                                                                           runType,
-                                                                           platform,
-                                                                           majMinProj,
-                                                                           self.name)
-            
-        message += '\n\nRegards,\n'
-        message += '     The RTT team\n\n'
-        message += '--------------------------------------------------------\n'
-        message += 'RTT results: http://cern.ch/atlasrtt/page0.php\n'
-        message += 'RTT home   : http://www.hep.ucl.ac.uk/atlas/AtlasTesting\n'
-
-        subject = 'Failures in %s (%s/%s/%s: %s)' % (str(self.name), release, branch, runType, platform)
-        sendMail(emailAddresses,subject, message)
-        self.logger.debug("Sent mail to: %s" % str(emailAddresses))
-            
-# ------------------------------------------------------------------------- 
-
-    def packageRequestsMailOnFailure(self):
-        return len(self.mailTo.strip()) > 0
-
-# -------------------------------------------------------------------------
-
-    def collectStats(self, minders):
-        def jobFailed(m):
-            jobStatus = m.status()
-            return jobStatus == 'error' and not 'Production Manager Veto' in m.errorMessages
-        def jobSucceeded(m):
-            return m.status()!='error'            
-        def testsFailed(m):
-            return [t.testId for t in m.finishedTests if t.result!=0]
-
-        if not self.packageRequestsMailOnFailure(): return
-        failed = {}
-        for m in minders:
-            jobName   = m.name
-            jobID     = m.jobSerialNumber
-            jobGroup  = m.jobGroup
-            if jobFailed(m):
-                failed[(jobName,str(jobID),jobGroup)] = []
-            elif jobSucceeded(m):
-                tf = testsFailed(m)
-                if tf:
-                    failed[(jobName,str(jobID),jobGroup)] = tf
-
-        self.emailStats.update(failed)
-   
diff --git a/Tools/RunTimeTester/src/PackageXMLFile.py b/Tools/RunTimeTester/src/PackageXMLFile.py
deleted file mode 100644
index 7e5c0ae5a738b491554b6c367e269252183ab113..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PackageXMLFile.py
+++ /dev/null
@@ -1,56 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, os.path
-from validateXMLFile import Validator
-from Evaluate        import Evaluate
-from exc2string2     import exc2string2
-from RTTSException   import RTTInputError
-
-class PackageXMLFile:
-    """
-    Do not stop the RTT if user input exceptions are thrown,
-    but keep the information and create error objects
-    (minders, test runs) later on for reporting to the user
-    """
-    
-    def __init__(self, pathToFile):
-        self.path    = pathToFile
-        self.errMsg  = ''
-        self._docEl  = None
-        self.isValid = self.validate_()
-
-    def exists(self):
-        return os.path.exists(self.path)
-
-    def validate_(self):
-        if not self.exists():
-            self.errMsg = 'User XML file %s does not exist' % self.path
-            return False
-
-        try:
-            validator = Validator(self.path) # throws exception if parse fails
-        except RTTInputError, e:
-            self.errMsg = 'User xml file does not parse:\n%s\nTraceback:\n%s' % (str(e), exc2string2())
-            return False
-
-        self._docEl = validator._dom.documentElement
-
-        try:
-            validator.validate() # throws exception if invalid
-        except RTTInputError, e:
-            self.errMsg = 'User input file does not validate:\n%s\nTraceback:\n%s' % (str(e), exc2string2())
-            return False
-        except:
-            return False
-        
-        return True
-
-    def __getitem__(self, tag):    
-        if not self.isValid: return None
-        return Evaluate(tag, self._docEl)
-
-    def __del__(self):
-        if not self.isValid: return
-        if self._docEl:
-            self._docEl.unlink()
-            del self._docEl
diff --git a/Tools/RunTimeTester/src/Page0XMLWriter.py.obsolete b/Tools/RunTimeTester/src/Page0XMLWriter.py.obsolete
deleted file mode 100755
index 4bef6160e80b6f1452218595cdcc8cb2ea9e5bea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Page0XMLWriter.py.obsolete
+++ /dev/null
@@ -1,151 +0,0 @@
-import os.path
-import xml.dom.ext
-from Tools2             import isValidReleaseNumber
-from makeRTTDOM         import makeRTTDOM
-from XMLTools           import xslRef, relPath, prependXSLRef
-from Evaluate           import getText, Evaluate
-from PageXMLWriter      import PageXMLWriter
-from Factory_Legals     import LegalsFactory
-
-import page0DefaultFiles
-
-class Page0XMLWriter(PageXMLWriter):
-
-    """
-    Page0XMLWriter collects together all information needed for Page0 into a DOM document,
-    and writes it out to the specified pag0.xml path.
-
-    The information includes the page1 locations. These are found by guessing: possible locatations
-    are given, and these are checked for the existence of page1.xml files.
-    """
-
-    
-    def __init__(self, tester, logger):
-        PageXMLWriter.__init__(self, tester, logger)
-        import xml.dom.ext, sys
-        docEl = self.tester.summaryDocument.documentElement
-
-        node = Evaluate('overview/resBaseDir', docEl)[0]
-        self.absPageLoc  = os.path.join(getText(node), 'page0.xml')
-        self.absXSLLoc   = os.path.join(self.absXSLDir, 'page0.xsl')
-
-
-        self.document = None
-        self.document = makeRTTDOM('RTTpage0')
-        prependXSLRef(xslRef(self.absXSLLoc, self.absPageLoc), self.document) 
-
-        factory = LegalsFactory(logger)
-        legals  = factory.create(tester.paths.releaseType, tester.paths.originalBranch, tester.paths.hardCoded)
-
-        self.possiblePage1Locations = []
-
-        # get legal names of nightly builds
-        nightlyNames = legals.legals['nightliesInRelease']
-        releaseNames = []
-        for branch in legals.legals['branch']:
-            releaseNames.extend([os.path.join(n,branch) for n in nightlyNames])
-
-        # now add in numbered releases
-        numberedBuilds = [f for f in os.listdir(self.absXSLDir)
-                          if os.path.isdir(os.path.join(self.absXSLDir,f)) and isValidReleaseNumber(f)]
-        
-        releaseNames.extend(numberedBuilds)
-                
-        for release in releaseNames:
-            for runType in legals.legals['runType']:                
-                self.possiblePage1Locations.append('%s/%s' % (release, runType))
-
-        self.createEmptyXMLDoc()
-
-        filesToWrite = {
-            'currentStatusN0X.js':       page0DefaultFiles.currentStatusN0X,
-            'currentStatusNX0.js':       page0DefaultFiles.currentStatusNX0,
-            'messageOfTheDay.js':        page0DefaultFiles.messageOfTheDay,
-            'page0.html':                page0DefaultFiles.page0,
-            'page0.noTopFrame.html':     page0DefaultFiles.page0_noTopFrame,
-            'status.html':               page0DefaultFiles.status,
-            'page0.xsl':                 page0DefaultFiles.page0XSL,
-            'page1.xsl':                 page0DefaultFiles.page1XSL,
-            'page2.xsl':                 page0DefaultFiles.page2XSL,
-            'page3.xsl':                 page0DefaultFiles.page3XSL,
-            'page4.xsl':                 page0DefaultFiles.page4XSL,
-            'RTTpage1_LinksAtTopOfPage.js': page0DefaultFiles.RTTpage1_LinksAtTopOfPage
-            }
-
-        def writeOut((fn, text), resBase):
-            fullPath = os.path.join(resBase, fn)
-            if os.path.exists(fullPath): return
-            of = open(fullPath,'w')
-            of.write(text)
-            of.close()
-
-        [writeOut(item, self.absXSLDir) for item in filesToWrite.items()]
-        self.update()
-
-    def createEmptyXMLDoc(self):
-        if self.document: self.document.unlink()
-        self.document     = makeRTTDOM('RTTpage0')
-        prependXSLRef(xslRef(self.absXSLLoc, self.absPageLoc),  self.document)
-        
-    def update(self, package=None, updatedMinders=None):
-
-        # self.logger.debug('Page0XMLWriter: start update')
-        p1 = 'page1.xml'
-        p1Locations = {}
-        
-        def fillDict(location, dict):
-            p1location = os.path.join(self.absXSLDir, location, p1)
-            # a location relative to resbase is provided because
-            # the string representing the location will be used by
-            # page0,html which MUST be located in resbase. The browser
-            # will then correctly use the relative path to page1.xml
-            p1relloc   = os.path.join(location, p1)
-            if os.path.exists(p1location): dict[location] = p1relloc
-            
-        [fillDict(l, p1Locations) for l in self.possiblePage1Locations]
-
-
-
-        #toRemove = self.document.documentElement.childNodes
-        #[self.document.documentElement.removeChild(n) for n in toRemove]
-        self.createEmptyXMLDoc()
-        
-        [self.addNode(p) for p in p1Locations.items()]
-
-
-
-        of = open(self.absPageLoc, 'w')
-        xml.dom.ext.PrettyPrint(self.document, of)
-        of.close()
-
-        # self.logger.info('Wrote out page0.xml to %s' % str(self.absPageLoc))
-        # self.logger.debug('Page0XMLWriter: end update')
-                        
-
-
-    def addNode(self, (name, location)):
-
-
-        n = self.document.createElement('page1')
-
-        n.appendChild(self.makeNode('page1Name',     name,   ) )
-        n.appendChild(self.makeNode('page1Location', location) ) 
-
-        self.document.documentElement.appendChild(n)
-
-    def makeNode(self, name, text):
-            
-        try:
-            tn = self.document.createTextNode(text)
-        except xml.dom.InvalidCharacterErr:
-            tn = self.document.createTextNode('invalid text')
-            
-        n  = self.document.createElement(name)
-        
-        n.appendChild(tn)
-
-        return n
-
-if __name__ == '__main__':
-    writer = Page0XMLWriter('/afs/cern.ch/user/s/sherwood/public/RTT/Results')
-    writer.update()
diff --git a/Tools/RunTimeTester/src/Page1XMLWriter.py.obsolete b/Tools/RunTimeTester/src/Page1XMLWriter.py.obsolete
deleted file mode 100755
index 9e0dcfd60438baf5c3551c908825dbe374a9b3de..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Page1XMLWriter.py.obsolete
+++ /dev/null
@@ -1,177 +0,0 @@
-import anydbm, os.path, sys
-from xml.dom.ext        import PrettyPrint
-from xml.dom.minidom    import parseString, parse
-from Tools              import mkdirIfNew
-from Evaluate           import Evaluate, getText
-from makeRTTDOM         import makeRTTDOM
-from RTTXMLFileSelector import RTTXMLFileSelector
-from RTTpath            import RTTpath
-from exc2string2        import exc2string2
-from PageXMLWriter      import PageXMLWriter
-from StringIO           import StringIO
-
-from XMLTools           import xslRef, prependXSLRef, relPath
-
-
-class Page1XMLWriter(PageXMLWriter):
-    def __init__(self, dbPath, key, tester, logger):
-
-        PageXMLWriter.__init__(self, tester, logger)
-
-        docEl = self.tester.summaryDocument.documentElement
-        resRunTypeDir   = Evaluate('overview/page1Dir/text()', docEl)
-        self.absPageLoc = os.path.join(resRunTypeDir, 'page1.xml')
-        self.absXSLLoc  = os.path.join(self.absXSLDir, 'page1.xsl')
-
-        self.document = None
-        self.createEmptyXMLDoc()
-
-        self.dbPath = str(dbPath)
-        
-        try:
-            # del key['release']
-            del key['targetCMTCONFIG']
-            del key['time']
-        except:
-            pass
-
-        self.key = key
-        
-        selectedData = RTTXMLFileSelector(self.dbPath,
-                                          self.key,
-                                          self.logger)
-
-    
-        nDBfiles = len(selectedData.keys())
-        m = 'Page1XMLWriter: selected %d runs from DB.' % nDBfiles
-        self.logger.info(m)
-        def makeDoms(key, xml):
-            
-            try:
-                dbDocument = parseString(xml)
-            except:
-                m  = 'Unable to parse moni DB entry with key %s \n' % str(key)
-                m += 'Skipping this value (it will not show up on page1.xml)\n'
-                m += exc2string2() + '\n'
-                self.logger.error(m)
-                dbDocument = None
-
-            return dbDocument
-
-        # get past and present summaries (present is also in the DB)
-        doms = [makeDoms(key,value) for key,value in selectedData.items()]
-        self.doms = [d for d in doms if not d is None]
-        self.doms.append(self.tester.summaryDocument)
-
-        self.selectSummaries()
-        self.logger.info('Number of doms in list after selection: %d' % len(self.doms))
-        return
-        
-    def createEmptyXMLDoc(self):
-        if self.document: self.document.unlink()
-        self.document     = makeRTTDOM('RTTpage1')
-        prependXSLRef(xslRef(self.absXSLLoc, self.absPageLoc),  self.document)
-
-    def selectSummaries(self):
-        '''remove undesired entries - allow only the most recent dom per targetCMTCONFIG
-        for a given release'''
-
-        domsByCMTCONFIG = {}
-
-        [domsByCMTCONFIG.setdefault(Evaluate('overview/targetCMTCONFIG/text()', d.documentElement), []).append(d)
-         for d in self.doms]
-        
-        for item in domsByCMTCONFIG.items():
-            domList = [(Evaluate('overview/startTime2/text()', d.documentElement), d) for d in item[1]]
-            domList.sort()
-            domsByCMTCONFIG[item[0]] = domList[-1]
-
-        self.doms = [dom  for (time, dom) in domsByCMTCONFIG.values()]
-        
-    def update(self, package=None, updatedMinders=None):
-
-        # self.logger.debug('Page1XMLWriter: start update')
-
-        # make a document that will hold information for runs read from
-        # the4 database.
-
-        # Calculate the page2 paths, check if they exist, and make relative
-        ndom = 0
-        #for d in doms:
-        #    temp = StringIO()
-        #    PrettyPrint(d, temp)
-        #    m = 'dump for dom %d\n%s' % (ndom, temp.getvalue())
-        #    self.logger.debug(m)
-        #    ndom += 1
-                                         
-        page2Paths = [(d, Evaluate('overview/resReleaseDir', d.documentElement)[0])
-                      for d in self.doms]
-        page2Paths = [(d, os.path.join(getText(p), 'page2.xml'))
-                      for d,p in page2Paths]
-
-        self.logger.debug('Page2 locations:\n%s' % str([str(p) for p in page2Paths]))
-        page2Paths = [(d,p) for d,p in page2Paths if os.path.exists(p)]
-
-        page2Paths = [(d,relPath(self.absPageLoc, p)) for d, p in page2Paths]
-
-
-        # transfer the relevant information to the page1 document
-        overviewNodes = [self.makeOverviewNode(p) for p in page2Paths]
-
-        #toRemove = self.document.documentElement.childNodes
-        #[self.document.documentElement.removeChild(n) for n in toRemove]
-        self.createEmptyXMLDoc()        
-        
-        [self.document.documentElement.appendChild(p) for p in overviewNodes]        
-
-        npage2 = len(page2Paths)
-
-            
-        m = 'Page1XMLWriter: %d files from DB, %d pages2 found' % (len(self.doms),
-                                                                   npage2)
-        self.logger.info(m)
-        of = open(self.absPageLoc, 'w')
-        PrettyPrint(self.document, of)
-        of.close()
-
-
-        
-        # self.logger.info('Wrote out page1.xml to %s' % str(self.absPageLoc))
-
-        # self.logger.debug('Page1XMLWriter: end update')
-
-    def makeOverviewNode(self, (srcDom, page2loc)):
-
-        locationNode = self.document.createElement('page2location')
-        textNode     = self.document.createTextNode(page2loc)
-        locationNode.appendChild(textNode)
-
-        overviewNode = Evaluate('overview', srcDom.documentElement)[0]
-        overviewNode = self.document.importNode(overviewNode, True)
-        overviewNode.appendChild(locationNode)
-        return overviewNode
-        
-    def finalise(self): self.update()
-
-if __name__== '__main__':
-    # This allows a the xml files required for the web display to be
-    # run manually.
-
-
-    class Logger:
-        def error(self, text): print text
-    # The full path to the database file with the xml files for the RTT runs/
-    dbPath = '/local_disk/sherwood/RTT02/moniFile.db'
-
-    # The directory in which to pace the page1 xml files. The xml files for
-    # subsequent pages will be placed in subdirectories labelled by branch.
-    page1Dir   = '/local_disk/sherwood/RTT02/Junk'
-    page2Dir   = '/local_disk/sherwood/RTT02/Junk'
-
-    # Supply a dictionary the has the selection parameters, e.g. the branch.
-    # The entries of the key will compared against the entries in the
-    # dictionary which is used to store the xml file in the db.
-    
-    key    = {u'originalBranch': 'N.X.0'}
-
-    rsr = Page1XMLWriter(dbPath, page1Dir, page2Dir, key, Logger())
diff --git a/Tools/RunTimeTester/src/Page2XMLWriter.py.obsolete b/Tools/RunTimeTester/src/Page2XMLWriter.py.obsolete
deleted file mode 100755
index 2de37d5d3567c279005ea9caee189bf15809c143..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Page2XMLWriter.py.obsolete
+++ /dev/null
@@ -1,41 +0,0 @@
-from makeRTTDOM     import makeRTTDOM
-from Evaluate	    import Evaluate, getText
-from xml.dom.ext    import PrettyPrint
-from XMLTools       import xslRef, prependXSLRef
-from PageXMLWriter  import PageXMLWriter
-from StringIO       import StringIO
-import os.path
-
-class Page2XMLWriter(PageXMLWriter):
-	def __init__(self, tester, logger):
-
-		PageXMLWriter.__init__(self, tester, logger)
-		absXSLLoc = os.path.join(self.absXSLDir, 'page2.xsl')
-
-		docEl = self.tester.summaryDocument.documentElement
-		node = Evaluate('overview/resReleaseDir', docEl)[0]
-		self.absPageLoc = os.path.join(getText(node),'page2.xml')
-
-		self.xslref = xslRef(absXSLLoc, self.absPageLoc)
-		
-		print '=========page2 loc ========', self.absPageLoc 
-		
-		
-	def update(self, package, updatedMinders):
-		# self.logger.debug('Page2XMLWriter: start update')
-		# transfer nodes
-
-		# temporary addition of the pointer to the xsl file for this page
-		xslRefNode = prependXSLRef(self.xslref,self.tester.summaryDocument)
-		
-		#self.tester.summaryDocument.appendChild(self.xslInstructionNode)
-
-		
-		of = open(self.absPageLoc,'w')
-		PrettyPrint(self.tester.summaryDocument, of)
-		#of.write(self.tester.summaryDocument.toxml())
-		of.close()
-
-		# self.tester.summaryDocument.removeChild(self.xslInstructionNode)
-		self.tester.summaryDocument.removeChild(xslRefNode)
-		xslRefNode.unlink()
diff --git a/Tools/RunTimeTester/src/Page3XMLWriter.py.obsolete b/Tools/RunTimeTester/src/Page3XMLWriter.py.obsolete
deleted file mode 100755
index 158315c6825488d6538cfe0cdc86b98fd29e6b05..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Page3XMLWriter.py.obsolete
+++ /dev/null
@@ -1,51 +0,0 @@
-from Evaluate      import Evaluate, getText
-from xml.dom.ext   import PrettyPrint
-from RTTpath       import RTTpath
-from XMLTools      import xslRef, prependXSLRef
-from makeRTTDOM    import makeRTTDOM
-from PageXMLWriter import PageXMLWriter
-
-import os.path
-
-class Page3XMLWriter(PageXMLWriter):
-	def __init__(self, tester, logger):
-		PageXMLWriter.__init__(self, tester, logger)
-		self.absXSLLoc = os.path.join(self.absXSLDir, 'page3.xsl')
-
-	def update(self, updatedPackage, updatedMinders):
-		# self.logger.debug('Page3XMLWriter: start update')
-		rttSummaryDocument = self.tester.summaryDocument
-
-		sDocumentElement =  rttSummaryDocument.documentElement
-		packageList = sDocumentElement.getElementsByTagName('package')
-		packageNode = [p for p in  packageList if Evaluate('packageName/text()', p) == updatedPackage][0]
-		
-		# create one page3 per package
-		overviewNode = Evaluate('overview', sDocumentElement)[0]
-		self.createPage( overviewNode, packageNode )
-		# self.logger.debug('Page3XMLWriter: end update')
-
-
-	def createPage(self, overviewNode, packageNode):
-		rttSummaryDocument = self.tester.summaryDocument
-		doc   = makeRTTDOM('RTTpage3')
-
-
-		absPageLoc = RTTpath(
-			Evaluate('resReleaseDir/text()', overviewNode),
-			Evaluate('packageName/text()',   packageNode),
-			'page3.xml')
-		
-		absPageLoc = str(absPageLoc)
-
-		prependXSLRef(xslRef(self.absXSLLoc, absPageLoc), doc)
-
-		doc.documentElement.appendChild(doc.importNode(overviewNode, True))
-		doc.documentElement.appendChild(doc.importNode(packageNode, True))
-
-		of = open(absPageLoc, 'w')
-		PrettyPrint(doc,of)
-		of.close()
-
-		doc.unlink()
-		# self.logger.debug('Wrote out page3.xml to %s' % str(absPageLoc))
diff --git a/Tools/RunTimeTester/src/Page4XMLWriter.py.obsolete b/Tools/RunTimeTester/src/Page4XMLWriter.py.obsolete
deleted file mode 100755
index e894065bc82503b18d2af8f9df3a8f509c9727e3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Page4XMLWriter.py.obsolete
+++ /dev/null
@@ -1,49 +0,0 @@
-from Evaluate      import Evaluate, getText
-from xml.dom.ext   import PrettyPrint
-from RTTpath       import RTTpath
-from XMLTools      import xslRef, prependXSLRef
-from makeRTTDOM    import makeRTTDOM
-from PageXMLWriter import PageXMLWriter
-import os.path
-
-class Page4XMLWriter(PageXMLWriter):
-   def __init__(self, tester, logger):
-
-      PageXMLWriter.__init__(self, tester, logger)
-      self.absXSLLoc = os.path.join(self.absXSLDir, 'page4.xsl')
-      self.doms  = {}
-
-   def update(self, package, updatedMinders):
-      self.logger.debug('Page4XMLWriter: start update')
-      rttDocEl = self.tester.summaryDocument.documentElement
-      overviewNode = Evaluate('overview', rttDocEl)[0]
-      minders   = Evaluate('package/minder', rttDocEl)
-      minders   = [m for m in minders if Evaluate('identifiedName/text()', m) in updatedMinders]
-      [self.createPage(overviewNode, minder) for minder in minders]
-      self.logger.debug('Page4XMLWriter: end update')
-
-
-   def createPage(self, overview, minder):
-
-      self.logger.debug('Page4XMLWriter: start createPage')
-      doc   = makeRTTDOM('RTTpage4')
-      docEl = doc.documentElement
-      docEl.appendChild(doc.importNode(overview, True))
-      docEl.appendChild(doc.importNode(minder,   True))
-	      
-      absPageLoc = str(RTTpath(
-         Evaluate('resultsPath/text()', minder),
-         'page4.xml'))
-
-      prependXSLRef(xslRef(self.absXSLLoc, absPageLoc), doc)
-
-      of = open(absPageLoc,'w')
-      PrettyPrint(doc,of)
-      of.close()
-
-
-      doc.unlink()
-      # print '=========page4 loc ========',absPageLoc
-      self.logger.debug('Wrote out page4.xml to %s' % absPageLoc)
-      self.logger.debug('Page4XMLWriter: end createPage')
- 
diff --git a/Tools/RunTimeTester/src/PageWriter.py.obsolete b/Tools/RunTimeTester/src/PageWriter.py.obsolete
deleted file mode 100755
index a7b71415289413c018a0fc3e53f72a36d925058a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PageWriter.py.obsolete
+++ /dev/null
@@ -1,636 +0,0 @@
-# Write out web page xml files for the RTT
-# Input is a single XML file which stores the RTT summary xml files
-# Structure of the code:
-# 5 page writer classes: Page0Writer,..,Page4Writer, all derived from PageWriter.
-# The Page(n)Writer receives information in its update() method (supplied by the base class PageWriter) in the form of an document node,
-# extracts the subnodes in the getSubNodes method, and instantiates the corresponding number of Page(n+1)Writers. Page(n)Writer can
-# find the location of the n+1 generation pages by asking the writers.
-# at the end of the method a call is made to writepage to write out the xml. In this way the order of writing out pages is
-# 4,3,2,1,0. The lifetime of the writer objects is the lifetime of the update method in Page0Writer.
-#
-# Page0Writer is different from the other page writers in that it deals with muliple summaries (bundled together into a single
-# document to allow processing with similar codee). All pages are initialised with the overview node of the document they will deal
-# with. A single overview node makes no sense in the case of page0writer and page1writer, which get a None object instead.
-#
-from Evaluate   import Evaluate
-from RTTpath    import RTTpath
-from makeRTTDOM import makeRTTDOM
-from XMLTools   import xslRef, prependXSLRef, relPath
-
-import os.path, string
-
-# print level: current settings are at 100. 0 = silent.
-debug = 0
-
-class SummaryLocationInterpreter:
-    """
-    useless C++-like interface
-    This class is acts as a reminder that location
-    is a general concept (place in database, url, unix file path.
-    Derived classes know how to handle specific representations of location
-
-    These classes are used to make the page names. The names are currently only
-    used to indentifiy the page for debug purposes.
-    """
-    def __init__(self): return
-    def linkedPageName(self): return ''
-    def writeTo(self):
-        'path to write current page to'
-        return ''
-    
-class SummaryPathInterpreter(SummaryLocationInterpreter):
-    """
-    location is implemented as a relative path. The code knows
-    how to interpret the path. In this case, it knows how
-    with parts of the path to use as pointers to the next page.
-    the pointer values are set in derived classes
-
-    Note that the location is not used to find the summary file,
-    rather information is extracted from the path.
-    """
-    
-    def __init__(self, location):
-        SummaryLocationInterpreter.__init__(self)
-
-
-        # if not os.path.isabs(location):
-        #    print location
-
-        # print location
-
-        # assert False
-
-        self.tokens = location.split('/')
-
-    def linkedPageName(self):
-        # return string.join(self.tokens[self.minIndex:self.maxIndex], '')
-
-        path = ''                      
-        for t in self.tokens[self.minIndex:self.maxIndex]:
-            path = os.path.join(path, t)
-        return path
-
-class Page0PathInterpreter(SummaryPathInterpreter):
-    """
-    provide pointers into the location path to point from page 0
-    to page 1
-    """
-
-    def __init__(self, summaryLocation):
-        self.minIndex = 1
-        self.maxIndex = 4
-        self.nextPage = 'page1.xml'
-        SummaryPathInterpreter.__init__(self, summaryLocation)
-
-
-class Page1PathInterpreter(SummaryPathInterpreter):
-    """
-    provide pointers into the location path to point from page 1
-    to page 2
-    """
-    
-    def __init__(self, summaryLocation):
-        # self.minIndex = 4
-        # self.maxIndex = 5
-        self.minIndex = 1
-        self.maxIndex = 6
-        self.nextPage = 'page2.xml'
-        SummaryPathInterpreter.__init__(self, summaryLocation)
-
-
-    def writeTo(self):
-        path = ''
-        for t in self.tokens[1:4]:
-            path = os.path.join(path, t)
-            # print path
-        return path
-
-class PageWriterBase:
-    def __init__(self, resBaseDir, name):
-        # print 'Constructor %s for page %s' %(self.__class__.__name__, name) 
-        self.resBaseDir      = resBaseDir       # paths to XSLT files
-        self.absXSLDir       = resBaseDir       # paths to XSLT files
-        self.absXSLLoc       = os.path.normpath(os.path.join(self.absXSLDir, self.xslFile))
-        self.name            = name             # a path-like string> ends up on the web page.
-        self.writerClass     = None             # class for child writers
-        self.childWriters    = {}               # location (where written) to page writer mapping
-                                                # the dictionary ensures there is only one such writer
-
-    
-        # print 'Constructor ends %s for page %s' %(self.__class__.__name__, name) 
-
-
-        
-    def dump(self, ngen, counts):
-        ngen += 1
-        tab = '   '*ngen
-        relLoc = relPath(self.resBaseDir, self.location)
-        s = '\n%s%s %s\n%s%s\n\n' % (tab,
-                                     self.__class__.__name__,
-                                     self.name,
-                                     tab,
-                                     relLoc)
-        #s+=  'from dump: no of children %d\n' % len(self.childWriters.keys())
-        for c in self.childWriters.values():
-            s += c.dump(ngen, counts)
-
-        cn = self.__class__.__name__
-        try:
-            counts[cn]+=1
-        except:
-            counts[cn]=1
-            
-        return s
-
-    def writePage(self, doc):
-        # occasionnally WPC has a permission denied problem.
-        # not sure why: perhaps temp AFS problem. Hence this retry solution.
-        thereAreProblems = True
-        attempts = 0
-        while attempts < 5 and thereAreProblems:
-            try:
-                self._writePage(doc)
-                thereAreProblems = False
-            except:
-                print "Attempt %d: PageWriterBase._writePage() threw an exception" % attempts
-                print exc2string2()
-                attempts += 1
-                time.sleep(10)
-
-    def _writePage(self, doc):
-        """
-        write out the page for a job.
-        """
-
-        if debug >= 50: print 'writing page %s::%s:\n    %s' % (self.__class__.__name__, self.name, self.location)
-        of = open(self.location,'w')
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-        doc.unlink()
-
-
-    def registerChild(self, childWriter):
-        """
-        childWriter passed in is a candidate: it is refused
-        if there is already a writer existing at this location.
-        otherwise children with same location (but presumably
-        different offspring) eliminarte each other.
-        """
-        cn = childWriter.__class__.__name__
-        writer = self.childWriters.setdefault(childWriter.location, childWriter)
-        # if cn in ['Page1Writer','Page2Writer','Page1Writer']:
-        #    print 'registering child of class',cn
-        #    print '--start dict---'
-        #    print self.childWriters
-        #    print '--end dict---'
-        # print 'registered child at location', childWriter.location
-        return writer
-    def deregisterChildren(self):
-        """
-        pages are deregistered if the cummary file no longer needs updating
-        to conserve memory
-        This base version passes the command on to the children - some
-        classes never take action on the commands. Others do so in
-        overrides.
-        """
-    
-        for c in self.childWriters.values(): c.deregisterChildren()
-
-
-class MultipleSummaryPageWriter(PageWriterBase):
-    """
-    Base class for pages that handle more than 1 summary file.
-    """
-    def __init__(self, resBaseDir, name):
-        PageWriterBase.__init__(self, resBaseDir, name)
-        
-    def update(self, cmttargets, packages, minders, writePages, summaryLocation, overviewForChild, nodeForChild):
-        """
-        cmttargets list of values to be updated
-        package    list of values to be updated
-        minders    list of values to be updated
-        write      flag to enable page writing (not always fone at init for the system)
-        """
-
-
-        if debug>=55: print '\nupdate start for', self.__class__.__name__, self.name
-        # if there are to be linked pages, make a new writer for each new location
-        writer = self.makeWriter(summaryLocation, overviewForChild, nodeForChild)
-        if debug >= 100: print '\nupdate made writer', writer.__class__.__name__
-
-        if debug >= 100: print '%s registering  writer %s' % (self.name, writer.name)
-        # use this writer as a child only of there is not an older one at this location
-        writer = self.registerChild(writer)
-        # write out the linked pages
-        # print 'registered, about to update'
-        nupdates = writer.update(cmttargets, packages, minders, writePages, summaryLocation, overviewForChild, nodeForChild)+1
-
-        # write out the current page
-        doc = self.composePage(overviewForChild, nodeForChild)
-
-
-        if writePages: self.writePage(doc)
-        if debug>30: print 'path to write to:', self.location
-                
-        doc.unlink()
-        # print '\nupdate end for', self.__class__.__name__
-        if debug>=55: print '\nupdate ends for', self.__class__.__name__, self.name
-
-        if debug>=40: print 'nupdates: ', nupdates, self.__class__.__name__
-        return nupdates
-    
-class SingleSummaryPageWriter(PageWriterBase):
-    def __init__(self, resBaseDir, name):
-        PageWriterBase.__init__(self, resBaseDir, name)
-
-        
-    def getSubNodes(self, cmttargets, packages, minders, nodeToSplit):
-        """
-        Get the minder nodes which will geneate the page4 files linked from page3.
-        """
-        # example - page 4: terminal page, pathToSubNodes = ''
-        pathToSubNodes = self.getXPathToSubNodes()
-        if not pathToSubNodes: return []
-        
-        # example: page 3 - pathToSubNodes = 'minder', ie the relative xpath from package to minder nodes
-        allNodes = Evaluate(pathToSubNodes, nodeToSplit)
-        # print 'name', self.name
-        # print 'pathToSubNodes', pathToSubNodes
-        # print 'no of subNodes', len(allNodes)
-        # if len(allNodes) == 0:
-        #    print self.node.nodeName
-        #    print self.node.toxml()
-        #    assert False
-        
-        # example: page3  - [ (<identifiedName>, <minder)....]
-        # create a 'path' like name to be able to identify the children of the node in the PageTable
-        def makeWriterName(parentName, nodeNamePath, node):
-            nodeName  = parentName+'/'
-            nodeName +=  Evaluate(nodeNamePath, a).strip()
-            return nodeName
-        
-        allNodes = [(makeWriterName(self.name, self.getUpdatedNodeName(), a), a) for a in allNodes]
-
-        # example: page 3 - absence of a list of updated minder identified names  means take all (init from db)
-        #                   otherwise get the list of identified names for the updated minders for this package
-        
-        updated = self.getUpdatedNodeNameValues(cmttargets, packages, minders)
-        if updated:
-            # evaluate the end of the name path - this has to match the update information.
-            # example: page3 - keep only the minder nodes that correspond to the updated identified names
-
-            
-
-            nodes = [(name, node) for name, node in allNodes if Evaluate(self.getUpdatedNodeName(), node) in updated]
-
-
-            if debug>= 100:
-                # m = '%s::%s subnodes:\n' % (self.__class__.__name__,self.name)
-                for name, node in nodes:
-                    m += '   %s\n ' % name
-                print m
-
-                print updated
-            
-            if not len(nodes) == len(updated):
-                print nodeToSplit.toprettyxml('   ')
-                assert False
-
-            return nodes
-        else:
-
-            if debug>=100:
-                m = '%s::%s subnodes:\n' % (self.__class__.__name__,self.name)
-                for name, node in allNodes:
-                    m += '   %s\n ' % name
-                print m
-
-            return allNodes
-        
-    def update(self, cmttargets, packages, minders, writePages, summaryLocation,
-               overviewForChild, nodeForChild):
-
-
-        # path to RTTSummary.xml relative to resBase.
-        # Used to caculate self.name
-        self.summaryLocation = summaryLocation
-        
-        'package and minders are hints for subnode selection'
-        if debug>=100: print '\nupdate start for', self.__class__.__name__
-
-        # print '\nupdate start for', self.__class__.__name__
-        # find all the nodes for linked pages
-        subNodes  = self.getSubNodes(cmttargets, packages, minders, nodeForChild)
-        if debug>=100: print self.name,': no of subnodes', len(subNodes)
-        
-        # if there are to be linked pages, make a new writer for each new location
-        nupdates = 0
-        if self.writerClass:
-            for name, subnode in subNodes:
-                writer = self.writerClass(self.resBaseDir, name, overviewForChild, subnode)
-                if writer.location == '': continue # vetoed packages have no place to write to. xslt handles these cases.
-
-                writer = self.registerChild(writer)
-
-                nupdates += writer.update(cmttargets, packages, minders, writePages, summaryLocation, overviewForChild, subnode)+1
-
-        # write out the current page
-        doc = self.composePage(overviewForChild, nodeForChild)
-        if writePages: self.writePage(doc)
-        doc.unlink()
-        # print '\nupdate end for', self.__class__.__name__
-
-        if debug>=40:  print 'nupdates: ', nupdates, self.__class__.__name__
-        return nupdates
-    
-    def deregisterChildren(self):
-        """
-        pages are deregistered if the cummary file no longer needs updating
-        to conserve memory
-        """
-        PageWriterBase.deregisterChildren(self)
-        for c in self.childWriters.keys(): del self.childWriters[c]
-
-
-
-class Page4Writer(SingleSummaryPageWriter):
-    def __init__(self, resBaseDir, name, overviewNode, node):
-        'node is a minder node'
-        self.xslFile = 'page4.xsl'
-        SingleSummaryPageWriter.__init__(self, resBaseDir, name)
-
-        # location - where the page is to be written.
-        # Bind as is needed by parent page
-        self.location  = self.getLocation(node)
-
-
-    def getLocation(self, node):
-        location = os.path.join(Evaluate('resultsPath/text()', node).strip(), 'page4.xml')
-        assert os.path.isabs(location)
-        return location
-
-    def getXPathToSubNodes(self): return ''
-
-    def getUpdatedNodeName(self): return ''
-
-    def getUpdatedNodeNameValues(self, cmttargets, packageTargets, minderTargets): return []
-
-    def composePage(self, overviewNode, node):
-        # self.logger.debug('Page4XMLWriter: start createPage')
-        doc   = makeRTTDOM('RTTpage4')
-        docEl = doc.documentElement
-        docEl.appendChild(doc.importNode(overviewNode, True))
-        docEl.appendChild(doc.importNode(node, True))
-
-        # print self.location
-        prependXSLRef(xslRef(self.absXSLLoc, self.location), doc)
-        # print self.location
-        # assert False
-        
-        return doc
-
-
-class Page3Writer(SingleSummaryPageWriter):
-    """
-    recieves a package node, passes minder nodes to Page4Writer
-    """
-    def __init__(self, resBaseDir, name, overviewNode, node):
-        'node is a package node'
-        
-        self.xslFile = 'page3.xsl'
-        SingleSummaryPageWriter.__init__(self, resBaseDir, name)
-        self.writerClass = Page4Writer
-
-        # location - where the page is to be written.
-        # Bind as is needed by parent page
-        self.location  = self.getLocation(node)
-        self.relPathFromParent = Evaluate('relPage2Page3Path/text()', node)
-
-
-
-    def getLocation(self, node):
-        # location = os.path.join(Evaluate('resReleaseDir/text()', overviewNode).strip(),
-        #                        Evaluate('packageName/text()', node).strip(),
-        #                        'page3.xml')
-        location = Evaluate('absPage3Path/text()', node)
-        # print 'location: ',location
-        # print node.toxml()
-        # vetoed packages have no location
-        if location != '': assert os.path.isabs(location)
-
-        return location
-
-    def getXPathToSubNodes(self): return 'minder'
-
-    def getUpdatedNodeName(self): return 'identifiedName/text()'
-
-    def getUpdatedNodeNameValues(self, cmttargets, packageTargets, minderTargets): return minderTargets
-    def composePage(self, overviewNode, node):
-        """
-        write out the page3 and call update for page4 objects for a package.
-        """
-        # self.logger.debug('Page4XMLWriter: start createPage')
-        doc   = makeRTTDOM('RTTpage3')
-        docEl = doc.documentElement
-        docEl.appendChild(doc.importNode(overviewNode,    True))
-        docEl.appendChild(doc.importNode(node, True))
-
-        prependXSLRef(xslRef(self.absXSLLoc, self.location), doc)
-
-        return doc
-
-
-
-
-class Page2Writer(SingleSummaryPageWriter):
-    """
-    recieves a summary node, passes package nodes to Page3Writer
-    """
-
-    def __init__(self, resBaseDir, name, overviewNode, node):
-        'node is the full summary document'
-        
-        self.xslFile = 'page2.xsl'
-
-        SingleSummaryPageWriter.__init__(self, resBaseDir, name)
-        self.writerClass = Page3Writer
-        # location - where the page is to be written.
-        # Bind as is needed by parent page
-        self.location  = self.getLocation(overviewNode)
-        self.relPathFromParent = Evaluate('relPage1Page2Path/text()', overviewNode)
-
-    def update(self, cmttargets, packages, minders, writePages, summaryLocation,
-               overviewForChild, nodeForChild):
-
-        self.overviewNode = overviewForChild # needed by Page 1
-        nupdates = SingleSummaryPageWriter.update(self, cmttargets, packages, minders, writePages, summaryLocation,
-                                           overviewForChild, nodeForChild)
-
-        if debug>=40:  print 'nupdates: ', nupdates, self.__class__.__name__
-
-        return nupdates # number of updated pages
-    
-    def getLocation(self, overviewNode):
-        # location = os.path.join(Evaluate('resReleaseDir/text()', overviewNode).strip(),
-        #                              'page2.xml') 
-        location = Evaluate('absPage2Path/text()', overviewNode)
-
-        assert os.path.isabs(location)
-        return location
-
-
-    def getXPathToSubNodes(self): return 'package'
-
-    def getUpdatedNodeName(self): return 'packageName/text()'
-
-    def getUpdatedNodeNameValues(self, cmttargets, packageTargets, minderTargets): return packageTargets
-        
-    def composePage(self, overviewNode, node):
-        
-        doc   = makeRTTDOM('RTTSummary')
-        docEl = doc.documentElement
-        [docEl.appendChild(doc.importNode(n, True)) for n in node.childNodes]
-        # print self.location
-        xslRefNode = prependXSLRef(xslRef(self.absXSLLoc, self.location), doc)
-        # print self.location
-        #  assert False
-
-        return doc
-
-
-
-
-class Page1Writer(MultipleSummaryPageWriter):
-    """
-    recieves a RelBranchType node, passes summary nodes to Page2Writer
-    """
-    # def __init__(self, resBaseDir, name, overviewNode, node):
-    def __init__(self, resBaseDir, name):
-        'node is RelBranchType'
-
-        # page0 has no overview node - it deals with
-        # more than one doms, and a single overview node makes no sens.
-        # the overview node passed in is None, and needs to be set here
-        # so it can be transmitted down to other page writers
-        self.xslFile = 'page1.xsl'
-        MultipleSummaryPageWriter.__init__(self, resBaseDir, name)
-        # location - where the page is to be written.
-        # Bind as is needed by parent page
-        try:
-            self.location  = self.getLocation()
-        except:
-            print "Page1Writer::getLocation error. Node = \n%s"  % node.toprettyxml()
-        # self.relPathFromParent = Evaluate('relPage0Page1Path/text()', overviewNode)
-
-    # def getLocation(self, overviewNode):
-    def getLocation(self):
-
-        location =  os.path.join(self.resBaseDir, 'page1.xml')
-
-
-        # location = Evaluate('absPage1Path/text()', overviewNode)
-
-        assert os.path.isabs(location)
-        return location
-
-
-    def composePage(self, overviewnode, node):
-
-        doc   = makeRTTDOM('RTTpage1')
-        docEl = doc.documentElement
-
-        for w in self.childWriters.values():
-
-            # inport page2 overview and location into the current document
-
-            relPathToPage2 = w.relPathFromParent
-            relPathToSummXML = os.path.join(os.path.dirname(relPathToPage2), 'RTTSummary.xml')
-            
-            on = doc.importNode(w.overviewNode, True)
-            el = doc.createElement('relPage1Page2Path')
-            el2 = doc.createElement('relPathToSummaryXML')
-            el.appendChild(doc.createTextNode(relPathToPage2))
-            el2.appendChild(doc.createTextNode(relPathToSummXML))
-            on.appendChild(el)
-            on.appendChild(el2)
-            docEl.appendChild(on)
-
-        prependXSLRef(xslRef(self.absXSLLoc, self.location), doc)
-
-
-        # print 'Page1.compose start: doc:'
-        # print doc.toprettyxml('   ')
-        # print 'Page1.compose: end'
-
-        return doc
-
-    def makeWriter(self, summaryLocation, overviewNode, node):
-        
-        summaryPathInterpreter = Page1PathInterpreter(summaryLocation)
-
-        pageName = string.join([self.name, summaryPathInterpreter.linkedPageName()], '/')
-        return Page2Writer(self.resBaseDir, pageName, overviewNode, node)
-
-class Page0Writer(MultipleSummaryPageWriter):
-    """
-    receives a packedSummaries node, passes CMTCONFIG to Page1Writer
-    """
-
-    def __init__(self, resBaseDir, name, overviewNode, node):
-        
-        self.xslFile = 'page0.xsl'
-        MultipleSummaryPageWriter.__init__(self, resBaseDir, name)
-        # location - where the page is to be written.
-        # Bind as is needed by parent page
-        self.location  = self.getLocation()
-
-
-    def getLocation(self):
-        # self.location =  os.path.join(Evaluate('resBaseDir/text()', self.overviewNode).strip(),
-        #                              'page0.xml')
-        location =  os.path.join(self.resBaseDir, 'page0.xml')
-        assert os.path.isabs(location)
-        return location
-
-
-    def composePage(self, overviewNode, node):
-        doc   = makeRTTDOM('RTTpage0')
-        docEl = doc.documentElement
-
-        # for w in registeredWriters:
-        for w in self.childWriters.values():
-            p1name = doc.createElement('page1Name')
-
-            # remove self.name for the front of the writers' names
-            p1name.appendChild(doc.createTextNode(w.name[len(self.name):]))
-            page1  = doc.createElement('page1')
-            page1.appendChild(p1name)
-
-            el = doc.createElement('relPage0Page1Path')
-            el.appendChild(doc.createTextNode(w.relPathFromParent))
-            page1.appendChild(el)
-
-            docEl.appendChild(page1)
-        
-        prependXSLRef(xslRef(self.absXSLLoc, self.location), doc)
-
-
-        return doc
-
-    def makeWriter(self, summaryLocation, overviewNode, node):
-        summaryPathInterpreter = Page0PathInterpreter(summaryLocation)
-        pageName = string.join([self.name, summaryPathInterpreter.linkedPageName()], '/')
-        return Page1Writer(self.resBaseDir, pageName, overviewNode, node)
-
-    def update(self, cmttargets, packages, minders, writePages, summaryLocation, overviewNode, de):
-        nupdates = MultipleSummaryPageWriter.update(self, cmttargets, packages, minders,
-                                         writePages, summaryLocation, overviewNode, de)
-        if debug>=40:  print 'nupdates: ', nupdates, self.__class__.__name__
-        
-        return nupdates # number of updated pages
-        
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/PageWriterTest.py.obsolete b/Tools/RunTimeTester/src/PageWriterTest.py.obsolete
deleted file mode 100755
index f8f114d619e3dcf2a5ee8c75c3380cdd8ab9ab3d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PageWriterTest.py.obsolete
+++ /dev/null
@@ -1,79 +0,0 @@
-from PageWriter import Page0Writer, Page1Writer, Page2Writer, Page3Writer, Page4Writer
-
-class Page4WriterTest(Page4Writer):
-
-    def __init__(self, register, o, r, na, no):
-        Page4Writer.__init__(self, register, o, r, na, no)
-
-    def writePage(self):
-        # will be overwritten many times...
-        of = open('page4Test.xml','w')
-        doc = self.composePage()
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-    def checkPage(self): return True
-
-
-class Page3WriterTest(Page3Writer):
-
-    def __init__(self, register, o, r, na, no):
-        Page3Writer.__init__(self, register, o, r, na, no)
-        self.writerClass = Page4WriterTest
-
-    def writePage(self):
-        # will be overwritten many times...
-        of = open('page3Test.xml','w')
-        doc = self.composePage()
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-    def checkPage(self): return True
-    
-
-class Page2WriterTest(Page2Writer):
-
-    def __init__(self, register, o, r, na, no):
-        Page2Writer.__init__(self, register, o, r, na, no)
-        self.writerClass = Page3WriterTest
-
-    def writePage(self):
-        # will be overwritten many times...
-        of = open('page2Test.xml','w')
-        doc = self.composePage()
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-    def checkPage(self): return True
-    
-
-class Page1WriterTest(Page1Writer):
-
-    def __init__(self, register, o, r, na, no):
-        Page1Writer.__init__(self, register, o, r, na, no)
-        self.writerClass = Page2WriterTest
-
-    def writePage(self):
-        of = open('page1Test.xml','w')
-        doc = self.composePage()
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-
-    def checkPage(self): return True
-    
-
-class Page0WriterTest(Page0Writer):
-
-    def __init__(self, register, o, r, na, no):
-        Page0Writer.__init__(self, register, o, r, na, no)
-        self.writerClass = Page1WriterTest
-
-    def writePage(self):
-        of = open('page0Test.xml','w')
-        doc = self.composePage()
-        of.write(doc.toprettyxml('   '))
-        of.close()
-
-    def checkPage(self): return True
-    
diff --git a/Tools/RunTimeTester/src/PageXMLWriter.py.obsolete b/Tools/RunTimeTester/src/PageXMLWriter.py.obsolete
deleted file mode 100755
index be53a9d55491fb16b919829474d71d956f5d7ca9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PageXMLWriter.py.obsolete
+++ /dev/null
@@ -1,19 +0,0 @@
-"""
-A base class for page writers
-"""
-import os.path
-from Evaluate    import Evaluate, getText
-import xml.dom.ext, sys
-class PageXMLWriter:
-    def __init__(self, tester, logger):
-        self.tester = tester
-        self.logger = logger
-
-        document= self.tester.summaryDocument
-        docEl = document.documentElement
-        # xml.dom.ext.PrettyPrint(document, sys.stdout)
-        node = Evaluate('overview/resBaseDir', docEl)[0]
-        self.absXSLDir = getText(node)
-
-    def finalise(self): pass
-    
diff --git a/Tools/RunTimeTester/src/ParallelDescriptor.py b/Tools/RunTimeTester/src/ParallelDescriptor.py
deleted file mode 100644
index 8e757930aae184ecb88adb47608f35954c581f48..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ParallelDescriptor.py
+++ /dev/null
@@ -1,9 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from ContainerDescriptor import ContainerDescriptor
-
-class ParallelDescriptor(ContainerDescriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-        ContainerDescriptor.__init__(self, paths, logger, jobSerialNumber)
-        self.name = 'Parallel'
-        self.identifiedName = self.name+str(self.jobSerialNumber)
diff --git a/Tools/RunTimeTester/src/ParametersForStampObject.py.obsolete b/Tools/RunTimeTester/src/ParametersForStampObject.py.obsolete
deleted file mode 100755
index cc36ac8c3a0bfd515cdfe42c51e5f2251adc0687..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ParametersForStampObject.py.obsolete
+++ /dev/null
@@ -1,26 +0,0 @@
-from Factory_HardCoded       import HardCodedFactory
-from Factory_SiteSpecifics   import SiteSpecificsFactory
-from Factory_Legals          import LegalsFactory
-from Factory_PathNameBuilder import PathNameBuilderFactory
-from RTTpath                 import RTTpath
-
-class ParametersForStampObject:
-    def __init__(self,rttConfig,logger):
-        self.runType    = rttConfig.get('runType')
-        releaseType     = rttConfig.get('releaseType')
-        originalBranch  = rttConfig.get('originalBranch')
-        cmtConfig       = rttConfig.get('cmtConfig')
-        site            = rttConfig.get('site')
-        hardCoded       = HardCodedFactory(logger).create(releaseType, self.runType)
-        siteSpecifics   = SiteSpecificsFactory(logger).create(site)
-        legals          = LegalsFactory(logger).create(releaseType,
-                                                       originalBranch,
-                                                       hardCoded)
-        self.pathNameBuilder = PathNameBuilderFactory(logger).create(releaseType,
-                                                                     rttConfig,
-                                                                     hardCoded,
-                                                                     legals,
-                                                                     siteSpecifics)
-        
-        self.distArea  = RTTpath(rttConfig.get('distArea'))
-            
diff --git a/Tools/RunTimeTester/src/PathCompleter.py.obsolete b/Tools/RunTimeTester/src/PathCompleter.py.obsolete
deleted file mode 100755
index f12e2cb7519483bb336ed20adaa2c5bbdf19ef8a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PathCompleter.py.obsolete
+++ /dev/null
@@ -1,102 +0,0 @@
-"""
-PathCompleter  builds paths to a file from a release and  package.
-
-"""
-
-import os, re
-from os.path import join
-from os.path import exists
-
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class PathCompleter:
-    def __init__(self):
-        pass
-        
-    def versionDir(self, package, paths):
-        "returns path to CMT added directory (ie package-xx-yy-zz) "
-
-        path = join(paths.releasePath,package)
-        #
-        # have to jump across the version directory (ie package-xx-yy-zz)
-        # that lies on the path
-        #
-        version = os.listdir(path)[0]
-        path = join(path, version)
-        #return '*'
-        return path
-
-    def fileFromReleasedFile(self, package, file, paths):
-        "returns path to file in a release"
-        
-        path = join(self.versionDir(package, paths), file)
-        logger.debug("path = "+ path)
-        return path
-    
-    def releaseAreaCMTfile(self, package, paths):
-        """
-        returns path to the cmt directory given release and 
-        package name
-        """
-
-        cmtDir = ''
-
-        # The directory to look for is /xx/xxx/x/xx/package-ab-cd-ef/cmt
-
-        for root, dirs, files in os.walk(paths.releasePath):
-            for dir in dirs:
-                if dir.find('-') != -1:
-                    if dir.find(package) != -1:
-                        if  'cmt' in os.listdir(os.path.join(root,dir)):
-                            cmtDir = os.path.join(root,dir,'cmt')
-                            break
-                    else:
-                        break
-            if cmtDir != '': break
-
-        return cmtDir
-    def releaseAreaCMTfile2(self, package, paths):
-        """
-        returns path to the cmt directory given release and 
-        package name
-
-        Timing tests show that releaseAreaCMTfile2 is a little faster than releaseAreaCMTfile
-        """ 
-
-        cmtDir = ''
-
-        # The directory to look for is /xx/xxx/x/xx/package-ab-cd-ef/cmt
-
-        # compPattern=re.compile(package+'-[0-9][0-9]-[0-9][0-9]-[0-9][0-9]')
-        
-        for root, dirs, files in os.walk(paths.releasePath):
-            for dir in dirs:
-                if dir == package:
-                    cmtDir = os.path.join(root, os.listdir(os.path.join(root,dir))[0], 'cmt')
-                    if exists(cmtDir):
-                        return cmtdir
-                    else:
-                        return ""
-        return ''
-
-if __name__ == '__main__':
-    from Paths import Paths
-    from time import time
-    
-    pc=PathCompleter()
-    paths = Paths()
-    paths.releasePath = '/afs/cern.ch/atlas/software/dist/10.0.0'
-
-    start = time()
-    # print pc.releaseAreaCMTfile('AtlfastEvent', paths)
-    end = time()
-    print 'time to run v1 = ',str(end-start)
-
-    start = time()
-    print pc.releaseAreaCMTfile2('AtlfastEvent', paths)
-    end = time()
-    print 'time to run v2 = ',str(end-start)
diff --git a/Tools/RunTimeTester/src/PathConstants.py b/Tools/RunTimeTester/src/PathConstants.py
deleted file mode 100644
index 8bf02f810cac1aebdd2b3acabac3ae47cdfd88ca..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PathConstants.py
+++ /dev/null
@@ -1,250 +0,0 @@
-"""Provide hardcoded path related strings shared by a number of modules.
-
-Trivial (value = name) strings:
-cmt
-share
-bin
-installArea
-
-Ex-Variables - names tht could once take more than one value:
-releaseType
-runType
-
-names that take one value
-cmtConfigFileMacroName
-
-RTT project renaming anomally:
-projectDict
-
-Require updates
-branches list of legal branches
-
-Dangerous Kludges
-projectMap - list of sub projects - use by findJobOptions - FIXME
-"""
-
-import os.path
-
-
-share = 'share'
-cmt = 'cmt'
-
-# historical - still ocurs in summary overview. Alternative was monolithic (?)
-releaseType = 'project'
-
-# historical - still ocurs in summary overview and results paths.
-# Alternative was kit. Note there is a PathNameBuilder.runType
-# (dev,prod,batch,shell) that is not the same thing as this relic.
-runType = 'build'
-
-installArea = 'InstallArea'
-
-# actual name of the macro is <package>_TestConfiguration
-cmtConfigFileMacroName = '_TestConfiguration'
-
-distArea = '/afs/cern.ch/atlas/software/builds'
-
-eosProjSitRTTPath = '/eos/atlas/atlascerngroupdisk/proj-sit/rtt'
-
-eosBinDir = '/afs/cern.ch/project/eos/installation/atlas/bin'
-eosLibDir = '/afs/cern.ch/project/eos/installation/atlas/lib64'
-eos = os.path.join(eosBinDir, 'eos.select')  # alias to the EOS command
-
-
-# The path to the private file containing RTT connection params for AMI.
-# This file is sourced before launching the user job.
-amiConnectionFile = '/afs/cern.ch/user/r/rtt/private/amiConnection.sh'
-
-rttExternals = '/afs/cern.ch/atlas/project/RTT/Externals'
-
-# 31/5/2012. DCubeRunner uses dcubeRefBaseDir as a base dir for data files
-dcubeRefBaseDir = '/afs/cern.ch/atlas/project/RTT/DCubeReference'
-
-# 31/5/2012. DCubeRunner uses dcubeCfgBaseDir as a base dir for cfg files
-dcubeCfgBaseDir = '/afs/cern.ch/atlas/project/RTT/Results/reference/'
-
-rttBases = {'prod':  '/afs/cern.ch/atlas/project/RTT/prod',
-            'dev':   '/afs/cern.ch/atlas/project/RTT/dev'
-            }
-
-projectDict = {'AtlasHLT':        'hlt',
-               'AtlasProduction': 'offline',
-               'AtlasPoint1':     'point1',
-               'AtlasTier0':      'tier0',
-               'AtlasP1HLT':      'p1hlt'}
-
-invProjectDict = {}
-for k, v in projectDict.items():
-    invProjectDict[v] = k
-
-resultsBaseURL = 'https://atlas-rtt.cern.ch/'
-
-projects_2017 = [
-    'Athena',
-    'AthenaP1',    
-    ]
-
-projects = [
-    'AtlasAnalysis',
-    'AtlasLogin',
-    'AtlasSetup',
-    'AtlasConditions',
-    'AtlasOffline',
-    'AtlasSimulation',
-    'AtlasCore',
-    'AtlasProduction',
-    'AtlasTrigger',
-    'AtlasEvent',
-    'AtlasReconstruction',
-    'AtlasHLT',
-    'AtlasPoint1',
-    'AtlasTier0',
-    'AtlasP1HLT',
-    'AtlasCAFHLT',
-    'TopPhys',
-    'TAGProd',
-    'WZBenchmarks',
-    'TrigMC',
-    'AtlasDerivation',
-    'AtlasProd1',
-    'AtlasProd2',
-    'AtlasP1MON',
-    'AthAnalysisBase',
-    'AtlasTestHLT',
-    'AtlasPhysics',
-    'Athena']
-
-branches_2017 = ['master',
-                 '21.0',
-                 '21.0.X',]
-
-
-branches = [
-    'cmake',
-    'dev',
-    'devval',
-    '19.0.X',
-    '19.0.X.Y',
-    '19.0.X-VAL',
-    '19.0.X.Y-VAL',
-    '19.0.X.Y-VAL2',
-    '19.1.X',
-    '19.1.X.Y',
-    '19.1.X.Y-VAL',
-    '19.2.X-VAL',
-    '19.2.X.Y',
-    '19.2.X.Y-VAL',
-    '19.1.X-VAL',
-    '19.1.0.Y-VAL',
-    '19.1.1.Y-VAL',
-    '19.1.3.Y-VAL',
-    '19.3.X.Y',
-    '19.3.X.Y-VAL',
-    '20.X.0',
-    '20.X.0-VAL',
-    '20.0.X.Y',
-    '20.0.X.Y-VAL',
-    '20.1.0.Y',
-    '20.1.0.Y-VAL',
-    '20.1.0.Y-VAL2',
-    '20.1.1.Y',
-    '20.1.1.Y-VAL',
-    '20.1.1.Y-VAL2',
-    '20.1.2.Y-VAL',
-    '20.1.2.Y-VAL2',
-    '20.1.3.Y-VAL',
-    '20.1.3.Y-VAL2',
-    '20.1.4.Y',
-    '20.1.4.Y-VAL',
-    '20.1.5.Y',
-    '20.1.5.Y-VAL',
-    '20.1.5.Y.Z-VAL',
-    '20.1.X',
-    '20.1.X-VAL',
-    '20.1.X.Y',
-    '20.1.X.Y-VAL',
-    '20.1.X.Y.Z',
-    '20.1.X.Y.Z-VAL',
-    '20.2.X',
-    '20.2.X-VAL',
-    '20.2.X.Y',
-    '20.2.X.Y-VAL',
-    '20.2.X.Y.Z-VAL',
-    '20.2.X.Y.Z-VAL2',
-    '20.3.X',
-    '20.3.X-VAL',
-    '20.3.X.Y',
-    '20.3.X.Y-VAL',
-    '20.3.3.Y-VAL',
-    '20.3.4.Y-VAL',
-    '20.3.0.Y-VAL',
-    '20.6.X',
-    '20.6.X-VAL',
-    '20.6.X.Y',
-    '20.6.X.Y-VAL',
-    '20.7.X',
-    '20.7.X-VAL',
-    '20.7.X.Y',
-    '20.7.X.Y-VAL',
-    '20.7.X.Y.Z-VAL',
-    '20.7.1.Y-VAL',
-    '20.7.2.Y-VAL',
-    '20.7.3.Y',
-    '20.7.3.Y-VAL',
-    '20.7.3.Y.Z-VAL',
-    '20.7.4.Y',
-    '20.7.4.Y-VAL',
-    '20.7.5.Y-VAL',
-    '20.7.5.Y.Z-VAL',
-    '20.7.6.Y-VAL',
-    '20.7.8.Y-VAL',
-    '20.7.9.Y-VAL',
-    '20.99.Y-VAL',
-    '20.11.X.Y-VAL',
-    '20.11.X.Y.Z-VAL',
-    '20.20.X',
-    '20.20.X-VAL',
-    '20.20.X.Y-VAL',
-    '21.0.X',
-    '21.0.X-VAL',
-    '21.0.X.Y-VAL',
-    '21.9.X-VAL',
-    '22.0.X-master',
-    '2.0.X',
-    '2.1.X',
-    '2.3.X',
-    '2.4.X',
-    'EXP',
-    'mig0',
-    'mig1',
-    'mig2',
-    'mig3',
-    'mig4',
-    'mig5',
-    'mig6',
-    'mig7',
-    'mig8',
-    'mig9',
-    'mig10',
-    'mig11',
-    'mig12',
-    'mig13',
-    'mig14',
-    'mig15',
-    'devmig1',
-    'devmig2',
-    'devmig3',
-    'devmig4',
-    'devmig5',
-    'devmig6',
-    'lcg4']
-
-
-_dbconnectionDir = '/afs/cern.ch/atlas/project/RTT/dbconnection/'
-cooldozer = '/afs/cern.ch/atlas/project/RTT/prod/Results/dozer/api/latest'
-db_location = {
-    ('rtt', 'prod'): os.path.join(_dbconnectionDir, 'oracle.prod'),
-    ('rttdev', 'dev'): os.path.join(_dbconnectionDir, 'oracle.dev')
-}
-
-branchdir_2017 = '/cvmfs/atlas-nightlies.cern.ch/repo/sw'
diff --git a/Tools/RunTimeTester/src/PathNameBuilder.py b/Tools/RunTimeTester/src/PathNameBuilder.py
deleted file mode 100755
index d381f033cae299fb9c5586e8493a45bd4f67bd33..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PathNameBuilder.py
+++ /dev/null
@@ -1,112 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-PathNameBuilder is a class that builds path names using
-inputs from HardCoded Legals and parameters set by the user
-"""
-
-from os.path import join
-from RTTpath import RTTpath
-
-class PathNameBuilder:
-  def __init__(self, argBag, hardCoded, legals, siteSpecificsObj, logger):
-
-    self.release         = argBag.release
-    self.cmtConfig       = argBag.cmtConfig
-    self.originalBranch  = argBag.originalBranch
-    self.isNightly       = argBag.isNightly
-    self.topProject      = argBag.topProject
-    self.otherProject    = argBag.otherProject
-    self.distArea        = argBag.distArea
-
-    self.hardCoded       = hardCoded
-    self.legals          = legals
-    self.logger          = logger
-
-  def latestCopiedReleaseLink(self, key, topProject):
-    l = join(
-      self.distArea,
-      self.hardCoded.getNightlyPathFragment(key[1]),
-      topProject,
-      self.hardCoded.getLatestCopiedReleaseLinkName(key)
-      )
-    
-    return l
-
-  def pathToFindReleaseDate(self):
-    "needed by the NICOS class"
-
-    path = RTTpath(
-      self.distArea,
-      self.hardCoded.nightlyDirIfAny(self.originalBranch, self.isNightly),
-      self.topProject,
-      self.legals.convertLogicalNightly(self.release),
-      )
-    
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'pathToFindReleaseDate()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-      self.logger.warning('isNightly %s' % self.isNightly)
-    return path
-    
-  def pathToNICOSLatestCopiedLink(self):
-    "needed by StampObject to determine latest nightly release name."
-    
-    key  = (self.cmtConfig, self.originalBranch, self.topProject) 
-    path = RTTpath(self.latestCopiedReleaseLink(key, self.topProject))
-    
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'pathToNICOSLatestCopiedLink()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-      
-    return path
-    
-  def getReleasePath(self, cvsPackage):
-    path = RTTpath(cvsPackage.getPackageCmtPath())
-
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'getReleasePath()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-      
-    return path
-
-  def getInstallArea(self, cvsPackage):
-    path =  RTTpath(str(self.getReleasePath(cvsPackage)),
-                    self.hardCoded.installAreaDirName())
-    
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'getInstallArea()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-        
-    return path
-
-  def getShareArea(self, cvsPackage):
-    path = RTTpath(str(cvsPackage.packageSharePath))
-
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'getShareArea()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-     
-    return path
-    
-  def cmtDir(self, cmtPath):
-    path = RTTpath(cmtPath, self.hardCoded.cmtAreaDirName())
-    
-    if not path.exists():
-      m  = 'PathNameBuilder::'
-      m += 'cmtDir()'
-      m += ' non-existent path  ' + str(path)
-      self.logger.warning(m)
-      
-    return path
-  
diff --git a/Tools/RunTimeTester/src/Paths.py b/Tools/RunTimeTester/src/Paths.py
deleted file mode 100755
index 1d0738eb7ddba78e4c0491d9bf50a2c034612a88..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Paths.py
+++ /dev/null
@@ -1,102 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection             import formatCollection
-from Tools                        import mkdirIfNew
-from extractDCubeBaseDir          import extractDCubeBaseDir
-from RTTSException                import RTTCodingError
-from RTTSException                import RTTInputError
-import os
-# -------------------------------------------------------------------------
-
-class Paths:
-    def __init__(self, noPackagePathsArgBag,
-                 cvsPackage, searchPaths,
-                 eCreator, tCreator, logger):
-
-        self.dCubeCfgFile    = noPackagePathsArgBag.pop('dCubeCfgFile')
-        self.pathNameBuilder = noPackagePathsArgBag.pop('pathNameBuilder')
-
-        self.workDirs        = noPackagePathsArgBag.pop('workDirs')
-        self.resultsDirs     = noPackagePathsArgBag.pop('resultsDirs')
-
-        self.__dict__.update(noPackagePathsArgBag.__dict__)
-        self.logger = logger
-
-        # cvsPackage is of type CVSPackage
-        
-        # Examples:
-        # ---------
-        # package.name = JiveXML
-        # package.pathToXMLfile = /some/base/path/JiveXML_TestConfiguration.xml
-        # package.fullName = graphics/JiveXML
-
-        self.packageXMLfile       = cvsPackage.packageXMLfile # instance of PackageXMLFile.py
-        self.package              = cvsPackage.name
-        self.confFile             = cvsPackage.packageXMLfile.path
-        self.fullPackageName      = cvsPackage.getFullName()
-        self.packageTag           = cvsPackage.getVersion()
-        self.local                = cvsPackage.local
-        self.cmtPath              = str(cvsPackage.getPackageCmtPath())
-        self.cmtDir               = str(self.pathNameBuilder.cmtDir(self.cmtPath))
-        self.installArea          = self.pathNameBuilder.getInstallArea(cvsPackage)
-        self.shareArea            = self.pathNameBuilder.getShareArea(cvsPackage)
-        self.containerPackage     = self.fullPackageName.split(self.cmtPath)[1].strip()[1:]
-        self.searchPaths          = searchPaths
-        try:
-            self.mailTo           = self.packageXMLfile['rtt/mailto[0]/text()']
-        except:
-            self.mailTo = ''
-            
-
-        self.dCubeRefBaseDir      = os.path.join(extractDCubeBaseDir(self.dCubeCfgFile),
-                                                 'offline',
-                                                 self.containerPackage)
-
-        self.elementCreator       = eCreator
-        self.textNodeCreator      = tCreator
-    
-
-        # handle the package information
-        self.releasePath = self.checkCriticalPath(self.pathNameBuilder.getReleasePath(cvsPackage))
-        self.shareArea   = self.checkCriticalPath(self.pathNameBuilder.getShareArea(cvsPackage))
-        self.installArea = self.checkCriticalPath(self.pathNameBuilder.getInstallArea(cvsPackage))
-
-        self.workDirs['package'] = os.path.join(self.workDirs['packages'], self.package)
-        self.resultsDirs['package'] = os.path.join(self.resultsDirs['packages'], self.package)
-    # def dumpHeader(self):
-    #    self.logger.info('=============Paths.dump================')
-        
-    #------------------------------------------------------------------------
-
-    def checkCriticalPath(self, path):
-        
-        if not path.exists():
-            m = 'Critical path: %s does not exist. Aborting run  ' % str(path)
-            self.logger.error(m)
-            raise RTTInputError(m)
-
-        return str(path)
-        
-    #------------------------------------------------------------------------
-
-    def checkPath(self, path):
-        if not path.exists():
-            m = 'Path: %s does not exist. Path creation ' % str(path)
-            self.logger.error(m)
-            try:
-                mkdirIfNew(path)
-                m += 'succeeded'
-            except:
-                
-                m += 'failed'
-
-            self.logger.error(m)
-        # for now, convert back to the old string format so we dont
-        # have to ptopogate RTTpath objects to other modules
-
-        return  str(path)
-
-
-        
-       
-
diff --git a/Tools/RunTimeTester/src/PhasePrinter.py b/Tools/RunTimeTester/src/PhasePrinter.py
deleted file mode 100644
index c716315fb05c4dc6f0a1e1074a6b15c9b29c5091..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PhasePrinter.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class PhasePrinter:
-    def __init__(self):
-        self.phase     = None
-        self.startTime = 0
-
-    def printIt(self, phase):
-        endTime = time.time()
-        if self.phase:
-            exeTime = endTime-self.startTime
-            logger.warning('\n================ %s ends, time = %d =============\n'   % (self.phase, exeTime))
-            
-            logger.warning('\n================ %s starts =============\n' % phase)
-        self.phase = phase
-        self.startTime = endTime
-
diff --git a/Tools/RunTimeTester/src/ProductionManagerVetoer.py b/Tools/RunTimeTester/src/ProductionManagerVetoer.py
deleted file mode 100644
index b79de4b44d277569c646f8f13767d330c779b112..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ProductionManagerVetoer.py
+++ /dev/null
@@ -1,147 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-"""
-module contains class ProductionManagerVetoer
-"""
-
-from xml.dom.minidom     import parse
-from Evaluate            import Evaluate
-from validateXMLFile     import Validator
-
-import re
-import os
-
-class ProductionManagerVetoer:
-    """
-    Used to determine if a job should be vetoed
-    according to identifiers provided in an external
-    file. Identiferes include the job hash identifier,
-    and the job CPC values. All of these are
-    attributes of the job descriptor.
-    """
-
-    def __init__(self, pathToVetoFile, logger):
-        "read the file with the veto information"
-
-        self.idsToVeto  = []
-        self.cpcsToVeto = {}
-
-        if not pathToVetoFile:
-            logger.debug('No Job veto file provided')
-            return
-
-        if not os.path.exists(pathToVetoFile):
-            msg = 'non existent Job veto file: %s', pathToVetoFile
-            logger.warning(msg)
-            return
-
-        self.validator = Validator(pathToVetoFile)
-        isValid, excInst = self.validator.validate()
-        if not isValid:
-            msg = 'File %s failed validation, no jobs vetoed' % str(pathToVetoFile)
-            logger.warning(msg)
-            return
-
-        documentElement = self.validator._dom.documentElement
-        self.setCPCs(documentElement)
-        self.setIDs(documentElement)
-
-    def setCPCs(self, element):
-        def loadTag(tagName, parentEl):
-            els = Evaluate(tagName, parentEl)
-            if not els: return [re.compile('[a-zA-Z0-9]*')]            
-            return [re.compile(Evaluate('text()', el)) for el in els]
-
-        cpcEl = Evaluate('CPC', element)        
-        if not cpcEl: return
-        
-        self.cpcsToVeto = {}        
-        for tag in ['displayClass', 'displayProcess', 'displayComponent']:
-            self.cpcsToVeto[tag] = loadTag(tag, cpcEl[0])
-
-    def setIDs(self, element):
-        idsEl = Evaluate('ids', element)
-        if not idsEl: return
-        ids = Evaluate('id', idsEl[0])
-        for id in ids:
-            try:
-                self.idsToVeto.append(int(Evaluate('text()', id)))
-            except:
-                continue
-
-    def idVetoed(self, jDesc):
-        if not self.idsToVeto: return False
-
-        if int(jDesc.hashString) in self.idsToVeto:
-            jDesc.logger.debug('Job %s: has hash ID %s which matches a vetoed ID. Is thus vetoed!' % (jDesc.identifiedName, jDesc.hashString))
-            return True
-
-        m = 'Job %s: does not match any vetoed IDs, so is not vetoed' % jDesc.identifiedName
-        jDesc.logger.debug(m)
-        # jDesc.logger.debug('%s %s' % (str(self.idsToVeto), jDesc.hashString))
-        return False
-
-    def cpcVetoed(self, jDesc):
-        def isMatched(vetoFileValues, userFileValues):
-            for uf in userFileValues:
-                if True not in [re.match(vf, uf)!=None for vf in vetoFileValues]: return False
-            return True
-
-        # not all descriptors (ie watchers) have CPCs. Do not make
-        # CPC tests for such descriptors. If the attrbute displayClass
-        # is present, so must be displayProcess, displayComponent
-        if not hasattr(jDesc, 'displayClass') or not self.cpcsToVeto:
-            jDesc.logger.debug('Returning False (%s)' % self.cpcsToVeto)
-            return False
-
-        matches = isMatched(self.cpcsToVeto['displayClass'], jDesc.displayClass)
-        matches = matches and isMatched(self.cpcsToVeto['displayProcess'],   jDesc.displayProcess)
-        matches = matches and isMatched(self.cpcsToVeto['displayComponent'], jDesc.displayComponent)
-        m = 'Job %s: does not match all CPC veto flags, is thus not vetoed' % jDesc.identifiedName
-        if matches: m = 'Job %s: matches all CPC veto flags. Is thus vetoed!' % jDesc.identifiedName
-
-        jDesc.logger.debug(m)
-        return matches
-        
-    def toBeVetoed(self, jDesc):
-        return self.idVetoed(jDesc) or self.cpcVetoed(jDesc)
-        
-
-if __name__ == "__main__":
-
-    from PseudoLogger import PseudoLogger
-    class Bag:
-        pass
-    
-    paths = Bag()
-    pathToVetoFile = '/afs/cern.ch/atlas/project/RTT/data/brinick/veto.xml'
-    logger = PseudoLogger()
-    jv = ProductionManagerVetoer(pathToVetoFile, logger)
-
-    desc = Bag()
-    desc.hashString       = '-689235331'
-    desc.displayClass     = ['displayProcess2']
-    desc.displayProcess   = ['Core']
-    desc.displayComponent = ['Athena-Core']
-    desc.logger = PseudoLogger()
-    desc.identifiedName = 'job1'
-
-    # print 'ids:'
-    # print 'desc: ', desc.hashString
-    # print jv.idsToVeto
-    print
-    # print 'CPCs'
-    # print 'desc: ', desc.displayClass, desc.displayProcess, desc.displayComponent
-    # print jv.cpcsToVeto
-    print 
-    # print 'match result', jv.toBeVetoed(desc)
-
-
-
-
-
-
-
-    
-    
diff --git a/Tools/RunTimeTester/src/PseudoDescriptor.py b/Tools/RunTimeTester/src/PseudoDescriptor.py
deleted file mode 100644
index 7826f13081a0e10f7ca9d46aa936484b6977f0f4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PseudoDescriptor.py
+++ /dev/null
@@ -1,15 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from AthenaJobDescriptor import AthenaJobDescriptor
-import os
-
-class PseudoDescriptor(AthenaJobDescriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-        #dict = {'name':     'pseudoDescriptor',
-        #        'jobGroup': 'Athena',
-        #        'jobOptions': ['pseudoJobOptions1.py']
-        #        }
-        #argDict.update(dict)
-        AthenaJobDescriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-        #self.__dict__['runPath'] =  '/User/peter/RTTdevelopment/RTTChain/Work/pseudoRunPath'
-        
diff --git a/Tools/RunTimeTester/src/PseudoJobMinder.py b/Tools/RunTimeTester/src/PseudoJobMinder.py
deleted file mode 100644
index 6625f5de61c1479669a8b61586095bf63861e100..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PseudoJobMinder.py
+++ /dev/null
@@ -1,131 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tracer            import Tracer
-from PseudoDescriptor  import PseudoDescriptor
-from WorkerJobMinder   import WorkerJobMinder
-from RTTSException     import RTTCodingError
-
-
-class PseudoChainFileCopier:
-    def copyFiles(self): return
-    
-class PseudoJobMinder(WorkerJobMinder, Tracer):
-    def __init__(self, argBag, desc):
-
-        commander = argBag.pop('commander')
-        argBag.postProcessCommand = 'pseudoPostProcessCommand'
-        argBag.submitCommand      = 'pseudoSubmitCommand'
-
-        WorkerJobMinder.__init__(self, argBag, desc)
-
-        self.nRetry       = 0
-        self.nIsRunning   = 0
-        self.finishedTests = []
-        self.chainFileCopier = PseudoChainFileCopier()
-        
-    def fire(self, rs): print 'script',rs
-    def checkJobSubmissionWasOK(self): return True
-    def checkForWallTimeTimeOut(self):return False
-    def jobIsRunning(self):
-        if self.nIsRunning<2:
-            self.nIsRunning += 1
-            return (True, True)
-        
-        return  (True, False)
-    def audit(self): pass
-    def areLogFilesInRunDir(self): return True
-    def checkForLogFileTimeOut(self): return False
-    def checkLog(self): return 
-    def runPostScripts(self): return
-    def runMoniActions(self): return
-    def runChecks(self): return
-    def copyKeepFiles(self): return
-    def setProtection(self): return
-    def calcPostProcessingResult(self): return
-    def retryTest(self):
-        self.nRetry += 1
-        if self.nRetry>2: return False
-        return True
-    def doPostProcessing(self): return  False
-    def prepareForRetry(self): pass
-    def cleanSpace(self): pass
-    def stopJob(self): pass
-    def handleBatchError(self): pass
-    def setDone(self):
-        print 'setDone for job ', self.identifiedName
-        self.done = True
-    def isDone(self):
-        return self.done
-    def isSuccess(self):
-        return True
-    # def poll(self):
-    #    self.stateEngine.poll()
-
-    def __str__(self):
-        s = 'PseudoMinder Dump: empty\n'
-        return s
-    def dump(self):
-        self.logger.debug(self.__str__())
-
-    def readTestResults(self):
-        return
-    
-    def fromQueued(self):
-        self.fire('runScript')
-        return 'running'
-
-    def fromRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-
-        if not queryOK:
-            msg = 'Unsuccesful query of a job running interactively'
-            raise RTTCodingError(msg)
-
-        if isRunning:  return 'running' # no state transition
-
-        self.fromRunningActions()
-        
-        if self.doPostProcessing():
-            return 'postProcessingQueued'
-        
-        return 'done'
-
-    def fromRunningActions(self):
-
-        self.checkLog()
-        
-        self.runChecks()
-        # chainFileCopier present only if running in a chain 
-        if hasattr(self, 'chainFileCopier'):
-            self.chainFileCopier.copyFiles()
-
-        # if keep file is also a chain file, this code  will silently fail. Intended that the chain
-        # file copier is interogated to find new location
-        self.copyKeepFiles()
-
-    def fromPostProcessingQueued(self):
-        self.fire('postProcessorScript')
-        return 'postProcessingRunning'
-            
-    def fromPostProcessingRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-
-        if not queryOK:
-            msg = 'Unsuccesful query of a job running interactively'
-            raise RTTCodingError(msg)
-
-        if isRunning:  return 'postProcessingRunning'
-
-        self.readTestResults()
-        self.copyKeepFiles()
-        self.cleanSpace()
-
-        return 'done'
-        
-    def fromOperatorCloseDown(self): return 'done'
-    def fromRetry(self): return 'done'
-    def fromBatchError(self): return 'done'
-    def fromTimedOut(self): return 'done'
-    def fromDone(self):
-        self.setDone()
-        return 'done'
diff --git a/Tools/RunTimeTester/src/PseudoLogger.py.obsolete b/Tools/RunTimeTester/src/PseudoLogger.py.obsolete
deleted file mode 100755
index 2f78133b6babe2ead22558508b8799f48dbff507..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/PseudoLogger.py.obsolete
+++ /dev/null
@@ -1,7 +0,0 @@
-class PseudoLogger:
-    def debug(self, str): print 'DEBUG: ',str
-    def info(self, str): print 'INFO: ',str
-    def warning(self, str): print str
-    def error(self, str): print str
-    def critical(self, str): print str
-    
diff --git a/Tools/RunTimeTester/src/Querier.py b/Tools/RunTimeTester/src/Querier.py
deleted file mode 100755
index de52ffd55f4c3b50a68b865882bba913ed15f78a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Querier.py
+++ /dev/null
@@ -1,119 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A series of classes that queries something.
-The query is a shell command.
-Replies can be expected and OK, undesired but known about, or a complete
-surprise.
-
-This module also has specialisation classes for querying batch queues,
-and which have different behaviours if the answer to the query is not the
-desired one.
-"""
-
-# import commands
-import time, string
-from ShellCommand import ShellCommand
-
-class Querier:
-    def __init__(self, logger, queryCmd):
-        self.queryCmd  = queryCmd
-        self.status    = 0
-        self.output    = ''
-        self.logger    = logger
-
-    def query(self):
-
-        try:
-            sc = ShellCommand([self.queryCmd], self.logger, 100, None, True)
-        except:
-            self.status = -1
-            self.output = 'query %s failed' % self.queryCmd
-            return
-        
-        self.status = 0
-        self.output = string.join(sc.getReply(), '\n')
-        # self.status, self.output = commands.getstatusoutput('bjobs -q atlasrttmedium')
-
-    def statusAndReply(self):
-        if self.status:
-            return (False, 'bad command status: %s output: %s' % (str(self.status), self.output))
-        else:
-            return (True, self.output)
-        
-class SleepingQuerier(Querier):
-    def __init__(self, logger, queryCmd, maxtry, sleepTime):
-        
-        self.maxtry    = maxtry
-        self.sleepTime = sleepTime
-        Querier.__init__(self, logger, queryCmd)
-
-
-    def statusAndReply(self):
-        
-        mtry = self.maxtry
-        while mtry:
-            self.query()
-            ok, comment = Querier.statusAndReply(self)
-            if not ok:
-                self.logger.debug('cmd %s: query failure %d sleeping %d' % (self.queryCmd,
-                                                                            self.maxtry-mtry,
-                                                                            self.sleepTime))
-                self.logger.debug('status %s reply %s' % (ok, comment))
-                time.sleep(self.sleepTime)
-                mtry -= 1
-            else:
-                return (ok, comment)
-
-        return (ok, comment)
-
-
-class LSF_bjobs1(SleepingQuerier):
-    def __init__(self, logger, queue, maxtry = 3, sleepTime = 5):
-        
-        queryCmd = 'bqueues ' + queue
-        SleepingQuerier.__init__(self, logger, queryCmd, maxtry, sleepTime)
-
-    def statusAndReply(self):
-        (status, comment) = SleepingQuerier.statusAndReply(self)
-        if not status: return (status, comment)
-        if self.output.find('No such queue')<0: return (True, self.output)
-        return (False, self.output)
-
-class LSF_bjobs2(SleepingQuerier):
-    def __init__(self, logger, pid, maxtry = 3, sleepTime = 5):
-        
-        queryCmd = 'bjobs -l ' + str(pid) + ';'
-        SleepingQuerier.__init__(self, logger, queryCmd, maxtry, sleepTime)
-
-    def statusAndReply(self):
-        (status, comment) = SleepingQuerier.statusAndReply(self)
-        if not status: return (status, comment)
-        if not self.output.find('The CPU time used is'): return (False, 'noCPUTime')
-        return (True, self.output)
-
-class LSF_bjobs3(SleepingQuerier):
-    def __init__(self, logger, pid, maxtry = 3, sleepTime = 5):
-        
-        queryCmd = 'bjobs -a ' + str(pid) + ';'
-        SleepingQuerier.__init__(self, logger, queryCmd, maxtry, sleepTime)
-
-    def statusAndReply(self):
-        (status, comment) = SleepingQuerier.statusAndReply(self)
-        if not status: return (status, comment)
-        tokens = self.output.split('\n')
-        if len(tokens)<2: return (False, 'reply has bad format')
-        tokens = tokens[1].split()
-        if len(tokens)<3: return (False, 'reply has bad format')
-        return (True, tokens[2])
-
-
-class LSF_bhist(SleepingQuerier):
-    def __init__(self, logger, pid, maxtry = 3, sleepTime = 5):
-        
-        queryCmd = 'bhist ' + str(pid) + ';'
-        SleepingQuerier.__init__(self, logger, queryCmd, maxtry, sleepTime)
-
-    def statusAndReply(self): return SleepingQuerier.statusAndReply(self)
-
-
diff --git a/Tools/RunTimeTester/src/RTTCloseDownKeyMaker.py b/Tools/RunTimeTester/src/RTTCloseDownKeyMaker.py
deleted file mode 100755
index c02e6604a816a93ae1950d16d5cb6bb73975f188..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTCloseDownKeyMaker.py
+++ /dev/null
@@ -1,314 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to manage the closeDown keys for the RTT.
-a closedownKey os is a dictionary that has been tunrd into an xml file.
-The dictionay must contain an entry 'startTime'.
-
-Tester appends such a key to a file closeDownKeyFile. Ongoing RTT TestRuns
-periodiocaly check to see if the they should shutdown. They should do so if
-there is a key that is more uptodate (ie its startTime is later than the time
-associated with the tester, and all entries in the closeDownKey are equal
-to a corresponding key in the TestRun key. This mismatching allows selective
-shutdowns of test tuns.
-
-Note that the closeDownKey need not have all the entries in the TestRunKey.
-This, providing a closeDown key with just the time will close down all ongoing
-TestRuns (no selectivity).
-
-To orevent the closeDownKeyFile growing, when a new key needs to be written
-the old keys are removed.
-
-
-The entrance methods are
-laterMatch()     -  to trigger shutdowns
-writ((dict=None) - to write out new keys 
-"""
-
-
-import os, xml.dom, sys, copy
-from StringIO          import StringIO
-from makeRTTDOM        import makeRTTDOM
-from exc2string2       import exc2string2
-from Evaluate          import Evaluate, getText
-from xml.dom.minidom   import parseString
-from xml.dom.ext       import PrettyPrint
-# -------------------------------------------------------------------------
-
-def mismatch(dict, refdict):    
-    keys = [ k for k in dict.keys() if k in refdict.keys()]
-
-    if 'startTime' in keys: keys.remove('startTime')
-    if not keys: return False # no mismatch if no more keys
-    # look for a mismatch (allows no entries to shutdown all testers)
-    for k in keys:
-        if dict[k] != refdict[k]:
-            return True
-    return False
-        
-# ------------------------------------------------------------------------
-
-
-# obsolete code kept while we still use the (almost) obsolete
-# module Minder.py
-def dictToDocument(dict, document):
-    keyNode = document.createElement('key')        
-
-
-    def addTextNode(parentNode, item):
-        try:
-            tn = document.createTextNode(str(item[1]).strip("'").strip())
-        except xml.dom.InvalidCharacterErr:
-            tn = document.createTextNode('invalid text')
-            
-        n  = document.createElement(str(item[0]).strip("'").strip())
-        
-        n.appendChild(tn)
-        parentNode.appendChild(n)
-
-
-    [addTextNode(keyNode, item) for item in dict.items()]
-    document.documentElement.appendChild(keyNode)
-
-    return document
-
-
-# ------------------------------------------------------------------------
-
-class RTTCloseDownKeyMaker:
-    def __init__(self, argBag):
-
-        self.logger        = argBag.pop('logger')
-        self.closeDownFile = argBag.pop('closeDownFile')
-        self.isNightly     = argBag.pop('isNightly')
-        self.testerStart   = argBag.pop('testerStart')
-        
-        self.__dict__.update(argBag.__dict__)
-        self.document         = None
-        self.docElement       = None
-    # ------------------------------------------------------------------------
-
-    def laterMatch(self, dict):
-
-        if 'startTime' not in dict.keys(): return False
-
-        # select keys that occur chronologically after the current key
-        dictsFromFile = self.read()
-        dictsFromFile = [d for (d,e) in  dictsFromFile]
-        dictsFromFile = [d for d in dictsFromFile if 'startTime' in d.keys()]
-        dictsFromFile = [d for d in dictsFromFile
-                         if float(d['startTime'])>float(dict['startTime'])]
-
-        if not dictsFromFile: return False
-
-        # there are some later keys.
-        # check to see the numbers of keys do not match
-        toRemove = [d for d in dictsFromFile if mismatch(d, dict)]
-
-        # print 'dictsFromFile %d toRemove %d' % (len(dictsFromFile),
-        #                                      len(toRemove))
-        # if no keys match, there was no later match
-        if len(dictsFromFile) > len(toRemove): return True
-        return False
-
-    # ------------------------------------------------------------------------
-
-    def read(self):
-        """
-        returns the information in the close down keys file in the form
-        of a  dictionary and as a DOM node
-        """
-        
-        self.document = self.fileToDocument()
-        self.docElement = self.document.documentElement
-
-        self.docElement = self.document.documentElement
-
-        keyFromFileTags = self.document.getElementsByTagName('key')
-        
-        dicts = []
-        for el in keyFromFileTags:
-            dict = {}
-            childTags = el.childNodes
-            childTags = [e for e in childTags if e.nodeType == e.ELEMENT_NODE]
-            def fillDict(e, dict):
-                dict[e.nodeName] = getText(e)
-            [fillDict(e, dict) for e in childTags]
-
-            if 'startTime' in dict.keys():
-                dict['startTime'] = float(dict['startTime'])
-            dicts.append( (dict, el) )
-
-
-        return dicts
-
-    # ------------------------------------------------------------------------
-
-    def write(self, dictToAdd=None):                
-
-        if dictToAdd is None:
-            dictToAdd = self.makeDict()
-        dictsRead = self.read()
-        dicts = self.removeOldKeys(dictsRead, dictToAdd)
-            
-        try:
-            newKeys = [self.dictToNode(d) for (d,n) in dicts]
-        except:
-            self.logger.error('Error adding info to the RTTClosedown key')
-            self.logger.error(exc2string2())
-            return
-
-        oldKeys = self.document.getElementsByTagName('key')
-        [self.document.documentElement.removeChild(n) for n in oldKeys]
-        [self.document.documentElement.appendChild(n) for n in newKeys]
-        [n.unlink() for n in oldKeys]
-        try:
-            self.xmlToFile()
-        except:
-            
-            m = 'Unable to write RTTCloseDownKey to file (%s)' % (
-                str(self.closeDownFile))
-            m += 'Traceback:\n%s' % (exc2string2())
-            self.logger.error(m)
-
-        m = 'Wrote an  RTTCloseDownKey to file (%s)' % (
-            str(self.closeDownFile))
-        self.logger.info(m)
-
-        
-    # ------------------------------------------------------------------------
-    
-    def fileToDocument(self):
-        'Parses close down key file, returns a DOM document '
-        
-        if os.path.exists(self.closeDownFile):
-            ifile = open(self.closeDownFile, 'r')
-            temp = ifile.read()
-            ifile.close()
-        else:
-            return  makeRTTDOM('RTTCloseDownKey')
-            
-        try:
-            return parseString(temp)
-        except:
-            m = 'Could not parse  %s' % str(self.closeDownFile)
-            m += 'traceback\n%s' % exc2string2()
-            self.logger.error(m)
-
-        return ''
-
-    # ------------------------------------------------------------------------
-
-    def removeOldKeys(self, dictsIn, newDict):
-        # prevent pile-up of old keys
-
-        dicts = []
-
-        # keep dicts which mismatch - they are from RTT runs
-        # with different parameters
-        [dicts.append( (d, node) ) for (d, node) in dictsIn
-                       if mismatch(d, newDict)]
-
-        what = 'dictionary'
-        number = len(dictsIn)-len(dicts)
-        if number != 1: what = 'dictionaries'
-        self.logger.info('removed %d %s' % (number, what))
-
-        dicts.append( (newDict, self.dictToNode(newDict)) )
-        
-        return dicts
-    # ------------------------------------------------------------------------
-
-    def makeDict(self):
-
-        els = ['cmtConfig', 'originalBranch', 'release',
-               'releaseType', 'runType', 'topProject', 'otherProject']
-
-        dict = {}
-        for el in els:
-            dict[el] = self.__dict__[el]
-
-        if self.isNightly: dict['release'] = 'nightly'
-
-        # subtract 0.5 seconds from start time. The time is written to a
-        # file and, due to rounding errors, the it may be in advance of
-        # the same time passed to TestRun, which will cause a shutdown.
-        startTime = self.testerStart-0.5
-        dict['startTime'] = startTime
-
-        return dict
-    
-    # ------------------------------------------------------------------------
-
-    def docToPrettyString(self):
-        buffer = StringIO()
-        PrettyPrint(self.document, buffer)
-        return buffer.getvalue()
-        
-    # ------------------------------------------------------------------------
-
-    def xmlToFile(self):
-
-        of = open(str(self.closeDownFile), 'w')
-        # print '------------'
-        # print 'xmlToFile ', of 
-        # print 'xmlToFile ', self.docToPrettyString()
-        # print '------------'
-        of.write(self.docToPrettyString())
-        of.close
-
-    # ------------------------------------------------------------------------
-
-    def dictToNode(self, dict):
-
-        document = self.document # nested scopes
-        keyNode  = document.createElement('key')        
-
-        
-        def addTextNode(parentNode, item):
-            try:
-                tn = document.createTextNode(str(item[1]).strip("'").strip())
-            except xml.dom.InvalidCharacterErr:
-                tn = document.createTextNode('invalid text')
-            
-            n  = document.createElement(str(item[0]).strip("'").strip())
-            n.appendChild(tn)
-            
-            parentNode.appendChild(n)
-
-
-        [addTextNode(keyNode, item) for item in dict.items()]
-
-        return keyNode
-    # ------------------------------------------------------------------------
-
-    def __del__(self):
-        if self.document: self.document.unlink()
-            
-if __name__ == '__main__':
-    import time
-    from RTTpath import RTTpath
-    class JLogger:
-        def debug(self, s): print 'debug',s
-        def info(self, s): print 'info', s
-        def warning(self, s): print 'warning',s
-        def error(self, s): print 'error',s
-        def crtical(self, s): print 'crtitical',s
-
-    class JPaths:
-        def __init__(self):
-
-            self.cmtConfig      = 'x86_64-slc4-gcc34-opt'
-            self.originalBranch = 'N.X.0'
-            self.release        = '12.0.2'
-            # self.release        = 'atlrel_4'
-            self.releaseType    = 'project'
-            self.runType        = 'release'
-            self.isNightly      = False
-            # self.isNightly      = True
-            self.closeDownFile  = RTTpath('./closeDownFile.xml')
-
-    logger = JLogger()
-    a = RTTCloseDownKeyMaker(JPaths(), time.time(), logger)
-    a.write()
-    # a.write({'startTime': 1159868860.})
diff --git a/Tools/RunTimeTester/src/RTTConfigParser.py b/Tools/RunTimeTester/src/RTTConfigParser.py
deleted file mode 100755
index 1aa30872e2cfb4925016f579ca0fd40eb816cd6f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTConfigParser.py
+++ /dev/null
@@ -1,675 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Evaluate            import Evaluate, getText
-from xml.dom.minidom     import parse
-from validateXMLFile     import Validator
-from Tools               import mkdirIfNew
-from Tools2              import fixSortBranches
-from formatCollection    import formatCollection
-from Factory_Legals      import LegalsFactory
-from Factory_HardCoded   import HardCodedFactory
-from Archiver            import ArchiverFactory, ArchivePathSelectorFactory, ArchivingCondition
-from Tools2              import isValidReleaseNumber
-from Tools               import machineNameToIPaddress
-from RTTSException       import RTTInputError
-from Factory_IOUtils     import IOUtilsFactory
-
-import os.path
-
-import logging 
-logger = logging.getLogger('rtt')
-
-# -------------------------------------------------------------------------
-
-def handleConfFiles(topElement, config):
-    """ make a
-
-dictionary confiles:
-key: package val:dict
-
-dictionary dict:
-key: conFilePath val: absolute path for package test configuration file
-key: cmtDir      val: absolute path to the package cmt directory 
-"""
-    confFiles  = {}
-    packageElements = Evaluate('packageConfFile', topElement)
-    for element in packageElements:
-        package = Evaluate('package/text()', element)
-        
-        dict = {}
-        dict['confFilePath'] = Evaluate('pathToFile/text()', element)
-        dict['cmtPath']      = Evaluate('cmtpath/text()', element)
-        dict['version']      = Evaluate('version/text()', element)
-        dict['installArea']  = Evaluate('installArea/text()', element)
-        dict['shareArea']    = Evaluate('shareArea/text()', element)
-        confFiles[package] = dict
-    config['confFiles'] = confFiles
-
-#----------------------------------------------------------------------
-
-def handleChainStoreBasePath(topElement, config):
-    area = Evaluate('chainStoreBasePath', topElement)
-    if area:
-        config['chainStoreBasePath'] = Evaluate('text()', area[0])
-
-#----------------------------------------------------------------------
-
-def handleConfFileSrc(topElement, config):
-
-    # place a dictionary {leaf package: absolute path to conf file}
-    # of rtt provided configuration files in the conf dictioanry 
-    packagesFromRTT = Evaluate('confFileSrc/rttProvides', topElement)
-    if packagesFromRTT:
-        config['rttProvidesConfFiles'] = True
-        handleConfFiles(packagesFromRTT[0], config)
-
-    localPackageTags = Evaluate('confFileSrc/localConfFiles', topElement)
-
-    # flag whether the config file declares local packages (ie checked out
-    # packahes, and make a list of packages so declared> Both flag and list
-    # are placed in the conf dictionary.
-    config['localConfFiles']     = False
-    if localPackageTags:
-        config['localConfFiles'] = True
-
-        config['userHomeCmtDir'] = Evaluate(
-            'confFileSrc/localConfFiles/userHomeCmtDir/text()', topElement)
-        config['testProjectCmtDir'] = Evaluate(
-            'confFileSrc/localConfFiles/testProjectCmtDir/text()', topElement)
-
-    # flag whether the release packages should be accessed.
-    installed = Evaluate('confFileSrc/installedConfFiles', topElement)
-    if installed: config['installedConfFiles'] = True
-            
-#----------------------------------------------------------------------
-
-def handleSize(sizeEls):
-    def check(sizeEls):
-        if not sizeEls: return ''
-        size = sizeEls[0]
-        try:
-            size = float(Evaluate('text()', size))
-        except:
-            m = '%s: illegal value for a size in bytes.' % size
-            raise RTTInputError(m)
-        else:
-            if not size > 0.0:
-                m = 'Must provide a file size to match > 0.0 bytes! (you gave %s)' % str(size)
-                raise RTTInputError(m)
-            else:
-                return size
-            
-    if not sizeEls: return {}
-    size_el = sizeEls[0]
-    size_gt = check(Evaluate('gt', size_el))
-    size_lt = check(Evaluate('lt', size_el))
-    if (size_gt and size_lt) and (size_gt <= size_lt):
-        m = 'Cannot simultaneously match files larger than X bytes _and_ less than Y bytes where X > Y.'
-        raise RTTInputError(m)        
-            
-    dict = {}
-    if size_lt and not size_gt:
-        dict = {'lt':size_lt}
-    elif size_gt and not size_lt:
-        dict = {'gt':size_gt}
-    else:
-        dict = {'lt':size_lt, 'gt':size_gt}
-    return dict
-
-def handleName(nameEls):
-    def check(nameEls):
-        if not nameEls: return []
-        return [Evaluate('text()', nameEl) for nameEl in nameEls]
-        
-    if not nameEls: return {}
-    name_el = nameEls[0]
-    dict = {}
-    dict['is_']    = check(Evaluate('is', name_el))
-    dict['isnot_'] = check(Evaluate('isnot', name_el))    
-    return dict
-
-def handleParentDir(parentdirEls):
-    def check(nameEls):
-        if not nameEls: return []
-        return [Evaluate('text()', nameEl) for nameEl in nameEls]
-    
-    if not parentdirEls: return {}
-    parentdirEl = parentdirEls[0]
-    dict = {}
-    dict['is_']    = check(Evaluate('is', parentdirEl))
-    dict['isnot_'] = check(Evaluate('isnot', parentdirEl))    
-    return dict    
-    
-def handleArchive(topElement, config):
-    # Example:
-    #    <archive>
-    #       <type>castor</type>
-    #       <base>/castor/cern.ch/user/j/joebloggs/archive_area</base>
-    #       <nvers>7</nvers>
-    #       <condition>
-    #          <size><gt>100000</gt></size>
-    #          <name><is>*.root</is><isnot>root.*</isnot></name>   
-    #       </condition>
-    #    </archive>
-    # 
-    from createDirectoryDictionary import findSeg
-    archives    = Evaluate('archive', topElement)
-    archiveList = []
-    
-    for archive in archives:
-        numb  = Evaluate('nvers/text()', archive)
-        type_ = Evaluate('type/text()', archive)
-        dest  = Evaluate('base/text()', archive)
-        try:
-            numb = int(numb)
-        except:
-            msg = 'Please give an integer for the number of required archives. You gave: <nvers>%s</nvers>' % numb
-            raise RTTInputError(msg)
-        if numb < 1:
-            msg = 'Require an archive to have >=1 versions (<nvers>), you gave %d' % numb
-            raise RTTInputError(msg)
-
-        utils = IOUtilsFactory(logger).create(type_)
-        if not utils.exists(dest): # complicated statement because we may need to check on Castor
-            msg = 'Archive <dest> %s: inexistant' % dest
-            raise RTTInputError(msg)
-
-        conditions = Evaluate('condition', archive)
-        condsList = []
-        for c in conditions:
-            cList = []
-
-            sizeDict = handleSize(Evaluate('size', c))
-            nameDict = handleName(Evaluate('name', c))
-            parentdirDict = handleParentDir(Evaluate('parentdir', c))
-            
-            if sizeDict.keys(): cList.append(ArchivingCondition.Size(sizeDict))
-            if nameDict.keys(): cList.append(ArchivingCondition.Name(nameDict))
-            if parentdirDict.keys(): cList.append(ArchivingCondition.ParentDir(parentdirDict))
-                                
-            # Now finally append the list of conditions to the master list
-            condsList.append(ArchivingCondition(cList))
-    
-        selector = ArchivePathSelectorFactory(logger).create(type_, dest, numb)
-        
-        prefix = config['release']
-        if prefix.startswith('rel_'): prefix='nightly'
-        run_frag = os.path.join(prefix,                     
-                                config['branch'],
-                                'build',
-                                config['cmtConfig'], 
-                                findSeg(config['topProject']))
-
-        # Select which archive we should use
-        # Archives have path like <base_archive_path>/<branch>/<archiveIndex>/<run_frag>/
-        # where, for example:
-        #     branch=15.2.X.Y,
-        #     archiveIndex=N (where 0<=N<size of archive (7 days for ex.))
-        #     run_frag=nightly/dev/build/i686-slc4-gcc34-opt/offline
-        # RTT when it chooses an archive, dumps an empty file called time_stamp below
-        # <run_frag> dir. Next run, by comparing time_stamps, the RTT can select the oldest archive to re-use.
-        archivepath = selector.select_archive(run_frag)
-        archiveList.append(ArchiverFactory(logger).create(type_, archivepath.split(run_frag)[0].strip(),
-                                                          run_frag, condsList))    
-    config['archivers'] =  archiveList
-
-#----------------------------------------------------------------------
-
-def handleBigFiles(topElement, config):
-    bigFilesElement = Evaluate('bigFiles', topElement)
-    config['bigFilesArea'] = ''
-    config['bigFilesSize'] = 0
-    
-    if not bigFilesElement: return
-    config['bigFilesArea'] = Evaluate('bigFilesArea/text()', bigFilesElement[0])
-    config['bigFilesSize'] = int(Evaluate('bigFilesSize/text()', bigFilesElement[0])) # size in MB
-
-#----------------------------------------------------------------------
-
-def handleAcceptPackage(topElement, config):
-    acceptElements = Evaluate('packageAccept', topElement)
-    acceptPackages = []
-    [acceptPackages.append(getText(element)) for element in acceptElements] 
-    config['packagesAccept'] = acceptPackages
-
-#----------------------------------------------------------------------
-
-def handleVetoPackage(topElement, config):
-    vetoElements = Evaluate('packageVeto', topElement)
-    vetoPackages = []
-    [vetoPackages.append(getText(element)) for element in vetoElements] 
-    config['packagesVeto'] = vetoPackages
-
-#----------------------------------------------------------------------
-
-def handleLibToolsLoc(topElement,config):
-    libTools = Evaluate('libToolsLoc', topElement)
-    if libTools: config['libToolsLoc'] = getText(libTools[0])
-
-#----------------------------------------------------------------------
-
-def handleMaxTime(topElement, config):
-    
-    hours=mins=secs=0
-
-    texthours   = Evaluate('maxTime/hours',   topElement)
-    if texthours:
-        texthours   = getText(texthours[0])
-        try:
-            hours = int(texthours)
-        except:
-            logger.error('error in hours spec '+texthours)
-
-
-    if hours < 0:
-        msg  = 'time spec out of range (hours) = '+ str(hours)
-        logger.error(msg)
-        hours = 0
-        
-    textmins   = Evaluate('maxTime/minutes',   topElement)
-    if textmins:
-        textmins   = getText(textmins[0])
-        try:
-            mins = int(textmins)
-        except:
-            logger.error('error in mins spec '+textmins)
-
-
-    if mins < 0 or mins >59:
-        msg  = 'time spec out of range (mins) = '+ str(mins)
-        logger.error(msg)
-        mins = 0
-        
-    total = hours*3600+mins*60
-    if total > 0: config['maxTime'] = total
-    return
-
-#----------------------------------------------------------------------
-
-def handleTopProject(topElement, config):
-    config['topProject'] = Evaluate('topProject/text()', topElement)
-    
-#----------------------------------------------------------------------
-
-def handleOtherProject(topElement, config):
-    # DTD guarantees presence
-    config['otherProject'] = Evaluate('otherProject/text()', topElement)
-
-#----------------------------------------------------------------------
-
-def handleRunType(topElement, config):
-    # DTD guarantees either release or kit
-    config['runType'] = 'build'
-    elements = Evaluate('runType/kitRun', topElement)
-    if elements:
-        config['runType'] = 'kit'
-        installAreaElements = Evaluate('runType/kitRun/installArea',topElement)
-        
-#----------------------------------------------------------------------
-
-def handleReleaseType(topElement, config):
-    releaseTypeElements = Evaluate('releaseType', topElement)
-    if releaseTypeElements: # DTD ensures 0 or 1 such elements
-        config['releaseType'] = getText(releaseTypeElements[0])
-    else:
-        config['releaseType'] = 'project' # default
-        
-                                        
-#----------------------------------------------------------------------
-
-def handlePilotJobFlag(topElement, config):
-
-    config['rttPilotJobs'] = False # default
-    config['atnPilotJobs'] = False # default
-
-    # DTD ensures 0 or 1 such elements
-    if Evaluate('rttPilotJobs', topElement): config['rttPilotJobs'] = True
-    if Evaluate('atnPilotJobs', topElement): config['atnPilotJobs'] = True
-                                        
-#----------------------------------------------------------------------
-
-def handleConcurrentFlag(topElement, config):
-    config['concurrent'] = False # default
-
-    # DTD ensures 0 or 1 such elements
-    if Evaluate('concurrent', topElement):
-        config['concurrent'] = True
-        config['wpcServer']  = machineNameToIPaddress(Evaluate('concurrent/wpcServer/text()',topElement))
-
-#----------------------------------------------------------------------
-
-def handleBooleanFlag(flagname, topElement, config):
-    config[flagname] = False
-    if Evaluate(flagname, topElement): config[flagname] = True
-
-#----------------------------------------------------------------------
-
-def handleRTTRunFlag(topElement, config):
-    handleBooleanFlag('rttRun', topElement, config)
-
-def handleTCTRunFlag(topElement, config):
-    handleBooleanFlag('tctRun', topElement, config)
-
-def handleFCTRunFlag(topElement, config):
-    handleBooleanFlag('fctRun', topElement, config)
-
-#----------------------------------------------------------------------
-
-def handleDisableMailToFlag(topElement, config):
-    handleBooleanFlag('disableMailTo', topElement, config)
-                                        
-#----------------------------------------------------------------------
-
-def handleDeleteNonKeepFiles(topElement, config):
-    handleBooleanFlag('deleteNonKeepFiles', topElement, config)
-
-#----------------------------------------------------------------------
-
-def handleDevRunFlag(topElement, config):
-    handleBooleanFlag('devRun', topElement, config)
-
-#----------------------------------------------------------------------
-
-def handleSandboxRunning(topElement, config):
-    handleBooleanFlag('sandboxRunning', topElement, config)
-
-#----------------------------------------------------------------------
-
-def ensureInequalityWorkResBase(workBase,resBase):
-    msg = ''
-    #if workBase == resBase:
-    #   msg  = 'You have given the same path in <workBasePath> and '
-    #   msg += '<resultsBasePath>. This is forbidden! Exiting...\n'
-
-    return msg
-
-#----------------------------------------------------------------------
-
-def checkPaths(config):
-    
-    pathsToMake = ['workBasePath',
-                   'resultsBasePath'
-                   ]
-
-    [mkdirIfNew(config[p]) for p in pathsToMake]
-
-    pathsThatMustExistAndBeAbs = ['workBasePath',
-                                  'resultsBasePath',
-                                  'jobGroupConfig',
-                                  'dataSetCatalog',
-                                  'libToolsLoc',
-                                  'dCubeCfgFile',
-                                  'pathToVetoFile']
-    
-
-
-    badPaths = []
-    badPaths.extend([config[xmlTagName]
-                     for xmlTagName in pathsThatMustExistAndBeAbs
-                     if checkPathsCondition(config,xmlTagName)])
-    
-    badPaths.extend([aPath for aPath in config.get('confiles', {}).values()
-    if checkPathsCondition2(aPath)])    
-
-    msg = ''
-    if badPaths: msg = 'Bad paths: %s' % str(badPaths)
-    return msg
-
-#----------------------------------------------------------------------
-
-def checkPathsCondition(config,key):    
-    conditionMet = key in config
-    conditionMet = conditionMet and not isDefault(key,config[key])
-    conditionMet = conditionMet and not (os.path.exists(config[key]) and os.path.isabs(config[key]))
-    return conditionMet
-
-#----------------------------------------------------------------------
-
-def checkPathsCondition2(aPath):
-    conditionMet = not (os.path.exists(aPath) and os.path.isabs(aPath))
-    return conditionMet
-
-#----------------------------------------------------------------------
-
-def checkPathsCondition3(config,key):
-    bad = False
-    if not  key in config: return bad
-
-    isRel         = not os.path.isabs(config[key])
-    # doesntExist    = not os.path.exists(config[key])
-    # bad = isRel or doesntExist
-    bad = isRel
-    return bad
-
-#----------------------------------------------------------------------
-
-def loadSimplePathsIntoConfig(documentElement,config):
-    simplePaths = [
-        'mode', 'release', 'cmtConfig', 'workBasePath', 'resultsBasePath',
-        'site', 'dataSetCatalog', 'branch',
-        'jobGroupConfig', 'versionDir','distArea','libToolsLoc','confFileSuffix',
-        'resultsBaseURL', 'devRun', 'dCubeCfgFile', 'pathToVetoFile']
-    
-    # DTD guarantees only a single element
-    for path in simplePaths:
-        elements = Evaluate(path, documentElement)
-        if elements:
-            element=elements[0]
-            config[path]  = getText(element)
-
-#----------------------------------------------------------------------
-
-defaults = {'packagesAccept'       : [],
-            'packagesVeto'         : [],
-            'maxTime'              : 60*60*20,
-            'versionDir'           : None,
-            'rttProvidesConfFiles' : False,
-            'localConfFiles'       : False,
-            'installedConfFiles'   : False,
-            'confFileSuffix'       : '_TestConfiguration',
-            'libToolsLoc'          : None,
-            'dCubeCfgFile'         : '',
-            'otherProject'         : '',
-            'pathToVetoFile'       : ''
-            }
-
-# -------------------------------------------------------------------
-
-def addDefaultsToConfig(config):
-    for item in defaults.items():
-        key          = item[0]
-        defaultValue = item[1]
-        config.setdefault(key,defaultValue)
-
-# -------------------------------------------------------------------
-
-def isDefault(key,val):
-    itIs = False
-    try:
-        if defaults[key] == val:
-            itIs = True
-    except:
-        itIs = False
-    return itIs
-
-# -------------------------------------------------------------------
-
-def rttConfiguration(fn):
-
-    Validator(fn).validate() # throws exception if invalid
-        
-    documentElement = parse(fn).documentElement    
-
-    config = {}
-    loadSimplePathsIntoConfig(documentElement, config)
-    
-    handleRunType(documentElement, config)
-    handleTopProject(documentElement, config)
-    handleOtherProject(documentElement, config)
-    handleConfFileSrc(documentElement, config)
-    handleAcceptPackage(documentElement, config)
-    handleVetoPackage(documentElement, config)
-    handleMaxTime(documentElement, config)
-    handleLibToolsLoc(documentElement,config)
-    handleReleaseType(documentElement, config)
-    handlePilotJobFlag(documentElement, config)
-    handleConcurrentFlag(documentElement, config)
-    handleDisableMailToFlag(documentElement, config)
-    handleDevRunFlag(documentElement, config) # if present, sets a smaller sleep time in TestRun
-    handleDeleteNonKeepFiles(documentElement, config)
-    handleChainStoreBasePath(documentElement, config)
-    handleArchive(documentElement, config)
-    # handleRTTRunFlag(documentElement, config)
-    # handleTCTRunFlag(documentElement, config)
-    # handleFCTRunFlag(documentElement, config)
-
-    # fill in blanks where necessry
-    addDefaultsToConfig(config)
-
-    logger.debug('Checked legality of top level configuration file:\n %s' % config)    
-    return config
-
-class RTTConfiguration:
-    def __init__(self, configFile):
-        self.config = rttConfiguration(configFile)
-        self.legals = LegalsFactory(logger).create(
-            self.config['releaseType'],
-            self.config['branch'],
-            HardCodedFactory(logger).create(self.config['releaseType'],
-                                            self.config['runType'])
-            
-            )
-        # make a copy of the branch as self.branch will be overwritten
-        # when handling numbered releases
-        self.originalBranch    = self.config['branch']
-        self.isNightly         = False
-        # the user may request that the release is to be discovered
-        # make a copy of the release here. This copy will be overwritten
-        # if the release is discovered
-        self.release         = self.config['release']
-        self.findNightly     = False
-        self.valid           = True
-        self.invalidText     = ''
-
-
-        self.isConfigFileLegal()
-        
-        if not self.valid:
-            print '\n\n*******************************************'
-            print '     ILLEGAL TOP LEVEL CONFIGURATION FILE!'
-            print '*******************************************\n\n'
-            logger.critical(self.invalidText)
-            raise RTTInputError(self.invalidText)
-
-        self.processConfigData()
-
-        
-    def isConfigFileLegal(self):
-        canTest = self.legals.validKeys()
-        willTest = [k for k in canTest if k in self.config.keys()]
-            
-        valids = [(k,
-                   self.config[k],
-                   self.legals.isLegal(k, self.config[k])) for k in willTest]
-
-        m = ''
-        for v in valids:
-            if not v[2]:
-                m += 'Illegal value (%s) for variable "%s".\n' % (v[1], v[0])
-                m += 'Legal values: \n'
-                for l in self.legals.vals(v[0]):
-                    m += '\t%s\n' % l
-                m += '******************************************************\n\n'
-        if m:
-            self.valid = False
-            self.invalidText += m
-
-
-        # need a test for webBranchVeto
-        
-            
-        msg = ensureInequalityWorkResBase(self.config['workBasePath'],
-                                          self.config['resultsBasePath'])
-        if msg:
-            self.valid = False
-            self.invalidText += msg+'\n'
-
-        msg = checkPaths(self.config)
-
-        if msg:
-            self.valid = False
-            self.invalidText += msg+'\n'
-    
-                
-    def processConfigData(self):
-        release          = self.config['release']
-        
-        # adjust the config file logical name for nightly
-        # releases to the name used as part of a path
-        if self.legals.isNightlyRelease(release):
-            self.release = self.legals.convertLogicalNightly(release)
-            self.isNightly = True
-        # set flag to find current nighlty release (user puts the word
-        # 'nightly' in the config file
-        if release in self.legals.vals('nightlyWord'):
-            self.findNightly = True
-            self.isNightly   = True
-
-        
-        numbered = isValidReleaseNumber(release)
-
-        if self.isNightly == numbered:
-            self.valid = False
-            msg  = 'Release %s is illegal.\n' % release
-            msg += 'Legal nightlies: %s \n' % str(self.legals.vals('nightlies'))
-            self.invalidText = msg
-        
-  
-    def discoverRelease(self): return self.findNightly
-    def setRelease(self, release): self.release = release
-    def get(self, key, defaultReturn=None):
-        keys = ['release', 'isNightly', 'originalBranch']
-        if key in keys: return self.__dict__[key]
-        if self.config.has_key(key): return self.config[key]
-        return defaultReturn
-    def text(self): return self.invalidText
-    def confDict(self): return self.config
-    def configIsValid(self): return self.valid
-    def __str__(self):
-        s = ''
-        for k in self.__dict__.keys():
-            if k != 'config':
-                s += '%s      %s\n' % (k, str(self.__dict__[k]))
-        s += '----\nconfig: \n%s\n' %formatCollection(self.config)
-        return s
-#----------------------------------------------------------------------
-
-if __name__ == '__main__':
-    #    import  getopt, sys
-    #    from findRTThome import findRTThome
-    #
-    #    try:
-    #        opts = getopt.getopt(sys.argv[1:],'c:')
-    #    except getopt.GetoptError:
-    #        print 'bad options'
-    #        sys.exit(2)
-    #
-    #    cfg = None
-    #    for opt in opts[0]:
-    #        print opt
-    #        if opt[0] == '-c': cfg = opt[1]
-    #
-    #
-    #    (base, cfg, src) = findRTThome(cfg)
-    #
-    #    if not os.path.exists(cfg):
-    #        print 'could not find cfg file',cfg
-    #
-    #    logger.info('config file: '+cfg)
-    #    
-    #    logger.info(formatCollection(rttConfiguration(cfg)))
-    #
-    b = ['Numbered','N.X.0','10.0.X']
-    print b
-    b =  fixSortBranches(b)
-    print b
diff --git a/Tools/RunTimeTester/src/RTTConstructionFactoryHelpers.py b/Tools/RunTimeTester/src/RTTConstructionFactoryHelpers.py
deleted file mode 100644
index 75221155e141eb1918daa9f68d38d121a77291ad..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTConstructionFactoryHelpers.py
+++ /dev/null
@@ -1,64 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Helper functions for RTT construction factories
-"""
-import string
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def makeArgDict(inlist, obj):
-
-    inlist = [a for a in inlist if a in obj.__dict__.keys()]
-    outDict = {}
-    for a in inlist: outDict[a] = obj.__dict__[a]
-    return outDict
-                
-class ArgBag:
-    def __init__(self, dict):
-        self.__dict__.update(dict)
-    def __str__(self):
-        longest = 0
-        for s in self.__dict__.keys():
-            if len(s) > longest: longest = len(s)
-        
-        s = ''
-        longest += 5
-        keys =  self.__dict__.keys()
-        keys.sort()
-
-        for v in keys:
-            vv = v+':'
-            vv = string.ljust(v, longest)
-            s += '%s%s\n' % (vv, self.__dict__[v])
-        return s
-
-    def pop(self, *args):
-        assert(len(args) in range(1,3))
-        if len(args) == 1:
-            return self.__dict__.pop(args[0])
-        return self.__dict__.pop(args[0], args[1])
-
-    def duplicate(self):
-        from copy import copy
-        new_ = {}
-        for k,v in self.__dict__.items():
-            new_[copy(k)] = copy(v)
-        return ArgBag(new_)
-
-def makeArgBag(inlist, obj):
-
-    missing = [a for a in inlist if a not in obj.__dict__.keys()]
-    if missing:
-        logger.info('makeArgBag Warning: cannot find %s' % str(missing))
-        
-    inlist = [a for a in inlist if a in obj.__dict__.keys()]
-    outDict = {}
-    for a in inlist: outDict[a] = obj.__dict__[a]
-    bag = ArgBag(outDict)
-    return bag
-                
-
diff --git a/Tools/RunTimeTester/src/RTTDateTime.py b/Tools/RunTimeTester/src/RTTDateTime.py
deleted file mode 100755
index a1fec5042330442503ae349ee03bd254aa4c400b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTDateTime.py
+++ /dev/null
@@ -1,50 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time
-
-class RTTDateTime:
-  def __init__(self, time=None):
-    self.ttime = time
-    self.valid = False
-    if time: self.valid = True
-
-  def isValid(self): return self.valid
-  
-  def date(self):
-    if self.ttime:
-      return time.strftime('%Y/%m/%d', time.gmtime(self.ttime))
-    else:
-      return '----/--/--'
-
-  def time(self):
-    if self.ttime:
-      return time.strftime('%H:%M:%S', time.gmtime(self.ttime))
-    else:
-      return '--:--:--'
-
-  def absTime(self): return self.ttime
-
-  def sameDate(self, other):
-    
-    gmt1 = time.gmtime(self.ttime)
-    
-    try:
-      gmt2 = time.gmtime(other.ttime)
-    except Exception:
-      return False
-    
-    return gmt1[0]==gmt2[0] and gmt1[1]==gmt2[1] and gmt1[2]==gmt2[2]
-  
-  def __str__(self):
-    return '%s %s' % (self.date(), self.time())
-
-  def __cmp__(self, other):
-    return cmp(self.ttime, other.ttime)
-    
-if __name__ == '__main__':
-  a = RTTDateTime(time.time())
-  time.sleep(3)
-  b = RTTDateTime(time.time())
-  print a<b
-  print b<a
-  
diff --git a/Tools/RunTimeTester/src/RTTFactory.py b/Tools/RunTimeTester/src/RTTFactory.py
deleted file mode 100644
index 3790ece5fa1a6e511a72af2912a4171f3318e2f4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTFactory.py
+++ /dev/null
@@ -1,355 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-PS 15/12/08
-RTTFactory is part a of a major refactorisation of the RTT,
-and is currently work in progress.
-
-We are moving object construction out of the main RTT classes
-(Tester, TestRun) and into the factory in order to
-
-- make the code dependencies clearer. Diminish the bag objects
-that carries much info to many places.
-
-- clarify the action structure of Tester et al
-
-- allow for shelving of objects to allow quick design turn around
-"""
-
-from Tester                    import Tester
-from PhasePrinter              import PhasePrinter
-from RTTpath                   import RTTpath
-from CMTProjectDirCalculator   import CMTProjectDirCalculator
-from RTTSException             import RTTCodingError
-from RTTSException             import RTTInputError
-from createDirectoryDictionary import findSeg
-from exc2string2               import exc2string2
-
-import os
-import shelve
-import sys
-
-from RTTConstructionFactoryHelpers import makeArgDict
-from RTTConstructionFactoryHelpers import ArgBag
-from RTTConstructionFactoryHelpers import makeArgBag
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-class RTTFactory:
-    def __init__(self, RTTSrcDir, RTTLibDir, topLevelRTTconfFileLocation, db, logDir):
-        
-        pp = PhasePrinter()
-
-        pp.printIt('getTimes')
-        from getTimes import getTimes
-        self.startTime1, self.startTime2 = getTimes()
-
-        self.RTTLibDir = RTTLibDir
-        self.RTTSrcDir = RTTSrcDir
-                    
-        
-        pp.printIt('validateInputParams')
-        
-        from validateInputParams import validateInputParams
-        validateInputParams(logger,  self.RTTSrcDir, self.RTTLibDir, topLevelRTTconfFileLocation)
-            
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('makeRTTDOM')
-        
-        from makeRTTDOM import makeRTTDOM
-        self.summaryDocument = makeRTTDOM('RTTSummary')
-        
-
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('readTopLevelConfigFile')
-
-        rttConfig = db.get('rttConfig', None)
-        if not rttConfig:
-            from  readTopLevelConfigFile import readTopLevelConfigFile
-            try:
-                rttConfig  = readTopLevelConfigFile(topLevelRTTconfFileLocation, logger)
-            except RTTCodingError, e:
-                print 'RTTCoding Error while reading top level configuration file'
-                print str(e)
-            except RTTInputError, e:
-                print 'RTTInput Error while reading top level configuration file'
-                print str(e)
-
-            db['rttConfig'] = rttConfig        
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('make noPackagePathsDict')
-        noPackagePathsDict = {}
-        noPackagePathsDict['RTTSrcDir'] = self.RTTSrcDir
-        noPackagePathsDict['RTTLibDir'] = self.RTTLibDir
-        fillWith = [
-            'cmtConfig',
-            'build',
-            'site',
-            'resultsBasePath',
-            'resultsBaseURL',
-            'disableMailTo',
-            'sandboxRunning',
-            'deleteNonKeepFiles',
-            'devRun',
-            'dCubeCfgFile',
-            'bigFilesArea',
-            'bigFilesSize',
-            'pathToVetoFile',
-            'chainStoreBasePath',
-            'isNightly',
-            'rttPilotJobs',
-            'atnPilotJobs',
-            'mode',
-            'rttProvidesConfFiles',
-            'localConfFiles',
-            'installedConfFiles',
-            'versionDir',
-            # 'rttRun',
-            # 'tctRun',
-            # 'fctRun',
-            'localPackages',
-            'dataSetCatalog',
-            'release',
-            'releaseType',
-            'runType',
-            'workBasePath',
-            'localRTTRun',
-            'originalBranch',
-            'branch',
-            'otherProject',
-            'distArea',
-            'topProject',
-            'confFileSuffix',
-            'confFiles',
-            'packagesAccept',
-            'packagesVeto',
-            'jobGroupConfig',
-            'devRun',
-            'archivers']
-
-        for k in fillWith: noPackagePathsDict[k] = rttConfig.get(k)
-
-        if rttConfig.get('localConfFiles'):
-            noPackagePathsDict['userHomeCmtDir']    = rttConfig.get('userHomeCmtDir')
-            noPackagePathsDict['cmtHomeDir']        = rttConfig.get('userHomeCmtDir')
-            noPackagePathsDict['testProjectCmtDir'] = rttConfig.get('testProjectCmtDir')
-            noPackagePathsDict['localRTTRun']    = True
-        else:
-            noPackagePathsDict['cmtHomeDir']     = os.path.join(noPackagePathsDict['workBasePath'], 'cmtHome')
-            noPackagePathsDict['cmtProjDir']     = os.path.join(noPackagePathsDict['workBasePath'], 'cmtProject')
-            noPackagePathsDict['localRTTRun']    = False
-
-        self.noPackagePaths = ArgBag(noPackagePathsDict)
-
-        pathsToCheck = [
-            'RTTSrcDir',
-            'RTTLibDir',
-            'resultsBasePath',
-            'dCubeCfgFile',
-            'installArea',
-            'bigFilesArea',
-            'pathToVetoFile',
-            'chainStoreBasePath',
-            'dataSetCatalog', 
-            'userHomeCmtDir',
-            'cmtHomeDir',
-            'cmtProjDir',
-            'workBasePath',
-            'distArea'
-            ]
-        
-        pathsToCheck = [self.noPackagePaths.__dict__.get(p, None) for p in pathsToCheck]
-        if self.noPackagePaths.rttProvidesConfFiles:
-            for cf in self.noPackagePaths.confFiles.values():
-                pathsToCheck.append(cf['installArea'])
-                pathsToCheck.append(cf['shareArea'])
-                pathsToCheck.append(cf['confFilePath'])
-                pathsToCheck.append(cf['cmtPath'])
-
-        pathsToCheck = [p for p in pathsToCheck if p]
-        badPaths = [p for p in pathsToCheck if not os.path.exists(p)]
-
-        if badPaths:
-            m =  'WARNING: bad paths\n'            
-            for p in badPaths: m+= '   %s\n' % p
-            logger.debug(m)
-            
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('pathSegment')
-        try:
-            self.noPackagePaths.pathSegment = findSeg(self.noPackagePaths.topProject)
-        except:
-            m = 'Could not get "segment" for topProject %s, aborting' % self.noPackagePaths.topProject
-            logger.error(m)
-            raise RTTCodingError(m)
-
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('HardCoded')
-        from Factory_HardCoded import HardCodedFactory
-        self.hardCoded      = HardCodedFactory(logger).create(self.noPackagePaths.releaseType, self.noPackagePaths.runType)
-
-        # --------------------------------------------------------------------------------------------------------        
-        
-        pp.printIt('Legals')
-        from Factory_Legals import LegalsFactory
-
-        self.legals    = LegalsFactory(logger).create(self.noPackagePaths.releaseType,
-                                                      self.noPackagePaths.originalBranch,
-                                                      self.hardCoded)
-        
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('SiteSpecifics')
-        from Factory_SiteSpecifics   import SiteSpecificsFactory
-        siteSpecificsFact            = SiteSpecificsFactory(logger)
-        self.siteSpecificsObj        = siteSpecificsFact.create(rttConfig.get('site'))
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('PathNameBuilder')
-        # to be injected into paths. was sitting in NoPackagePaths, but is not used there
-        from Factory_PathNameBuilder    import PathNameBuilderFactory
-        factory = PathNameBuilderFactory(logger)
-
-        labels = ['release', 'cmtConfig', 'originalBranch', 'isNightly', 'topProject', 'otherProject', 'distArea']
-        argBag = makeArgBag(labels, self.noPackagePaths)
-        self.pnb   = factory.create(self.noPackagePaths.releaseType, argBag, self.hardCoded, self.legals, self.siteSpecificsObj)
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('make StampObject')
-        from StampObject import StampObject
-        stampFile = str(self.pnb.pathToNICOSLatestCopiedLink())
-        self.stampObject = StampObject(stampFile)
-        
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('find nightly release')
-        if rttConfig.findNightly:
-            from findNightlyRelease import findNightlyRelease
-            release = findNightlyRelease(self.stampObject, logger)
-            if not release: sys.exit(1) # exit, no nightly flag came
-
-            self.noPackagePaths.release = release
-            self.pnb.release = release
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('CMTConfigurer')
-        from Factory_CMTConfigurer import CMTConfigurerFactory
-        factory = CMTConfigurerFactory(logger)
-
-        factoryArgBag = makeArgBag(['runType', 'localRTTRun'], self.noPackagePaths)
-        argDict       = makeArgDict(['cmtHomeDir', 'release', 'branch', 'topProject', 'distArea',
-                                     'cmtProjDir', 'userHomeCmtDir'], self.noPackagePaths)
-
-        argDict['hardCoded'] = self.hardCoded
-
-        cmtConfigurer = factory.create(factoryArgBag, argDict)
-
-        self.noPackagePaths.cmtVersion = cmtConfigurer.cmtVersion
-
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('configure CMT dirs')
-        from configureCMTDirs import configureCMTDirs
-        configureCMTDirs(self.noPackagePaths.localRTTRun, cmtConfigurer, logger)
-
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('make CMTProjectDirCalculator')
-
-        self.projectCMTDirs = db.get('projectCMTDirs', None)
-        if not self.projectCMTDirs:
-            cmtProjectDirCalculator = CMTProjectDirCalculator(self.noPackagePaths.topProject,
-                                                              self.noPackagePaths.originalBranch,
-                                                              self.noPackagePaths.release,
-                                                              self.noPackagePaths.otherProject,
-                                                              self.noPackagePaths.isNightly,
-                                                              self.noPackagePaths.cmtConfig,
-                                                              self.noPackagePaths.cmtHomeDir,
-                                                              logger)
-            
-            self.projectCMTDirs = cmtProjectDirCalculator.projectCMTDirs
-            db['projectCMTDirs'] = self.projectCMTDirs
-
-
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('make results and work directory dictionaries')
-        from createDirectoryDictionary import createDirectoryDictionary
-        argBag = makeArgBag(['release', 'isNightly', 'originalBranch', 'runType', 'cmtConfig',
-                             'topProject', 'otherProject'], self.noPackagePaths)
-        argBag.base = self.noPackagePaths.workBasePath
-        try:
-            self.noPackagePaths.workDirs = createDirectoryDictionary(argBag)
-        except Exception, e:
-            m  = 'Error calling createDirectoryDictionary: arguments:\n%s\n' % str(e)
-            m += 'Exception:\n%s\n' % str(argBag.__dict__)
-            m += 'Traceback:\n%s' % exc2string()
-            logger.error(m)
-            raise RTTCodingError(m)
-
-        argBag.base = self.noPackagePaths.resultsBasePath
-        try:
-            self.noPackagePaths.resultsDirs = createDirectoryDictionary(argBag)
-        except Exception, e:
-            m  = 'Error calling createDirectoryDictionary: arguments:\n%s\n' % str(e)
-            m += 'Exception:\n%s\n' % str(argBag.__dict__)
-            m += 'Traceback:\n%s' % exc2string()
-            logger.error(m)
-            raise RTTCodingError(m)
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('make closeDownKeyPath')
-        
-        self.noPackagePaths.RTTSummary     = os.path.join(self.noPackagePaths.resultsDirs['packages'], 'RTTSummary.xml')
-        self.noPackagePaths.closeDownFile  = os.path.join(self.noPackagePaths.workDirs['base'], 'closeDownFile.xml')
-
-        if not self.noPackagePaths.isNightly: self.noPackagePaths.branch = 'Numbered' 
-        limitedNightlies = [self.legals.nightlies.get(n, n) for n in self.hardCoded.limitedNightlies]
-        self.noPackagePaths.limitedNightlyVeto   = not self.noPackagePaths.release in limitedNightlies
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('Outputting PID of this python process')
-        from Tools import mkdirIfNew
-
-        basepath = os.path.join(self.noPackagePaths.resultsBasePath,'rtt.prod.pids')
-        mkdirIfNew(basepath)
-        if not os.path.exists(basepath):
-            logger.error('Unable to create inexistant %s' % basepath)
-        else:
-            my_pid = str(os.getpid())
-            pidfile = os.path.join(basepath,my_pid)
-            if os.path.exists(pidfile):
-                logger.warning('%s: already exists! Doing nothing.' % pidfile)
-            else:
-                runFrag = self.noPackagePaths.resultsDirs['packages']
-                os.system('echo %s %s > %s' % (self.startTime2, runFrag, pidfile))
-        
-        # --------------------------------------------------------------------------------------------------------
-        
-        pp.printIt('store NoPackagePaths')            
-        db['noPackagePathsArgBag'] = self.noPackagePaths
-        
-
-if __name__ == '__main__':
-    pass 
-    # RTTSrcDir                   = '/Users/peter/RTTdevelopment/Refactor/Directories/src'
-    # RTTLibDir                   = '/Users/peter/RTTdevelopment/Refactor/Directories/share' 
-    # topLevelRTTconfFileLocation = '/Users/peter/RTTdevelopment/Refactor/Directories/cfg/cfg.xml'
-    # logDir                      = '/Users/peter/RTTdevelopment/Refactor/Directories/logs'
-
-    # factory = RTTFactory(RTTSrcDir, RTTLibDir, topLevelRTTconfFileLocation, logDir)
diff --git a/Tools/RunTimeTester/src/RTTFactory2.py b/Tools/RunTimeTester/src/RTTFactory2.py
deleted file mode 100644
index 762fae84234c9a0e80db25b4bae688eeb1dfe69b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTFactory2.py
+++ /dev/null
@@ -1,611 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-PS 15/12/08
-RTTFactory is part a of a major refactorisation of the RTT,
-and is currently work in progress.
-
-We are moving object construction out of the main RTT classes
-(Tester, TestRun) and into the factory in order to
-
-- make the code dependencies clearer. Diminish the bag objects
-that carries much info to many places.
-
-- clarify the action structure of Tester et al
-
-- allow for shelving of objects to allow quick design turn around
-"""
-
-from Tester                        import Tester
-from RTTCloseDownKeyMaker          import RTTCloseDownKeyMaker
-from PhasePrinter                  import PhasePrinter
-from exc2string2                   import exc2string2
-from ATNSummaryRetriever           import ATNSummaryRetriever
-from RTTSException                 import RTTCodingError
-from RTTSException                 import RTTInputError
-from Tools                         import nowSecs
-
-from RTTConstructionFactoryHelpers import makeArgDict
-from RTTConstructionFactoryHelpers import ArgBag
-from RTTConstructionFactoryHelpers import makeArgBag
-
-import os
-import shelve
-import time
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-class RTTFactory2:
-    def __init__(self, factory, logDir, db):
-        noPackagePaths   = factory.noPackagePaths
-        legals           = factory.legals
-        hardCoded        = factory.hardCoded
-        siteSpecificsObj = factory.siteSpecificsObj
-        pnb              = factory.pnb
-        stampObject      = factory.stampObject
-        RTTSrcDir        = factory.RTTSrcDir
-        RTTLibDir        = factory.RTTLibDir
-        startTime1       = factory.startTime1
-        startTime2       = factory.startTime2
-        summaryDocument  = factory.summaryDocument
-        projectCMTDirs   = factory.projectCMTDirs
-        
-        # db = shelve.open('RTTFactory.dbm', 'c')
-        
-        pp = PhasePrinter()
-
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make BIGIO')
-        
-        # watch out - has global! was also being set self.bigFileIO in tester, but no more!
-        from BigFileIO import createBigFileIO
-        self.bigFileIO = createBigFileIO(noPackagePaths.site,
-                                         noPackagePaths.bigFilesArea,
-                                         noPackagePaths.workDirs,
-                                         noPackagePaths.isNightly)
-        
-
-        # --------------------------------------------------------------------------------------------------------
-                                      
-        pp.printIt('make NICOS Dates')
-        nicosDate          = 'n/a'
-        nicosDateFormatted = 'n/a'
-
-        from NICOS import getReleaseDate
-        try:
-            nDate = getReleaseDate(pnb)
-        except:
-            logger.warning('Invalid NICOS date')
-        else:
-            if nDate.isValid():
-                nicosDate          = str(nDate.absTime())
-                nicosDateFormatted = str(nDate)
-
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make CMT lines commands')
-        from Factory_CmtLines import CmtLinesFactory
-        argBagFactory = makeArgBag(['runType','localRTTRun', 'topProject'], noPackagePaths)
-
-        labels = ['release', 'branch', 'originalBranch', 'topProject', 'otherProject', 'cmtHomeDir', 'cmtConfig',
-                  'userHomeCmtDir', 'testProjectCmtDir']
-        argBag                = makeArgBag(labels, noPackagePaths)
-        argBag.projectCMTDirs = projectCMTDirs
-        argBag.logger         = logger
-        
-        cmtLinesFactory = CmtLinesFactory(logger).create(argBagFactory, argBag)
-        cmtLinesCmds = cmtLinesFactory.lines()
-
-        # --------------------------------------------------------------------------------------------------------                   
-        pp.printIt('make UserStuffRetriever')
-        # --------------------------------------------------------------------------------------------------------                   
-        usr = None
-        if noPackagePaths.installedConfFiles or noPackagePaths.localConfFiles:
-            from Factory_UserStuffRetriever import UserStuffRetrieverFactory
-            fact       = UserStuffRetrieverFactory(logger)
-
-            factoryArgBag = makeArgBag(['localRTTRun', 'runType', 'release', 'branch'] , noPackagePaths)
-
-            labels = ['cmtConfig', 'topProject', 'otherProject', 'confFileSuffix']
-            argBag                = makeArgBag(labels, noPackagePaths)
-            argBag.hardCoded      = hardCoded
-            argBag.cmds           = cmtLinesCmds[:] # copy, don't just reference. USR appends to this which messes things up otherwise.
-            argBag.projectCMTDirs = projectCMTDirs
-            
-            usr   = fact.create(factoryArgBag, argBag)
-
-        # --------------------------------------------------------------------------------------------------------                   
-        pp.printIt('make GetUserData')
-        # --------------------------------------------------------------------------------------------------------                   
-
-        from GetUserConfigData          import GetUserConfigData
-        labels = ['installedConfFiles', 'localConfFiles', 'rttProvidesConfFiles', 'confFileSuffix', 'confFiles',
-                  'packagesAccept', 'packagesVeto', 'cmtHomeDir', 'cmtProjDir']
-        argBag        = makeArgBag(labels, noPackagePaths)
-        argBag.usr    = usr
-        argBag.logger = logger
-        argBag.pnb    = pnb
-        
-        getUserConfigData = GetUserConfigData(argBag)
-
-        # --------------------------------------------------------------------------------------------------------
-                               
-        pp.printIt('make cvsPackages')
-        
-        cvsPackages = getUserConfigData.userConfigData()
-        
-        from cullPackages import cullPackages
-        cvsPackages = cullPackages(cvsPackages, logger)
-        if len(cvsPackages) == 0:
-            m = 'No packages left after RTTFactory::cullPackages(), abandoning RTT run'
-            logger.error(m)
-            raise RTTInputError(m)
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make Kits')
-        
-        # createKits sets up a Singleton
-        from createKits import createKits
-        jobGroupKits = createKits(cvsPackages, noPackagePaths.jobGroupConfig, logger)
-        # print str(jobGroupKits)
-        # assert(False)
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('make joSearchPaths')
-        
-        from findJobOptionsSearchPath import findJobOptionsSearchPath
-        joSearchPaths = db.get('joSearchPaths', None)
-
-        argBag = makeArgBag(['isNightly', 'release'], noPackagePaths)
-
-        if not joSearchPaths:
-            try:
-                joSearchPaths = findJobOptionsSearchPath(cmtLinesCmds, argBag, logger)
-            except Exception, e:
-                print e
-            else:
-                db['joSearchPaths'] = joSearchPaths
-        
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make Paths')
-        from Paths import Paths
-
-        packagePaths = db.get('packagePaths', None)
-            
-        if not packagePaths:
-
-            argLabels = [
-                'RTTSrcDir',
-                'RTTLibDir',
-                'cmtConfig',
-                'build',
-                'site',
-                'resultsBasePath',
-                'resultsBaseURL',
-                'disableMailTo',
-                'sandboxRunning',
-                'deleteNonKeepFiles',
-                'devRun',
-                'dCubeCfgFile',
-                'installArea',
-                'bigFilesArea',
-                'bigFilesSize',
-                'pathToVetoFile',
-                'chainStoreBasePath',
-                'isNightly',
-                'rttPilotJobs',
-                'atnPilotJobs',
-                'mode',
-                'dataSetCatalog',
-                'rttProvidesConfFiles',
-                'localConfFiles',
-                'installedConfFiles',
-                'versionDir',
-                'cmtHomeDir',
-                'cmtProjDir',
-                'localPackages',
-                'workBasePath',
-                'release',
-                'runType',
-                'topProject',
-                'otherProject',
-                'workDirs',
-                'resultsDirs',
-                'closeDownFile',
-                'originalBranch',
-                'branch',
-                'devRun',
-                'limitedNightlyVeto',
-                'localRTTRun',
-                'distArea',
-                'pathSegment',
-                'archivers']
-
-            # make argumentBag an attribute to allow resuse when creating Paths objects
-            argBag = makeArgBag(argLabels, noPackagePaths)
-            argBag.siteSpecificsObj = siteSpecificsObj
-            argBag.legals           = legals
-            argBag.projectCMTDirs   = projectCMTDirs
-            argBag.hardCoded        = hardCoded
-            argBag.pathNameBuilder  = pnb
-            argBag.logDir           = logDir
-
-            packagePaths = [(p, Paths(argBag.duplicate(),
-                                      p,
-                                      joSearchPaths,
-                                      summaryDocument.createElement,
-                                      summaryDocument.createTextNode,
-                                      logger)) for p in cvsPackages]
-
-            for p in packagePaths: p[1].logger = None
-
-        for p in packagePaths: p[1].logger = logger
-        
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make tester XMLConverter')
-        from TesterXMLConverter2 import TesterXMLConverter
-
-        elementCreator  = summaryDocument.createElement
-        textNodeCreator = summaryDocument.createTextNode
-
-        labels =['resultsDirs', 'RTTSrcDir', 'RTTLibDir', 'topProject', 'otherProject',
-                 'cmtConfig', 'originalBranch', 'site', 'release', 'releaseType', 'runType',
-                 'RTTSummary', 'devRun']
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('Tester Logger setup')
-
-
-        from makeLocalLog                 import makeLocalLog
-        testerLogger = makeLocalLog(
-            logDir,
-            'TesterError.log',
-            'TesterWarning.log',
-            'TesterInfo.log',
-            'TesterDebug.log',
-            'Tester'
-            )
-        
-
-        argBag = makeArgBag(labels, noPackagePaths)
-        argBag.legals             = legals
-        argBag.logger             = testerLogger
-        argBag.startTime1         = startTime1
-        argBag.startTime2         = startTime2
-        argBag.nicosDate          = nicosDate
-        argBag.nicosDateFormatted = nicosDateFormatted
-        argBag.node               = summaryDocument.documentElement
-
-        testerXMLConverter = TesterXMLConverter(argBag, elementCreator, textNodeCreator)
-
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make closeDownKeyMaker')
-        labels = ['closeDownFile', 'cmtConfig', 'originalBranch', 'release', 'releaseType', 'runType',
-                  'topProject', 'otherProject', 'isNightly']
-        argBag              = makeArgBag(labels, noPackagePaths)
-        argBag.logger       = testerLogger
-        argBag.testerStart  = startTime2
-
-        self.closeDownKeyMaker = RTTCloseDownKeyMaker(argBag)
-
-        # --------------------------------------------------------------------------------------------------------
-        
-        from DataSetCatalog import DataSetCatalog
-        dataSetCatalog = DataSetCatalog(noPackagePaths.dataSetCatalog, testerLogger)
-        
-        # --------------------------------------------------------------------------------------------------------
-                   
-        pp.printIt('make Factories for TestRuns')
-
-        from Factory_TestRunXMLConverter import TestRunXMLConverterFactory
-        xf = TestRunXMLConverterFactory(logger)
-
-        from Factory_JobGroupDirectoryMaker import JobGroupDirectoryMakerFactory
-        jobGroupDirectoryMakerFactory = JobGroupDirectoryMakerFactory(logger)
-
-        from Factory_Commander import CommanderFactory
-        commander = CommanderFactory(noPackagePaths.mode, logger).create()
-
-        from UnifiedConfigurationTestRun  import UnifiedConfigurationTestRun
-        from FailureTestRun               import FailureTestRun
-        from Factory_JobDescriptor        import DescriptorFactory
-        from Factory_Minder               import MinderFactory
-        from ProductionManagerVetoer      import ProductionManagerVetoer
-        from Factory_JobsXMLReader        import JobsXMLReaderFactory
-        from cullJobDescriptors           import cullJobDescriptors
-        from PackageMailer                import PackageMailer
-        from ChainStoreMaker              import ChainStoreMaker
-        
-        def makeTestRun((cvsPackage, paths)):
-
-
-            pkgName = cvsPackage.name
-            pp.printIt('make TestRun: %s' % pkgName)
-
-            labels = ['originalBranch', 'cmtConfig', 'package', 'release', 'runType', 'topProject',  'otherProject']
-            closeDownKey = makeArgBag(labels, paths)
-            for l in labels: assert(hasattr(closeDownKey, l))
-
-            closeDownKey.runStartTime = startTime2
-            if paths.isNightly: closeDownKey.release = 'nightly'
-
-            # make package directories
-            directoriesToMake = [paths.workDirs['package'], paths.resultsDirs['package']]
-
-            sleeptimes=[30,60,300]
-            for sleeptime in sleeptimes:
-                try:
-                    [os.mkdir(d) for d in directoriesToMake if not os.path.exists(d)]
-                    break
-                except:
-                    
-                    time.sleep(sleeptime)
-
-            # Log files: make directories and logger that routes to 4 files
-            pLogDir = os.path.join(logDir, pkgName)
-            if not os.path.exists(pLogDir): os.makedirs(pLogDir)
-
-
-            # make the logger for this test run unique (stops cross talking betweeen testruns for the same package which run in different RTT instances
-            # this occcurs while testing the RTT
-            time.sleep(1)
-            timedPackageName = pkgName+nowSecs() 
-            trLogger = makeLocalLog(pLogDir, pkgName+'Error.log', pkgName+'Warning.log', pkgName+'Info.log', pkgName+'Debug.log',timedPackageName)
-
-            # make a mailer pbject to handle emails to developers in case of problems
-            mailer = PackageMailer(pkgName,
-                                   paths.mailTo,
-                                   paths.release,
-                                   paths.runType,
-                                   paths.originalBranch,
-                                   paths.cmtConfig,
-                                   paths.pathSegment,
-                                   paths.packageTag,
-                                   paths.resultsDirs,
-                                   paths.resultsBaseURL,
-                                   paths.confFile,
-                                   paths.disableMailTo,
-                                   trLogger)
-
-            # make a Failure TestRun if there is already a known problem
-            if not cvsPackage.runJobs:
-                logger.warning('Package %s RTT operator veto\n' % pkgName)
-                testRun =  FailureTestRun(pkgName, closeDownKey, startTime2, self.closeDownKeyMaker, trLogger, 'RTT operator veto', True)
-                xmlConverter = xf.create(testRun,
-                                         paths.resultsDirs['package'],
-                                         pkgName,
-                                         paths.packageTag,
-                                         paths.containerPackage,
-                                         testerXMLConverter,
-                                         elementCreator,
-                                         textNodeCreator,
-                                         trLogger)
-                testRun.setXMLConverter(xmlConverter)
-                return testRun
-            if not cvsPackage.noError():
-                logger.warning('Package %s CVS package error: %s\n' % (pkgName, cvsPackage.errorMessage))
-                testRun =  FailureTestRun(pkgName, closeDownKey, startTime2, self.closeDownKeyMaker, trLogger, cvsPackage.errorMessage, False)
-                xmlConverter = xf.create(testRun,
-                                         paths.resultsDirs['package'],
-                                         pkgName,
-                                         paths.packageTag,
-                                                    paths.containerPackage,
-                                         testerXMLConverter,
-                                         elementCreator,
-                                         textNodeCreator,
-                                         trLogger)
-                testRun.setXMLConverter(xmlConverter)
-                return testRun
-
-            
-            labels = ['searchPaths', 'workDirs', 'resultsDirs',  # needed by descriptors
-                      'dCubeCfgFile', 'installArea'              # needed by paths
-                      ]
-            
-            argBag                      = makeArgBag(labels, paths)
-            argBag.legals               = legals
-            argBag.jobOptionsSearchPath = joSearchPaths
-            argBag.pathNameBuilder      = pnb  # needed to instantiate paths
-            smallPaths                  = Paths(argBag, cvsPackage, paths.searchPaths, elementCreator, textNodeCreator, trLogger)
-            descFactory                 = DescriptorFactory(smallPaths, trLogger)
-
-            labels = ['release', 'originalBranch', 'runType', 'cmtConfig', 'topProject', 'isNightly', 'package', 'otherProject']
-            argBag                      = makeArgBag(labels, paths) 
-            argBag.base                 = paths.chainStoreBasePath
-            enableMake                  = True
-            chainStoreMaker             = ChainStoreMaker(argBag, enableMake)
-            
-            # the xml files that contain the user information that species the jobs can come in various flavors
-            # (RTT, ATN....) which are parsed by different parsers.
-            readerFactory = JobsXMLReaderFactory(logger)
-            xmlReader = readerFactory.create(paths.mode,
-                                             cvsPackage.packageXMLfile['rtt/jobList[0]'],
-                                             paths.rttPilotJobs,
-                                             pkgName,
-                                             dataSetCatalog,
-                                             chainStoreMaker,
-                                             descFactory)
-            
-            descs = xmlReader.getDescriptors()
-
-            try:
-                atnNode = cvsPackage.packageXMLfile['atn[0]']
-            except:
-                pass
-            else:                
-                xmlReader = readerFactory.create(paths.mode,
-                                                 atnNode,
-                                                 paths.rttPilotJobs,
-                                                 pkgName,
-                                                 dataSetCatalog,
-                                                 chainStoreMaker,
-                                                 descFactory)
-                descs.extend(xmlReader.getDescriptors())
-
-            logger.debug('No of descriptors: %d' % len(descs))
-            descs = cullJobDescriptors(descs, paths.isNightly, paths.limitedNightlyVeto, paths.rttPilotJobs)
-            logger.debug('No of descriptors after vetos: %d', len(descs))
-
-
-
-            labels = [
-                'workDirs',
-                'cmtConfig',
-                'RTTSrcDir',
-                'RTTLibDir',
-                'releasePath',
-                'cmtDir',
-                'package',
-                'pathToProjectCMTDir',
-                'topProject',
-                'mode'
-                ]
-            argBag               = makeArgBag(labels, paths)
-            argBag.cmtLines      = cmtLinesFactory.cmtLinesScript()
-            argBag.siteSpecifics = siteSpecificsObj
-            from Factory_ScriptWriter import ScriptWriterFactory
-            scriptWriterFactory = ScriptWriterFactory(argBag, logger)
-            
-            # make an object to allow vetos on the new minders, pass this to a minder factory
-            pathToVetoFile = paths.__dict__.get('pathToVetoFile', None)
-            jobVetoer     = ProductionManagerVetoer(pathToVetoFile, trLogger)
-
-            labels = ['shareArea',
-                      'logDir',
-                      'package',
-                      'confFile',
-                      'bigFilesSize',
-                      'bigFilesArea',
-                      'deleteNonKeepFiles',
-                      'mode',
-                      'resultsBasePath',
-                      'elementCreator',
-                      'textNodeCreator',
-                      'release',
-                      'workDirs',
-                      'isNightly',
-                      'site',
-                      'dCubeCfgFile',
-                      'installArea',
-                      'containerPackage',
-                      'runType',
-                      'cmtPath',
-                      'localRTTRun',
-                      'distArea',
-                      'topProject',
-                      'branch',
-                      'cmtConfig',
-                      'archivers'
-                      ]
-            argBag                      = makeArgBag(labels, paths)
-            argBag.commander            = commander
-            argBag.scriptWriterFactory  = scriptWriterFactory
-            argBag.RTTLibDir            = RTTLibDir
-            argBag.RTTSrcDir            = RTTSrcDir
-            argBag.cmtLinesCmds         = cmtLinesCmds
-
-            minderFactory = MinderFactory(jobVetoer,
-                                          scriptWriterFactory,
-                                          jobGroupDirectoryMakerFactory,
-                                          jobGroupKits,
-                                          trLogger)
-
-            argBag.minderFactory = minderFactory # used by container minders
-            try:
-                minders = [minderFactory.create(argBag, d) for d in descs]
-            except Exception, e:
-                msg =  'Error creating minders:\nException dump starts:\n %s\nTraceBack\n%s \nException dump ends' % (str(e), exc2string2())
-                logger.error(msg)
-                raise RTTCodingError(msg)
-
-
-            logger.debug('Made %d minders' % len(minders))
-            for m in minders: logger.debug(m.__class__.__name__)
-            # we now give the minders to the TestRun, which looks after all the jobs for a package
-            testRun =  UnifiedConfigurationTestRun(minders, pkgName, closeDownKey,
-                                                   self.closeDownKeyMaker, paths.mode, paths.devRun,
-                                                   mailer, trLogger, '')
-            # xmlconverter is the summary callback object for TestRuns
-            # Needs the tester xmlConverter to set up the call back
-            xmlConverter = xf.create(testRun,
-                                     paths.resultsDirs['package'],
-                                     pkgName,
-                                     paths.packageTag,
-                                     paths.containerPackage,
-                                     testerXMLConverter,
-                                     elementCreator,
-                                     textNodeCreator,
-                                     trLogger)
-            
-            testRun.setXMLConverter(xmlConverter)
-            return testRun
-
-        testRuns = [makeTestRun(p) for p in packagePaths]
-        msg = '%d test runs created' % len(testRuns)
-        ftrs = [t for t in  testRuns if t.__class__.__name__ == 'FailureTestRun']
-        msg +=  ' of which %d were FailureTestRuns' % len(ftrs)
-        logger.debug(msg)
-        
-
-        # --------------------------------------------------------------------------------------------------------
-                                      
-        pp.printIt('make ATNretriever')
-
-        atnSummary = None
-        try:
-            retriever = ATNSummaryRetriever(
-                noPackagePaths.cmtConfig,
-                noPackagePaths.originalBranch,
-                noPackagePaths.release,
-                logger)
-        except:
-            logger.info('Could not instantiate ATNSummaryRetriever %s ' % exc2string2())
-        else:
-            atnSummary = retriever.getDom()
-
-        
-        # --------------------------------------------------------------------------------------------------------
-        pp.printIt('add ATNsummary')
-        
-        if atnSummary:
-            deep = True
-            importedSummary = summaryDocument.importNode(atnSummary.documentElement, deep)
-            summaryDocument.documentElement.appendChild(importedSummary)
-        else:
-            logger.info('Unable to obtain ATN summary')
-
-        # --------------------------------------------------------------------------------------------------------
-
-        pp.printIt('make Tester')
-        
-        labels              = ['release', 'branch', 'releaseType', 'runType', 'versionDir', 'cmtConfig',
-                               'originalBranch', 'site']
-        argBag                 = makeArgBag(labels, noPackagePaths)
-        argBag.stampObject     = stampObject
-        argBag.legals          = legals
-        argBag.summaryDocument = summaryDocument
-        argBag.logger          = testerLogger
-        argBag.xmlConverter    = testerXMLConverter 
-        argBag.testRuns        = testRuns
-        argBag.startTime2      = startTime2
-        
-        self.tester = Tester(argBag)
-
-        # --------------------------------------------------------------------------------------------------------
-                                   
-        pp.printIt('Done creating objects')
-
-        # db.close()
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/RTTFileLogger.py b/Tools/RunTimeTester/src/RTTFileLogger.py
deleted file mode 100755
index 9797e87a25f03a5ab1caa742bd6e486a797eec65..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTFileLogger.py
+++ /dev/null
@@ -1,44 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import logging, os, logging.handlers
-
-class RTTFileHandler(logging.handlers.MemoryHandler):
-    def __init__(self, ofn):
-        logging.handlers.MemoryHandler.__init__(self, 1000, logging.DEBUG)
-        self.ofn = ofn
-        try:
-            os.remove(ofn)
-        except:
-            pass
-        
-    def emit(self, record):
-        of = open(self.ofn, 'a')
-        of.write(self.formatter.format(record)+'\n')
-        of.close()
-    def shouldFlush(self, record): return True
-
-def getLogger(name, ofn):
-    msgfmt = '%(asctime)s %(filename)s %(lineno)s %(levelname)s %(message)s'
-    timefmt   = '%d%b%y %H:%M:%S'
-    formatter=logging.Formatter(msgfmt, timefmt)
-
-
-
-    handler = RTTFileHandler(ofn)
-    handler.setFormatter(formatter)
-    handler.setLevel(logging.DEBUG)
-    
-    logger = logging.getLogger(name)
-    logger.addHandler(handler)
-    logger.setLevel(logging.DEBUG)
-
-    return logger
-
-def useLogger(s, logger):
-    logger.debug(s)
-
-def testFunction():
-    logger = getLogger('rttlogger', 'logtest.txt')
-    logger.debug('aaa')
-    logger.error('bbb\nccc')
-
diff --git a/Tools/RunTimeTester/src/RTTHistoDataMaker.py.obsolete b/Tools/RunTimeTester/src/RTTHistoDataMaker.py.obsolete
deleted file mode 100755
index b5c1bd1fe3c1759d81745edbdbb34b9d5b0c36bb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTHistoDataMaker.py.obsolete
+++ /dev/null
@@ -1,79 +0,0 @@
-from makeRTTDOM         import makeRTTDOM
-from RTTXMLFileSelector import RTTXMLFileSelector
-from Evaluate           import Evaluate, getText
-from xml.dom.minidom    import parseString
-from xml.dom.ext        import PrettyPrint
-from exc2string2        import exc2string2
-from StringIO           import StringIO
-
-import time, sys
-
-def getSummedData(key, dom, tagPaths, logger):
-
-    def getTagContent(xpath, key, dom):
-        tags = Evaluate(xpath, dom.documentElement)
-
-        if not tags:
-            #temp = StringIO()
-            #PrettyPrint(dom, temp)
-            m = 'Unable to find XML path %s in moni DB entry with key %s\n' % (xpath, str(key))
-            
-            logger.info(m)
-            return 0.
-        
-        data = [float(getText(tag)) for tag in tags]
-        sum = reduce(lambda x,y: x+y, data, 0)
-        return sum
-
-    point = []
-    [point.append(getTagContent(tag, key, dom)) for tag in tagPaths]
-    return point
-        
-
-def  RTTHistoDataMaker(dbPath, selector, nDays, tagpaths, logger):
-
-    # returns a dictionary whose keys are matching DB keys, values = db values
-    matchingDataDict = RTTXMLFileSelector(str(dbPath), selector, logger)
-    
-    nSecs = nDays*24*60*60
-    t0    = time.time()-nSecs
-
-    doms = []
-
-    items =  matchingDataDict.items()
-    logger.info('No of points found in db: %d' % len(items))
-        
-    for key,value in  items:
-        logger.debug('parsing file with key %s' % str(key))
-        try:
-            dom = parseString(value)
-        except:
-            logger.error('Error trying to parseString in RTTHistoDataMaker. Key with unparseable value: %s' % str(key))
-            logger.error(exc2string2())
-        else:
-            item = (key,dom)
-            doms.append(item)
-
-    logger.debug('No of doms recovered from db: %d' % len(doms))
-    doms = [d for d in doms
-            if float(getText(Evaluate('overview/nicosDate2', d[1].documentElement)[0]))>t0]
-
-    points = [getSummedData(d[0], d[1],tagpaths, logger) for d in doms]
-    logger.debug('No of doms selected: %d' % len(points))
-
-    return points
-
-
-if __name__ == '__main__':
-    
-    class Logger:
-        def error(self, text): print text
-    # The full path to the database file with the xml files for the RTT runs/
-    dbPath = '/local_disk/sherwood/RTT02/moniFile.db'
-
-    
-    key    = {'originalBranch': u'N.X.0'}
-    tagpaths = ['nicosDate2','nJobs', 'nJobsOK']
-    days = 7
-    points =  RTTHistoDataMaker(dbPath, key, days, tagpaths, Logger())
-    print points
diff --git a/Tools/RunTimeTester/src/RTTHistoDataOut.py.obsolete b/Tools/RunTimeTester/src/RTTHistoDataOut.py.obsolete
deleted file mode 100755
index 70d772653f623683a6f2044bd3ee94eb415f2e29..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTHistoDataOut.py.obsolete
+++ /dev/null
@@ -1,45 +0,0 @@
-from RTTHistoDataMaker import RTTHistoDataMaker
-
-class RTTHistoDataOut:
-    def __init__(self, outfilename, dbPath, selector, days, tagpaths, logger):
-        self.fn       = outfilename
-        self.dbPath   = dbPath
-        self.selector = selector # dict with database key (a dict) to match on retrieval
-        self.days     = days    # get data for the last number of days
-        self.tagpaths = tagpaths  # points to plot
-        self.logger   = logger
-
-    def update(self):
-
-        points = RTTHistoDataMaker(self.dbPath, self.selector, self.days, self.tagpaths, self.logger)
-
-        
-        def pointToString(point):
-            s = ''
-            for d in point:
-                s += '%d ' % d
-            s += '\n'
-            return s
-
-        spoints = [pointToString(point) for point in points]
-        self.logger.debug('RTTHistoDataOut: writing out %d points' % len(spoints))
-        
-        of = open(self.fn, 'w')
-        [of.write(point) for point in spoints]
-        of.close()
-        
-if __name__ == '__main__':
-
-        
-    class Logger:
-        def error(self, text): print text
-    # The full path to the database file with the xml files for the RTT runs/
-    dbPath = '/local_disk/sherwood/RTT02/moniFile.db'
-
-    
-    key    = {'originalBranch': u'N.X.0'}
-    tagpaths = ['nicosDate2','nJobs', 'nJobsOK']
-    days = 8
-
-    rho = RTTHistoDataOut('./junkData.dat', dbPath,  key, days, tagpaths, Logger())
-    rho.update()
diff --git a/Tools/RunTimeTester/src/RTTJobCounter.py.obsolete b/Tools/RunTimeTester/src/RTTJobCounter.py.obsolete
deleted file mode 100755
index 86ef07accddfae353f1ffb55cecf6c36cdd9bc20..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTJobCounter.py.obsolete
+++ /dev/null
@@ -1,26 +0,0 @@
-import time, os.path
-from Tools import toJavaScript
-class RTTJobCounter:
-    def __init__(self, testRuns, odir, testerStatus):
-
-        self.testRuns     = testRuns
-        self.ofname       = os.path.join(odir, 'runInfo.js')
-        self.update()
-        
-    def update(self, dummy=None):
-
-        minders      = []
-        [minders.extend(tr.jobMinders) for tr in self.testRuns]
-        nTotal    = len(minders)
-        nDone     = len([m for m in minders if m.done])
-
-        date = time.strftime('%d/%m/%y %H:%M', time.gmtime())
-        testerStatus = '%s\njobs %d done %d' % (date, nTotal, nDone)
-        
-        
-        js = toJavaScript(testerStatus)
-
-        of = open(self.ofname, 'w')
-        of.write(js)
-        of.close()
-
diff --git a/Tools/RunTimeTester/src/RTTRunMonitor.py.obsolete b/Tools/RunTimeTester/src/RTTRunMonitor.py.obsolete
deleted file mode 100755
index 9d478a30bc336dbcd9f166c25e499a2effcec11e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTRunMonitor.py.obsolete
+++ /dev/null
@@ -1,130 +0,0 @@
-import os.path,os, shutil, fnmatch, anydbm, sys
-from popen2 import Popen4
-from Tools import pathIsOwnedByMe, changePerm
-from RTTpath import RTTpath
-from shellCommand import shellCommand
-from formatCollection import formatCollection
-
-from Factory_SiteSpecifics import SiteSpecificsFactory
-from exc2string import exc2string
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-class RTTRunMonitor:
-    #    def __init__(self, dirWithSummaries, pathToMacro, site):
-    def __init__(self, dirWithSummaries, pathToMacro, paths):
-        self.dws         = str(dirWithSummaries)
-        self.paths       = paths
-        #        self.theSite     = site
-        self.macroName   = 'rttSummaryMacro.cxx'
-        self.pathToMacro = str(RTTpath(pathToMacro,self.macroName))
-
-        if not os.path.exists(self.pathToMacro):
-            msg = 'Unable to run the RTT run monitor macro! It does not exist in given area: ' +str(self.pathToMacro)
-            logger.error(msg)
-            raise Exception, msg
-        
-    # --------------------------------------------------------------------
-    
-    def createMonitoringPlots(self):
-        holddir = os.getcwd()
-        os.chdir(self.dws)
-        logger.debug('createMonitoringPlots: about to cwd to %s to run summary macro ' % self.dws)
-        self.copyMacroToCurrentDir()
-        logger.debug('createMonitoringPlots: about to call  runSummaryMacro')
-        self.runSummaryMacro()
-        logger.debug('createMonitoringPlots: returning to ori dir')
-        os.chdir(holddir)
-        logger.debug('createMonitoringPlots: done')
-
-    # --------------------------------------------------------------------
-
-    def copyMacroToCurrentDir(self):
-        src = self.pathToMacro
-        dest = os.path.join(os.getcwd(),os.path.basename(self.pathToMacro))
-        shutil.copy(src,dest)
-        print 'copyMacroToCurrentDir changing perm on %s to 0775' % dest
-        nperm = changePerm(dest, 0775)
-        print 'copyMacroToCurrentDir new perm %s' % nperm
-        
-    # --------------------------------------------------------------------
-    
-    def runSummaryMacro(self):
-        #        rootCmd         = SiteSpecificsFactory(logger).create(self.theSite).rootCommand()
-        macroName           = os.path.basename(self.pathToMacro) + '+'
-        #cmd             = rootCmd +' '+macro
-
-        cmds = self.ROOTenvSetupAndLaunch()
-        cmd = cmds['localSetupScriptCmd'] + cmds['releaseSetupScriptCmd'] + cmds['launchRootCmd'] + macroName
-        
-        logger.info('running rttSummaryMacro.cxx from %s ' % os.getcwd())
-        logger.info('cmd : '+cmd)
-
-        try:
-
-            sumMacroReply = shellCommand([cmd])
-
-            logger.info('RTT run monitoring, root command output \n' + formatCollection(sumMacroReply))
-            
-        except Exception, e:
-            msg = 'Unable to successfully run the summary macro: ' + str(self.pathToMacro)
-            logger.error(msg)
-            logger.error(exc2string(sys.exc_info()))
-            logger.error(str(e))
-            raise Exception, msg
-    
-    # --------------------------------------------------------------------
-    
-    def copyPSFiles(self, src, dest):
-        logger.info('Copying monitoring plots from '+src+' to '+dest)
-        for file in os.listdir(src):
-            if fnmatch.fnmatch(file, '*.ps'):
-                logger.info("found file "+ file)
-                srcFile = os.path.join(src, file)
-                desFile = os.path.join(dest, file)
-                shutil.copyfile(srcFile, desFile)
-                
-                changePerm(str(desFile),0775)
-
-    # --------------------------------------------------------------------
-    
-    def ROOTenvSetupAndLaunch(self):        
-        info = {}
-        aDir = os.path.join(self.paths.hardCoded.distBase,self.paths.hardCoded.nightlyPathFragment[self.paths.branch])
-        sendToDevNull = ' >& /dev/null;'
-        
-        if self.paths.releaseType=='project':
-            pathToLocalSetupScript = os.path.join(self.paths.workDirs['workBase/'],'cmtHome/setup.sh')
-            localScriptTags = ' -tag=opt,AtlasOffline,' + str(self.paths.release)
-            atlasOfflineRunTimeDir = os.path.join(aDir,'AtlasOffline',self.paths.release,'AtlasOfflineRunTime/')        
-            pathToAtlasOfflineRunTimeSetupScript = os.path.join(atlasOfflineRunTimeDir,'cmt/setup.sh')
-        else:
-            pathToLocalSetupScript = os.path.join(self.paths.workDirs['workBase/'],'setup.sh')
-            localScriptTags        = ' -tag=opt'
-            atlasOfflineRunTimeDir = os.path.join(aDir,self.paths.release,'AtlasOfflineRunTime/')
-            pathToAtlasOfflineRunTimeSetupScript = os.path.join(atlasOfflineRunTimeDir,
-                                                                os.listdir(atlasOfflineRunTimeDir)[0],
-                                                                'cmt/setup.sh')
-
-
-        info['localSetupScriptCmd']   = 'source ' + pathToLocalSetupScript + localScriptTags + sendToDevNull
-        info['releaseSetupScriptCmd'] = 'source ' + pathToAtlasOfflineRunTimeSetupScript + sendToDevNull
-        info['launchRootCmd']         = 'root -b -q '
-        return info
-
-    #def storeSummaryXML(file):
-    #    monDB  = anydbm.open('MonitoringDB.db', 'c')
-    #    key    = getKey()
-    #    monDB[key]= file
-        
-
-
-if __name__ == "__main__":
-    logger.info('RTT Monitoring')
-    rttMon = RTTRunMonitor('/home/en/RTT/Tools/RunTimeTester/RunTimeTester-00-00-72/src')
-
-    rttMon.createMonitoringPlots()
diff --git a/Tools/RunTimeTester/src/RTTRunSummariser2.py.obsolete b/Tools/RunTimeTester/src/RTTRunSummariser2.py.obsolete
deleted file mode 100755
index f46b91063d3c5e4f8fddc03fa8d70a95a60aeb52..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTRunSummariser2.py.obsolete
+++ /dev/null
@@ -1,297 +0,0 @@
-"""
-A class to provide an aml file summary of an RTT run
-"""
-import os, string, copy, anydbm, fnmatch
-from time import localtime, strftime, time
-from Evaluate import Evaluate
-from Tools import changePerm
-
-from getInheritanceHeirarchy import isWorkerMinder, isWatcherMinder
-import threading
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class RTTRunSummariserCallBackHandler:
-    """
-    Call back handler to allow (eg) minders to initiate an update
-    the monitoring plots.
-    """
-    
-    def __init__(self, tester):
-
-        self.updateEvery = 10
-        self.updateCount = 0
-        self.tester      = tester
-        
-        minders = []
-        [minders.extend(tr.jobMinders) for tr in tester.testRuns]
-        [m.setRunSummaryCallBackHandler(self) for m in minders]
-        
-    def update(self, importance):
-        if importance != 'high':
-            self.updateCount += 1
-            if self.updateCount != self.updateEvery:
-                return False # not yet time for low level update
-            else:
-                self.updateCount = 0 # reset and proceed with update
-                
-        return self.tester.updateMonitors()
-
-def sum(list):
-    try:
-        floats = [float(l) for l in list]
-    except:
-        return  0
-
-    sum = 0
-    for f in floats: sum += f
-
-    return sum
-
-class RTTRunSummariser:
-    def __init__(self, tester):
-
-        self.tester     = tester
-        self.outputDir  = tester.paths.resultsDirs['packages']
-
-        self.nJobs                = 0
-        self.nPackages            = 0
-        self.nTests               = 0
-        self.nTestsPassed         = 0
-
-        self.nJobsByPackage        = {}
-        self.nJobsDoneByPackage    = {}
-        self.nJobsOpCloseByPackage = {}
-        self.nJobsSuccByPackage    = {}
-        self.nJobsFailByPackage    = {}
-        self.nTestsSuccByPackage   = {}
-        self.nTestsFailByPackage   = {}
-        self.nPPSuccByPackage      = {}
-        self.nPPFailByPackage      = {}
-        self.nRetriesByPackage     = {}
-        self.nJobsTOByPackage      = {}
-        self.nTestsByPackage       = {}
-        self.jobsByPackage         = {}
-
-
-    def dirWithSummaries(self): return self.outputDir
-
-
-    def update(self, package, updatedMinders):
-        # need the dummy arg so we can run via mutex.lock (why do we need thew arg??)
-        logger.info('Update method starts for RTTRunSummariser')
-        document = self.tester.summaryDocument
-        if document is None:
-            logger.info('Tester summary document unavailable, returning')
-            return
-        docElement = document.documentElement
-        
-        # self.minders = []
-        # [self.minders.extend(tr.jobMinders) for tr in self.testRuns]
-
-        
-        self.nPackages = len(Evaluate('package',docElement))
-        self.nJobs = sum(Evaluate('package/nJobsInPackage/text()',docElement))
-
-
-        self.nTests = sum(
-            Evaluate('package/nTestsInPackage/text()', docElement))
-
-        
-        
-        self.nJobsByPackage        = {}
-        self.nJobsDoneByPackage    = {}
-        self.nJobsOpCloseByPackage = {}
-        self.nJobsSuccByPackage    = {}
-        self.nJobsFailByPackage    = {}
-        self.nPPSuccByPackage      = {}
-        self.nPPFailByPackage      = {}
-        self.nTestsSuccByPackage   = {}
-        self.nTestsFailByPackage   = {}
-        self.nRetriesByPackage     = {}
-        self.nJobsTOByPackage      = {}
-        self.nTestsByPackage       = {}
-        
-        
-
-        self.nTestsPassed   = sum(Evaluate('package/nTestsOKInPackage/text()',
-                                           docElement))
-
-        packageNodes = Evaluate('package', docElement)
-
-        for n in packageNodes:
-            name = Evaluate('packageName/text()', n).strip()
-            
-            self.nJobsByPackage[name]        = int(Evaluate('nJobsInPackage/text()', n))
-            self.nJobsDoneByPackage[name]    = int(Evaluate('nJobsDoneInPackage/text()', n))
-            self.nJobsOpCloseByPackage[name] = int(Evaluate('nJobsOpCloseInPackage/text()', n))
-            self.nJobsSuccByPackage[name]    = int(Evaluate('nJobsSuccessInPackage/text()', n))
-            self.nJobsFailByPackage[name]    = int(Evaluate('nJobsFailureInPackage/text()', n))
-            self.nPPSuccByPackage[name]      = int(Evaluate('nPPSuccessInPackage/text()', n))
-            self.nPPFailByPackage[name]      = int(Evaluate('nPPFailureInPackage/text()', n))
-            self.nTestsSuccByPackage[name]   = int(Evaluate('nTestsSuccessInPackage/text()', n))
-            self.nTestsFailByPackage[name]   = int(Evaluate('nTestsFailureInPackage/text()', n))
-            self.nRetriesByPackage[name]     = int(Evaluate('nRetriesInPackage/text()', n))
-            self.nJobsTOByPackage[name]      = int(Evaluate('nTimeOutsInPackage/text()', n))
-            self.nTestsByPackage[name]       = int(Evaluate('nTestsInPackage/text()', n))
-        
-
-        self.summaryToFile()
-        logger.info('Update method ends for RTTRunSummariser')
-
-
-
-    def finalise(self): pass
-    
-    def stringToFile(self, string, fn):
-
-        filePath = os.path.join(self.outputDir, fn)
-        file = open(filePath, 'w')
-        os.chmod(filePath, 02755)
-        file.write(string)
-        # try:
-        file.close()
-        #except:
-        #    print 'Error closing filepath',filePath
-
-        logger.debug('wrote %s' % filePath)
-
-    def summaryToFile(self):
-        self.stringToFile(self.__str__(),     'RTTSummary.txt')
-
-    def info(self):
-
-        # self.update()
-        timestring = strftime("%d %b %Y %H:%M:%S", localtime())
-
-
-        de = self.tester.summaryDocument
-        if de is None:
-            m = 'RTTSummariser: Tester summary doc is None, returning'
-            self.logger.info(m)
-            return
-        
-        de = self.tester.summaryDocument.documentElement
-        on = Evaluate('overview', de)[0]
-        
-        simpleInfo = (
-            ('time                 ', timestring),
-            ('time2                ', time()),
-            ('NICOStime            ', Evaluate('nicosDate1/text()',     on)),
-            ('NICOStime2           ', Evaluate('nicosDate2/text()',     on)),
-            ('rttSrcDir            ', Evaluate('rttSrcDir/text()',      on)),
-            ('rttLibDir            ', Evaluate('rttLibDir/text()',      on)),
-            ('release              ', Evaluate('release/text()',        on)),
-            ('refRelease           ', Evaluate('refRelease/text()',     on)),
-            ('runType              ', Evaluate('runType/text()',        on)),
-            ('cmtConfig            ', Evaluate('cmtConfig/text()',      on)),
-            ('branch               ', Evaluate('originalBranch/text()', on)),
-            ('nPackages            ', self.nPackages),
-            )
-        
-
-        label = string.ljust('totalRTT', 30)
-
-        def sumList(list):
-            sum = 0
-            for l in list: sum += int(l)
-            return sum
-        
-        totalInfo = (label,
-                     sumList(self.nJobsByPackage.values()),
-                     sumList(self.nRetriesByPackage.values()),
-                     sumList(self.nJobsTOByPackage.values()),
-                     sumList(self.nJobsSuccByPackage.values()),
-                     sumList(self.nJobsFailByPackage.values()),
-                     sumList(self.nJobsOpCloseByPackage.values()),
-                     sumList(self.nJobsDoneByPackage.values()),
-                     sumList(self.nTestsByPackage.values()),
-                     sumList(self.nTestsSuccByPackage.values())
-                     )
-        justifiedKeys = [string.ljust(k, 30) for k in self.nJobsByPackage.keys()]
-        jobGroupInfo  = zip(
-            justifiedKeys,
-            self.nJobsByPackage.values(),
-            self.nRetriesByPackage.values(),
-            self.nJobsTOByPackage.values(),
-            self.nJobsSuccByPackage.values(),
-            self.nJobsFailByPackage.values(),
-            self.nJobsOpCloseByPackage.values(),
-            self.nJobsDoneByPackage.values(),
-            self.nTestsByPackage.values(),
-            self.nTestsFailByPackage.values()
-            )
-
-        return simpleInfo, totalInfo, jobGroupInfo
-
-    
-    def __str__(self):
-
-        timestring = strftime("%d %b %Y %H:%M:%S", localtime())
-
-        simpleInfo, totalInfo, jobGroupInfo = self.info()
-
-        line = '-'*40+'\n'
-        s  = line
-        s += '|'+string.center('RunTimeTester RunSummary', 28)+'|\n'
-        s += '|'+string.center('%20s' % timestring, 28)       +'|\n'
-        s += line
-        s += '\n'
-
-        for item in simpleInfo:
-            s+='%20s    %40s\n' % item
-            
-        s += '\n'
-        s += 'Counts by Group:\n\n'
-        spacer = ' '*30
-        s += '%s%s%s%s%s%s%s%s%s%s\n' % (string.ljust('Package',     30),
-                                         string.rjust('Jobs',        10),
-                                         string.rjust('Job',         10),
-                                         string.rjust('Job',         10),
-                                         string.rjust('Jobs',        10),
-                                         string.rjust('Jobs',        10),
-                                         string.rjust('Operator',    10),
-                                         string.rjust('Process',     10),
-                                         string.rjust('Tests',       10),
-                                         string.rjust('Tests',       10)
-                                         )
-        s += '%s%s%s%s%s%s%s%s%s%s\n\n' % (string.ljust('name',      30),
-                                           string.rjust('submitted', 10),
-                                           string.rjust('retries',   10),
-                                           string.rjust('timeouts',  10),
-                                           string.rjust('success',   10),
-                                           string.rjust('fail',      10),
-                                           string.rjust('closeDown', 10),
-                                           string.rjust('complete',  10),
-                                           string.rjust('total',     10),
-                                           string.rjust('pass',      10) )
-        
-
-        logger.debug("No of job groups %d" % len(self.nJobsByPackage.keys()))
-        fString = '%30s%10d%10d%10d%10d%10d%10d%10d%10d%10d\n'
-        s+= fString % totalInfo
-        for group in jobGroupInfo:
-            s+=fString % group
-
-        return s
-
-if __name__ == '__main__':
-    oDir = '/local_disk/sherwood'
-
-    from xml.dom.minidom import parse
-    class A:
-        def __init__(self):
-            self.resultsDirs = {'packages':'/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc3-gcc323-opt/RTTSummary.xml'}
-    class Tester:
-        def __init__(self, paths):
-            self.summaryDocument = parse('/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc3-gcc323-opt/RTTSummary.xml')
-            self.paths = paths
-    s = RTTRunSummariser(Tester(A()))
-    s.update()
-
-
-        
-    
-        
diff --git a/Tools/RunTimeTester/src/RTTRunner.py b/Tools/RunTimeTester/src/RTTRunner.py
deleted file mode 100644
index c4733319e34e9f78110985941fff1c2282e89c8c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTRunner.py
+++ /dev/null
@@ -1,84 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-import os
-import os.path
-
-from RTTFactory                  import RTTFactory
-from RTTFactory2                 import RTTFactory2
-from installWebFiles             import installWebFiles
-from installBaseDirectories      import installBaseDirectories
-from installStartTimeStamp       import installStartTimeStamp
-from adjustPathEnvVar            import adjustPathEnvVar
-from exc2string2                 import exc2string2
-
-import shelve
-
-from LoggerSetup import setupLoggers
-import logging
-os.environ['RTTLogDir'] = setupLoggers(os.environ['RTTLogDir'])
-logger = logging.getLogger('rtt')
-
-def handleError(e):
-    print "%s  error - please fix:\nException:\n%s\nTraceback:\n%s" % (e.__class__.__name__, str(e), exc2string2())
-    sys.exit(0)
-
-class Runner:
-    def __init__(self):
-        self.RTTSrcDir                        = os.environ['RTTSrcDir']
-        self.RTTLibDir                        = os.environ['RTTShareDir']
-        self.topLevelRTTconfFileLocation      = os.environ['RTTCfgFile']
-        self.logDir                           = os.environ['RTTLogDir']
-        self.useDB                            = os.environ.has_key('RTTFactoryUseDB')
-
-    def makeTester(self):
-        db = {}
-        if self.useDB:
-            dbPath = 'RTTFactory.dbm'
-            db =  shelve.open(dbPath, 'c')
-            logger.info('Using an RTT object database for this run: %s' % dbPath)
-        else:
-            db = {}
-            logger.info('NOT using an RTT object database for this run')
-                           
-        
-        try:
-            factory  = RTTFactory(self.RTTSrcDir, self.RTTLibDir, self.topLevelRTTconfFileLocation, db, self.logDir)
-        except Exception, e:
-            handleError(e)
-
-        # logger = factory.logger
-        
-        adjustPathEnvVar(factory.siteSpecificsObj)
-        
-        sys.path.append(factory.noPackagePaths.RTTLibDir)
-        
-        installWebFiles(factory.noPackagePaths.RTTLibDir,
-                        factory.noPackagePaths.resultsBasePath)
-        installBaseDirectories(factory.noPackagePaths, logger)
-        installStartTimeStamp(factory.noPackagePaths,  factory.startTime2, logger)
-        
-        try:
-            factory2 = RTTFactory2(factory, self.logDir, db)
-        except Exception, e:
-            handleError(e)
-        
-        factory2.closeDownKeyMaker.write() # install file to allow operator to close down the RTT
-        
-        tester = factory2.tester
-
-        if self.useDB: db.close()
-        return tester
-
-
-    def runTester(self, tester): tester.execute()
-
-if __name__ == '__main__':
-    r = Runner()
-    t = r.makeTester()
-    try:
-        sys.argv[1]
-        print 'Not running the instantiated Tester object as you gave a CL argument.'
-    except:
-        r.runTester(t)
-    
diff --git a/Tools/RunTimeTester/src/RTTRunner_NoEnv.py b/Tools/RunTimeTester/src/RTTRunner_NoEnv.py
deleted file mode 100644
index 4a668f90754f4f91dd74c1575827d6ab774a36a6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTRunner_NoEnv.py
+++ /dev/null
@@ -1,95 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-import os
-import os.path
-
-from RTTFactory                  import RTTFactory
-from RTTFactory2                 import RTTFactory2
-from installWebFiles             import installWebFiles
-from installBaseDirectories      import installBaseDirectories
-from installStartTimeStamp       import installStartTimeStamp
-from adjustPathEnvVar            import adjustPathEnvVar
-from exc2string2                 import exc2string2
-from makeLocalLog                import makeLocalLog
-from Tools                       import now
-from Tools                       import mkdirIfNew
-
-import shelve
-
-def handleError(e):
-    print "%s  error - please fix:\nException:\n%s\nTraceback:\n%s" % (e.__class__.__name__, str(e), exc2string2())
-    sys.exit(0)
-
-class Runner:
-    def __init__(self, RTTSrcDir, RTTShareDir, RTTCfgFile, RTTLogDir, useDB, name):
-
-        #self.RTTSrcDir                        = os.environ['RTTSrcDir']
-        #self.RTTLibDir                        = os.environ['RTTShareDir']
-        #self.topLevelRTTconfFileLocation      = os.environ['RTTCfgFile']
-        #self.logDir                           = os.environ['RTTLogDir']
-        #self.useDB                            = os.environ.has_key('RTTFactoryUseDB')
-
-        self.RTTSrcDir                        = RTTSrcDir
-        self.RTTLibDir                        = RTTShareDir
-        self.topLevelRTTconfFileLocation      = RTTCfgFile
-        self.logDir                           = os.path.normpath(os.path.join(RTTLogDir, 'logfiles_'+now()))
-        self.useDB                            = useDB
-        self.name                             = name
-
-        mkdirIfNew(self.logDir)
-        self.logger = makeLocalLog(self.logDir,
-                                   'RTTRunnerError',
-                                   'RTTRunnerWarning',
-                                   'RTTRunnerInfo',
-                                   'RTTRunnerDebug',
-                                   name+'_log'
-                                   )
-
-
-    def makeTester(self):
-        db = {}
-        if self.useDB:
-            dbPath = 'RTTFactory.dbm', 'c'
-            db =  shelve.open(dbPath, 'c')
-            self.logger.info('Using an RTT object database for this run: %s' % dbPath)
-        else:
-            db = {}
-            self.logger.info('NOT using an RTT object database for this run')
-                           
-        
-        try:
-            factory  = RTTFactory(self.RTTSrcDir, self.RTTLibDir, self.topLevelRTTconfFileLocation, db, self.logDir)
-        except Exception, e:
-            handleError(e)
-
-        # logger = factory.logger
-        
-        adjustPathEnvVar(factory.siteSpecificsObj)
-        
-        sys.path.append(factory.noPackagePaths.RTTLibDir)
-        
-        installWebFiles(factory.noPackagePaths.RTTLibDir,
-                        factory.noPackagePaths.resultsBasePath)
-        installBaseDirectories(factory.noPackagePaths, self.logger)
-        installStartTimeStamp(factory.noPackagePaths,  factory.startTime2, self.logger)
-        
-        try:
-            factory2 = RTTFactory2(factory, self.logDir, db)
-        except Exception, e:
-            handleError(e)
-        
-        factory2.closeDownKeyMaker.write() # install file to allow operator to close down the RTT
-        
-        tester = factory2.tester
-
-        if self.useDB: db.close()
-        return tester
-
-
-    def runTester(self, tester): tester.execute()
-
-if __name__ == '__main__':
-    r = Runner()
-    r.runTester(r.makeTester())
-    
diff --git a/Tools/RunTimeTester/src/RTTSException.py b/Tools/RunTimeTester/src/RTTSException.py
deleted file mode 100755
index 5bce7fc2d548e22f0f8dbd685da2cc5e1d5eba84..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTSException.py
+++ /dev/null
@@ -1,29 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""Exception class that carries an optional datum, eg error string"""
-
-class NewRTTError(Exception):
-  def __init__(self, msg=''):
-    try:
-      self.msg = str(msg)
-    except:
-      self.msg = 'Input message not a string'
-  def __str__(self): return self.msg
-    
-class RTTCodingError(NewRTTError):
-    def __init__(self, msg):
-        NewRTTError.__init__(self, msg)
-
-class RTTInputError(NewRTTError):
-    def __init__(self, msg):
-        NewRTTError.__init__(self, msg)
-
-class RTTSystemResponseError(NewRTTError):
-    def __init__(self, msg):
-        NewRTTError.__init__(self, msg)
-
-class RTTTimeOut(NewRTTError):
-    def __init__(self, msg):
-        NewRTTError.__init__(self, msg)
-
diff --git a/Tools/RunTimeTester/src/RTTSummForRoot.py.obsolete b/Tools/RunTimeTester/src/RTTSummForRoot.py.obsolete
deleted file mode 100755
index af9a209f2939f9ce6da03134d3a8263651941a55..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTSummForRoot.py.obsolete
+++ /dev/null
@@ -1,393 +0,0 @@
-from Evaluate            import Evaluate, getText
-from xml.dom.minidom     import parse
-from Tools               import unique, setFileProtection
-from RTTConfigParser     import isValid
-from formatCollection    import formatCollection
-
-import copy, string, os
-
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-"""
-A script that takes a list of possible locations of RTT summary files.
-RTT summary files are xml files.
-The existence and validity against a DTD of the files is checked.
-
-A data structure is set up using the contents of the files.
-
-data structure:
-
-identifier = [runType, cmtConfig, branch]
-dict = 
-{identifier: releaseDict}
-releaseDict =
-{release: jobGroupDict}
-jobGroupDict =
-{jobGroup: dataDict}
-dataDict = 
-{
-"njobsLabel",  njobs,
-"njSuccLabel", njSucc
-"ntestsLabel", ntests,
-"ntSuccLabel", ntSucc
-
-
-This data structure allows the reshuffling of the RTT summary information
-to ease the wring of the root macro used to display the results.
-
-The results for a giveb identifier (see above)
-are written to a file specified in the method output().
-
-A list of the files written out is written to rttSummaryList.txt
-to tell the root macro its input files.
-"""
-
-def makeIdentifier(element):
-    """
-    return a list of contents for selected elements which are
-    immmediate subelements of the input subelement.
-    """
-    # DTD guarantees exactly one of the following simple tags
-    tags=['runType', 'cmtConfig', 'branch']
-
-    id = []
-    for tag in tags:
-        id.append(str(getText(Evaluate(tag, element)[0])))
-
-    return str(id)
-
-def getJobGroupInfo(element, dict):
-    jobGroupName = Evaluate('jobGroupName', element)
-    dataDict = {}
-    dataDict['nJobs']           = Evaluate('nJobs', element)
-    dataDict['nJobsSuccessful'] = Evaluate('nJobsSuccesful', element)
-    dataDict['nTests']          = Evaluate('nTests', element)
-    dataDict['nTestsPassed']    = Evaluate('nTestsPassed', element)
-    dict[jobGroupName]          = dataDict
-    
-def allJobGroups(dict):
-    # find all mentioned jobGroups:
-    #first, make a list of dictionaires with jobGroup as the key:
-    releaseDicts = dict.values()
-    groupDicts = []
-    [groupDicts.extend(d.values()) for d in releaseDicts]
-    jobGroups = []
-    [jobGroups.extend(dict.keys()) for dict in groupDicts]
-    return  unique(jobGroups)
-
-def zeroDataDict():
-    dataDict = {}
-    dataDict['nJobs']           = 0
-    dataDict['nJobsSuccessful'] = 0
-    dataDict['nTests']          = 0
-    dataDict['nTestsPassed']    = 0
-    return dataDict
-
-def fillInZeros(dict):
-    # find all mentioned jobGroups:
-    #first, make a list of dictionaires with jobGroup as the key:
-    jobGroups = allJobGroups(dict)
-
-    rlDicts = dict.values()
-    jgDicts = []
-    [jgDicts.extend(d.values()) for d in rlDicts]
-
-    for d in jgDicts:
-        for g in jobGroups:
-            if g not in d.keys(): d[g]=zeroDataDict
-
-def makeFileName(key):
-    fn = key[1:-1]
-    fn = string.replace(fn,', ','_')
-    fn = string.replace(fn,"'",'')
-    fn = 'RTTSummary_'+fn+'.dat'
-    return fn
-
-def summaryToString(doms, odir, dict):
-    """
-    Write the summary to a string suitable for writing to a file.
-    The data structure dict is descripbed at the top of the module.
-    """
-
-    #
-    # releasesInOrder is a list of ntuples
-    # ntuple = (timeInSeconds, releaseLabels, timeAndDateString)
-    releasesInOrder = orderedReleases(doms)
-
-    lines = []
-
-
-    # Now running from the RTT. Should not be handling more than one value
-    # for the identifier (see notes at the top of the module) at a time.
-    assert( len(dict.keys()) <= 1)
-    
-    summaries = {}
-    for key in dict.keys():
-        summaryFileName = os.path.join(odir, makeFileName(key))
-        lines = [] # this will be the file contents
-        lines.append('%d\n'% len(releasesInOrder)) # number of releases
-        # a line with each release name
-        [lines.append('%s\n' % str(rel[1])) for rel in releasesInOrder]
-        # a line with each release time
-        [lines.append('%s\n' % str(rel[2])) for rel in releasesInOrder]
-        # retrieve the information for the releases
-        jgDict = dict[key]
-        # a line giving the amount of info that is being stored
-        lines.append('%d\n' % len(jgDict.keys()))  
-        for jg in jgDict.keys():
-            rlDict = jgDict[jg]
-            # print rlDict
-            lines.append(jg+'\n')
-            jobsLine        = ''
-            jobsSuccessLine = ''
-            testsLine       = ''
-            testsPassedLine = ''
-            for release in releasesInOrder:
-                dataDict = rlDict[release[1]]
-                jobsLine        += ' %d' % int(dataDict['nJobs'])
-                jobsSuccessLine += ' %d' % int(dataDict['nJobsSuccessful'])
-                testsLine       += ' %d' % int(dataDict['nTests'])
-                testsPassedLine += ' %d' % int(dataDict['nTestsPassed'])
-            lines.append(jobsLine+'\n')
-            lines.append(jobsSuccessLine+'\n')
-            lines.append(testsLine+'\n')
-            lines.append(testsPassedLine+'\n')            
-
-        summaryFileContent = ''
-        for line in lines: summaryFileContent += line
-        summaries[summaryFileName] =  summaryFileContent
-
-
-    # make a list of the summary files (this is read in by the root
-    # macro to get the names of the files to read'
-
-    fileContent = ''
-    for key in summaries.keys():
-        fileContent += key+'\n'
-
-    fileName = os.path.join(odir, 'rttSummaryList.txt')
-
-    summaries[fileName] = fileContent
-    return summaries
-
-def output(summaries):
-    """
-    Write the list of output files to a list file.
-    Write one output file per identifier
-    The output files contain repeated
-
-    summaries is a dictionary of file name: file content
-    """
-
-    for fileName in summaries.keys():
-        logger.debug('Writing summary file %s' % fileName)
-        ofile = open(fileName, 'w')
-        ofile.write(summaries[fileName])
-        ofile.close()
-        setFileProtection(fileName,02775)
-
-    
-
-        # def output(doms, dict, odir):
-        #     """
-        #     Write the list of output files to a list file.
-        #     Write one output file per identifier
-        #     The output files contain repeated 
-        #     """
-        # 
-        #     logger.debug('output dir %s' % odir)
-        # 
-        #     #
-        #     # releasesInOrder is a list of ntuples
-        #     # ntuple = (timeInSeconds, releaseLabels, timeAndDateString)
-        #     releasesInOrder = orderedReleases(doms)
-        #                                   
-        #     for key in dict.keys():
-        #         fn = os.path.join(odir, makeFileName(key))
-        #         logger.info ('Opening monitoring info for root fn = %s' % fn)
-        #         ofile = open(fn, 'w')
-        #         ofile.write('%d\n'% len(releasesInOrder))
-        #         [ofile.write('%s\n' % str(rel[1])) for rel in releasesInOrder]
-        #         [ofile.write('%s\n' % str(rel[2])) for rel in releasesInOrder]
-        #         jgDict = dict[key]
-        #         ofile.write('%d\n' % len(jgDict.keys()))  
-        #         for jg in jgDict.keys():
-        #             rlDict = jgDict[jg]
-        #             # print rlDict
-        #             ofile.write(jg+'\n')
-        #             jobsLine        = ''
-        #             jobsSuccessLine = ''
-        #             testsLine       = ''
-        #             testsPassedLine = ''
-        #             for release in releasesInOrder:
-        #                 dataDict = rlDict[release[1]]
-        #                 jobsLine        += ' %d' % int(dataDict['nJobs'])
-        #                 jobsSuccessLine += ' %d' % int(dataDict['nJobsSuccessful'])
-        #                 testsLine       += ' %d' % int(dataDict['nTests'])
-        #                 testsPassedLine += ' %d' % int(dataDict['nTestsPassed'])
-        #             ofile.write(jobsLine+'\n')
-        #             ofile.write(jobsSuccessLine+'\n')
-        #             ofile.write(testsLine+'\n')
-        #             ofile.write(testsPassedLine+'\n')            
-        #         ofile.close()
-        #         setFileProtection(fn,0775)
-        # 
-        # 
-
-def dataDict(el):
-    keys = ['nJobs', 'nJobsSuccessful', 'nTests', 'nTestsPassed']
-    dict = {}
-    for key in keys:
-        dict[key] = getText(Evaluate(key, el)[0])
-    return dict
-
-def groupsDict(dictToReplace, element):
-    jobGroupElements = Evaluate('jobGroup', element)
-
-    for jobGroupElement in jobGroupElements:
-        group = getText(Evaluate('jobGroupName', jobGroupElement)[0])
-        dictToReplace[group] = releaseDict(dictToReplace[group],
-                                           element,
-                                           jobGroupElement)
-    return dictToReplace
-
-def releaseDict(dictToReplace, element, jobGroupElement):
-    release = getText(Evaluate('release', element)[0])
-    dictToReplace[release] = dataDict(jobGroupElement)
-    return dictToReplace
-
-def processDom(dict, element):
-    """
-    Different RTT runs can have the same identifier, but be differentiated
-    by release. Add the entry into the release dictionary for this release.
-    """
-    identifier = makeIdentifier(element)
-    groupsDict(dict[identifier], element)
-
-def findAllPathContents(doms, path):
-
-    jobNameElements = []
-    [jobNameElements.extend(Evaluate(path, dom)) for dom in doms]
-    return unique([getText(element) for element in jobNameElements])
-
-def findAllIdentifiers(doms):
-    return unique([makeIdentifier(dom) for dom in doms])
-
-
-def makeEmptyDict(ids, releases, groups):
-
-    relDict = {}
-    for release in releases:
-        relDict[release] = zeroDataDict()
-
-    groupDict = {}
-    for group in groups:
-        groupDict[group]=copy.copy(relDict)
-
-    dict = {}
-    for id in ids:
-        dict[str(id)] = copy.copy(groupDict)
-
-    return dict
-
-def orderedReleases(doms):
-    timedReleases = []
-    for dom in doms:
-        release = getText(Evaluate('release', dom)[0])
-        time    = getText(Evaluate('time2', dom)[0])
-        date    = getText(Evaluate('time', dom)[0])[:-3] # drop the seconds
-        date    = string.replace(date,' ','_')
-        timedReleases.append((time, release, date))
-
-    timedReleases.sort()
-    # return [release for (time, release) in timedReleases]
-    return timedReleases
-
-def doit(files, odir):
-    """
-    creates special format files for root processing with all the data
-    needed for the (no of jobs, no of tests) x (pass, fail) histograms.
-    the information is returned in the form os a dictionary of
-    {filename: filecontent} (file content is a string) for further processing.
-
-    Call to output writes the files (commented out)
-    """
-    # make a document for each xml file.
-    doms = [parse(fn) for fn in files]
-
-    # get a list of all releases in the system
-    allReleases    = findAllPathContents(doms, 'release')
-    allJobGroups   = findAllPathContents(doms, 'jobGroup/jobGroupName')
-    allIdentifiers = findAllIdentifiers(doms)
-    logger.debug('allJobGroups   %s' % allJobGroups)
-    logger.debug('allReleases    %s' % allReleases)
-    logger.debug('allIdentifiers %s' % allIdentifiers)
-    dict = makeEmptyDict(allIdentifiers, allReleases, allJobGroups)
-    logger.debug( 'empty dict:')
-    # logger.debug( formatCollection(dict))
-    # returns a master dictionary of raw date.
-    # the key for each entry is a dictionsary specifiying RTT run parameters:
-    # - kit or release, cmtConfig, and branch (the identifier)
-    # the value is a dictionary with key the job groups (including "total") and
-    # value a fixed set of data. currently njobs(ran, passed), ntests(ran, passed)
-    # the data has is raw because the zeros for corresponding to jobs groups
-    # not present for a given identifier have not been added.
-
-    [processDom(dict, dom) for dom in doms]
-    logger.debug( 'filled dict:')
-    # logger.debug( formatCollection(dict))
-
-
-    summaries = summaryToString(doms, odir, dict)
-    output(summaries)
-    return summaries
-
-def summaryByBranch(oDir, paths):
-    """
-    Collects all the summary files written out by the invidual RTT runs
-    which lie below the dir oDir. The caller's choice of oDir determines
-    such parameters as branch, cmtConfig etc as these are represented by (=have)
-    different directories.
-
-    We also asumme a relation between where the files are stored and odir.
-    """
-
-    
-    
-    possibleFiles = [os.path.join(oDir, dir, 'RTTSummary.xml')
-                     for dir in paths.legals.nightlies.values()]
-
-    print 'RTTSummForRoot, possible files =',possibleFiles
-        # 'atlrel_0/RTTSummary.xml',
-        # 'atlrel_1/RTTSummary.xml',
-        # 'atlrel_2/RTTSummary.xml',
-        # 'atlrel_3/RTTSummary.xml',
-        # 'atlrel_4/RTTSummary.xml',
-        # 'atlrel_5/RTTSummary.xml',
-        # 'atlrel_6/RTTSummary.xml'
-        # ]
-
-    files = [fp for fp in possibleFiles if os.path.exists(fp)]
-    logger.debug('No of files before validation %i', len(files))
-    
-    badFiles = [file for file in files if not isValid(file)]
-    [files.remove(file) for file in badFiles]
-    logger.debug('No of files after validation %s' % len(files))
-    doit(files, oDir)
-
-if __name__ == '__main__':
-    # writes output to cwd
-    branch =  'N.0.X'
-    sdir = '/unix/www_data/atlas/NightlyTests/opt/N.X.0/release'
-    odir = '.'
-
-    odir = '/unix/www/html/atlas/atlfast/NightlyTestsDev3/opt/N.X.0/release'
-    
-    collectFiles(branch, odir, sdir)
-    # collectFiles('N.O.X', '.', sdir)
-    #    collectFiles('N.X.O', '.', sdir)
-
diff --git a/Tools/RunTimeTester/src/RTTSummaryFilePathsWriter.py.obsolete b/Tools/RunTimeTester/src/RTTSummaryFilePathsWriter.py.obsolete
deleted file mode 100644
index dfabb4cb4d3bb205b56be6f94c48d054a58f828f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTSummaryFilePathsWriter.py.obsolete
+++ /dev/null
@@ -1,185 +0,0 @@
-# This class acts as the entrance to the system that writes out the xml files
-# used to generate web pages.
-# It manages incomoming summaries, establishes if the web summary is more recent than
-# others with the same parameter set.
-#
-# It then wraps most recent the summaries into a DOM document and passes them to PageWriter.
-# the document is destroyed after the web pages have been written.
-#
-
-import os, types, string
-import time
-# import page0DefaultFiles
-
-# from PageWriter import Page1Writer
-# from xml.dom.minidom            import parseString, parse
-
-# from Evaluate  import Evaluate
-from listFiles import listFiles
-# from XMLTools  import relPath
-
-from exc2string2 import exc2string2
-
-debug = 10
-
-# ================================================================
-# HELPER FUNCTIONS
-# ================================================================
-
-# taken from the python cookbook
-def unique(s):
-    """ return a list of the elements in s in arbitrary order,
-    but without duplicates"""
-
-    # get the special case of empty s out of the way rapidly
-    n = len(s)
-    if n==0: return []
-
-    # try using a dict - fast, usaually works
-    u={}
-    try:
-        for x in s: u[x]=1
-    except TypeError:
-        del u # move to next method
-    else:
-        return u.keys()
-
-    # can't hash elements: sort, and remove identical
-    # elements in a single pass
-
-    try:
-        t = list(s)
-        t.sort()
-    except TypeError:
-        del t # move to next method
-    else:
-        assert n>0
-        last = t[0]
-        lasti = i= 1
-        while i<n:
-            if t[i] != last:
-                t[lasti] = last= t[i]
-                lasti += 1
-            i += 1
-
-        return t[:lasti]
-
-    # only brute force remains
-
-    u = []
-    for x in s:
-        if x not in u:
-            u.append(x)
-    return u
-    
-def readSummaryPathsFile(resultsBaseDir, summaryListFileName):
-    """
-    set up a dictionary of location: document pairs
-    location is  a general concept, here represented by a relative path
-
-    the locations of the summary files are stored in the file summaryFilePaths.txt
-    for rapid access. If this file does not exist, the file system is searched
-    for the summaryFiles.
-    """
-    
-    if os.path.exists(summaryListFileName):
-        handle=open(summaryListFileName, 'r')
-        summaryPaths = handle.readlines()
-        handle.close()
-        
-        summaryPaths = unique([s.strip() for s in summaryPaths])
-    else:
-        summaryPaths =  listFiles(resultsBaseDir, 'RTTSummary.xml')
-        of = open(summaryListFileName, 'w')
-        [of.write('%s\n' % l) for l in summaryPaths]
-        of.close()
-
-    summaryPaths = [s for s in summaryPaths if os.path.exists(s)]
-
-    return summaryPaths
-
-
-# ================================================================
-# MAIN STUFF
-# ================================================================
-
-import mutex
-accessLock = mutex.mutex()
-
-class RTTSummaryFilePathsWriterBase:
-    def __init__(self, resBaseDir):
-
-        self.resBaseDir = resBaseDir
-        self.summaryListFileName = os.path.join(self.resBaseDir, 'summaryFilePaths.txt')
-
-        # Look for all summary files below resBaseDir
-        self.pathsToSummaryFiles = readSummaryPathsFile(resBaseDir, self.summaryListFileName)
-
-        self.updates = []
-        
-    def update(self, pathToSummaryFile):
-        if debug>=50: print time.ctime(), 'WebPageWriter start update'
-        startTime = time.ctime()
-        timePair = (startTime,0)
-        self.updates.append(timePair)
-        accessLock.lock(self.lockedUpdate, pathToSummaryFile)
-
-        timePair = (startTime,time.ctime())
-        self.updates[-1] = timePair
-        if debug >=50: print time.ctime(), 'WebPageWriter ends update'
-        
-        return 0
-    
-    def lockedUpdate(self, pathToSummaryFile):
-        if debug>=100: print 'start lockedUpdate'
-
-
-        if os.path.isabs(pathToSummaryFile):
-            pathToSummaryFile = pathToSummaryFile.strip().split(self.resBaseDir)[1]
-            if pathToSummaryFile.startswith('/'): pathToSummaryFile = pathToSummaryFile[1:]
-            
-        if not pathToSummaryFile in self.pathsToSummaryFiles: # add it
-            self.pathsToSummaryFiles.append(pathToSummaryFile)
-            handle = open(self.summaryListFileName, 'r')
-            allFiles = [a.strip() for a in handle.readlines()]
-            allFiles = [a+'\n' for a in allFiles]
-            handle.close()
-            handle = open(self.summaryListFileName, 'w')
-            allFiles.append(pathToSummaryFile.strip()+'\n')
-            handle.writelines(allFiles)
-            handle.close()
-            
-        accessLock.unlock()
-        return 0
-
-    def getDocs(self):
-        if self.docs: return self.docs
-        return 0
-
-    def getOutputDir(self): return self.resBaseDir
-
-    def areYouAlive(self): return 'yes'
-    
-    def dumpTable(self):
-        counts = {}
-        s = str(self.updates) + '\n\n' + self.writer.dump(0, counts)
-        s += '\n\ncounts by page type \n\n'
-
-        ks = counts.keys()
-        ks.sort()
-        for k in ks:
-            s += '%10s %d\n' % (k, counts[k])
-        print s
-        return 0
-
-class RTTSummaryFilePathsWriter(RTTSummaryFilePathsWriterBase):
-    def __init__(self, resBaseDir):
-        RTTSummaryFilePathsWriterBase.__init__(self, resBaseDir)
-
-
-if __name__ == '__main__':
-    resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results'
-    writer = RTTSummaryFilePathsWriter(resBaseDir)
-
-    # print writer.dumpTable()
-    
diff --git a/Tools/RunTimeTester/src/RTTSummaryToText.py b/Tools/RunTimeTester/src/RTTSummaryToText.py
deleted file mode 100755
index 2777b41e69ee197c66f701d1c94b53f16626d368..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTSummaryToText.py
+++ /dev/null
@@ -1,119 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from xml.dom.minidom import parse
-from Evaluate import Evaluate
-import os.path, string
-
-#root = '/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc3-gcc323-opt'
-
-import xml.dom, time
-import xml.dom.minidom
-def makeDict(parentNode):
-    els = [e for e in parentNode.childNodes if e.nodeType == e.ELEMENT_NODE]
-    
-    dict = {}
-    def filldict(e):
-        dict[string.ljust(e.tagName.strip(), 30)] = Evaluate('text()', e).strip()#
-        
-    [filldict(e) for e in els]
-    return dict
-
-
-def RTTSummaryToText(dom):
-    
-    de = dom.documentElement
-
-
-    dict = makeDict(Evaluate('overview', de)[0])
-
-    lines = []
-    
-    line  = '|'+'-'*60 +'|'
-    vert  = '|'+' '*60 +'|'
-    title = '|'+string.center('RTT Run Summary', 60)+'|'
-    tim   = '|'+string.center(time.strftime('%d %B %Y  %H:%M  %Z', time.gmtime()), 60) +'|'
-    blank = ' '
-    lines.extend([line, vert, title, tim, vert, line, blank, blank])
-
-    [lines.append('%30s    %s'% i) for i in dict.items()]
-    lines.extend([blank, blank])
-
-    
-    tags = ['nPPSuccessInPackage',
-            'nRetriesInPackage',
-            'nTimeOutsInPackage',
-            'nBatchErrorsInPackage',
-            'packageName',
-            'nTestsSuccessInPackage',
-            'nJobsOpCloseInPackage',
-            'nTestsFailureInPackage',
-            'containerPackage',
-            'nJobsDoneInPackage',
-            'nJobsSuccessInPackage',
-            'nTestsInPackage',
-            'packageTag',
-            'nPPFailureInPackage',
-            'nJobsFailureInPackage'
-            ]
-
-    packageNodes = dom.getElementsByTagName('package')
-    precords = []
-    total = [string.ljust('totRTT', 30), '-', 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
-    for n in packageNodes:
-        record = ( (Evaluate('packageName/text()', n).strip()).ljust(30),
-                   Evaluate('phase/text()', n).strip(),
-                   int(Evaluate('nJobsTotalPackage/text()', n).strip()),
-                   int(Evaluate('nJobsInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsSuccessInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsFailureInPackage/text()', n).strip()),
-                   int(Evaluate('nTimeOutsInPackage/text()', n).strip()),
-                   int(Evaluate('nBatchErrorsInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsOpCloseInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsSuccessInPackage/text()', n).strip()),
-                   int(Evaluate('nTestsFailureInPackage/text()', n).strip()),
-                   int(Evaluate('nJobsDoneInPackage/text()', n).strip()))
-
-        for i in range(len(total[2:])):
-            j = i+2
-            total[j] += record[j]
-                  
-
-        precords.append(record)
-
-    precords.sort()
-    records = [tuple(total)]
-    records.extend(precords)
-
-    print
-    labels = (' ', 'pkge', 'jobs', 'jobs', 'jobs', 'jobs', 'jobs', 'jobs','oper', 'tests','tests', 'tests', 'process')
-    lines.append( '%30s %5s %5s %5s %5s %5s %5s %5s %5s %5s %5s %5s  %8s' % labels)
-    labels = (' ', 'phase', 'total', 'submt','succ','fail', 't_out', 'b_err', 'abort', 'total', 'succ', 'fail', 'complete')
-    lines.append('%30s %5s %5s %5s %5s %5s %5s %5s %5s %5s %5s %5s  %8s' % labels)
-
-    for r in records:
-
-        lines.append( '%s %5s %5d %5d %5d %5d %5d %5d %5d %5d %5d %5d  %8d' % r)
-
-    result = ''
-    for l in lines:
-        result += l  + '\n'
-
-    return result
-
-
-if __name__ == '__main__':
-
-
-    import sys
-    if sys.argv[2] == 'slc4' : conf = 'i686-slc4-gcc34-opt'
-    if sys.argv[2] == 'slc3' : conf = 'i686-slc3-gcc323-opt'
-    # root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/dev/build/i686-slc3-gcc323-opt' % sys.argv[1]
-    root = '/afs/cern.ch/atlas/project/RTT/Results/rel_%s/dev/build/%s' % (sys.argv[1], conf)
-    sys.path.append('../src')
-    fn = os.path.join(root, 'RTTSummary.xml')
-    dom = xml.dom.minidom.parse(fn)
-    print 'parsed ',fn
-
-    
-    print RTTSummaryToText(dom)
diff --git a/Tools/RunTimeTester/src/RTTTestRunner.py b/Tools/RunTimeTester/src/RTTTestRunner.py
deleted file mode 100755
index 6ca8a9886ccd351a3333c6b6d3df556e55f10ca4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTTestRunner.py
+++ /dev/null
@@ -1,73 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This is a wrapper class for RTT tests and actions.
-The purpose of the class is to allow a test to be run by a computing element,
-as opposed to the processor running the RTT. Instances of RTTT3estRunner have
-a test as an attribute. The instance is to be "shelved": in the run
-directory, "unshelved" on the computing node, run, reshelved, and examined
-by the RTT to determine the outcome.
-"""
-
-import shelve, sys, string, traceback, types
-
-class RTTTestRunner:
-    def __init__(self, dbPath, test, isAction):
-        # test is a 3-item tuple
-        self.dbPath   = dbPath
-        self.test     = test[0] # action/test instance
-        self.testId   = test[1]
-        self.userPos  = test[2]
-        self.result   = -1   # Failure
-        self.error    = False
-        self.errmsg   = ''
-        self.logger   = None # to be retrieved from the test
-        self.isAction = isAction # test or action?
-        
-    def autoShelve(self):
-        db = shelve.open(self.dbPath)
-        key = "(%s, %s)" % (str(self.userPos), str(self.testId))
-        db[key] = self
-        db.close()
-        
-    def run(self):
-        try:
-            self.result = self.test.run()
-        except:
-            self.error = True
-            list   = sys.exc_info()
-            self.errmsg  = str(list[0])
-            self.errmsg += str(list[1])
-            self.errmsg += string.join(traceback.format_tb(list[2]))
-
-        self.logger = self.test.logger
-        self.autoShelve()
-
-    def transferLog(self, logger):
-        if not self.logger: return # log not yet available
-
-        levels = {
-            'error'   : (logger.error, self.logger.errorL),
-            'warning' : (logger.warning, self.logger.warningL),
-            'info'    : (logger.info, self.logger.infoL),
-            'debug'   : (logger.debug, self.logger.debugL)
-            }
-       
-        actionOrTest = {True:'action', False:'test'}.get(self.isAction)
-
-        for level in levels.keys():
-            to_, from_ = levels[level]
-
-            for line in from_:
-                if type(line) in [types.ListType, types.TupleType]:
-                    for each_thing in line:
-                        message = '%s => %s: %s' % (actionOrTest, self.testId, each_thing.strip())
-                        to_(message)
-                elif type(line) is types.DictType:
-                    for k,v in line.items():
-                        message = '%s => %s: Key: %s, Value: %s' % (actionOrTest, self.testId, k, v)
-                        to_(message)
-                else:
-                    line = str(line)
-                    message = '%s => %s: %s' % (actionOrTest, self.testId, line.strip())
-                    to_(message)
diff --git a/Tools/RunTimeTester/src/RTTXMLConverter.py.obsolete b/Tools/RunTimeTester/src/RTTXMLConverter.py.obsolete
deleted file mode 100755
index 43475d8abfff595aa46d9d158ae1285f4b7f3bca..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTXMLConverter.py.obsolete
+++ /dev/null
@@ -1,432 +0,0 @@
-"""
-A class to provide an aml file summary of an RTT run
-"""
-import os, string, copy, anydbm, fnmatch
-from time              import localtime, strftime, time
-from StringIO          import StringIO
-from Tools             import changePerm
-from makeRTTDOM        import makeRTTDOM
-from NICOS             import NICOS
-from xml.dom.minidom   import parseString
-
-import xml.dom, sys
-from exc2string import exc2string
-from exc2string2 import exc2string2
-
-from Defs import RTTdefs
-
-from Evaluate import getText
-
-# -------------------------------------------------------------------------
-import Reporter
-globalReporter = Reporter.getGlobalReporter()
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-        
-class RTTXMLConverter:
-    def __init__(self, tester):
-
-
-        self.tester            = tester
-        # make the summary document easily available for other monitors.
-        self.document          = tester.summaryDocument
-        self.docElement        = self.document.documentElement
-        self.moniDBpath        = str(self.tester.paths.moniDB)        
-
-        self.update()
-        
-    def update(self):
-        self.enforceDBIntegrity()
-         
-        try:
-            [self.docElement.removeChild(child)
-             for child in self.docElement.childNodes]
-            self.document   = makeRTTDOM('RTTSummary')
-            self.docElement = self.document.documentElement
-            self.addTesterInfo()
-            self.addTestRuns()
-        except:
-            logger.error('Unable to convert this RTT run into XML.')
-            logger.error(exc2string2())
-            return
-        try:
-            self.xmlToDB()
-            self.xmlToFile()
-            # WebPageWriter(self.document)
-        except:
-            logger.error('Unable to output run XML to monitoring DB! (%s)' % str(self.moniDBpath))
-            logger.error('Consequence: cannot update web pages with latest info.')
-            logger.error(exc2string2())
-
-        self.tester.summaryDocument = self.document
-        self.enforceDBIntegrity()
-
-    # --------------------------------------------------------------------------
-
-    def openMoniDB(self):
-        handle = None
-        flag = 'n'
-        if os.path.exists(self.moniDBpath):
-            flag = 'c'
-        else:
-            logger.info('Monitoring DB (%s) does not exist, creating it.' % self.moniDBpath)
-
-        try:
-            handle = anydbm.open(self.moniDBpath, flag)
-            logger.info('Successfully opened the monitoring database.')
-        except:
-            logger.warning('Unable to open the monitoring DB!')
-            logger.warning(exc2string2())
-            
-        return handle
-    
-    # --------------------------------------------------------------------------
-                                                                                                                             
-    def corruptDBKeys(self,handle):
-        keys    = handle.keys()
-        keysThatAreCorrupt    = []
-        keysWithCorruptValues = []
-        
-        for key in keys:
-            isPrintableKey  = len([c for c in key if c not in string.printable])==0
-            if not isPrintableKey:
-                keysThatAreCorrupt.append(key)
-            else:
-                try:
-                    parseString(handle[key])
-                except:
-                    keysWithCorruptValues.append(key)
-                    
-        return (keysThatAreCorrupt,keysWithCorruptValues)
-
-    # --------------------------------------------------------------------------
-
-    def enforceDBIntegrity(self):
-        handle = self.openMoniDB()
-        if handle==None:
-            logger.warning('Unable to check moni DB integrity!')
-            return
-        
-        badKeys,keysWithBadValues = self.corruptDBKeys(handle)
-        if badKeys:
-            m = 'The monitoring DB was found to contain %d corrupt keys. These will be removed.' % len(badKeys)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in badKeys]
-        elif keysWithBadValues:
-            m = 'The monitoring DB was found to contain %d keys with corrupt values. These will be removed if not recoverable.' % len(keysWithBadValues)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in keysWithBadValues]
-        else:
-            m = 'Moni DB OK: no corrupt keys to remove.'
-            logger.info(m)
-
-        handle.close()
-        
-    # --------------------------------------------------------------------------
-
-    def recoverOrDeleteDBEntry(self,handle,key,badValue = True):
-        """For now, we just plain delete the corrupt entry.
-        An improvement could be to grab the file version, create a dom and stick in in the DB.
-        badValue = True if the value if corrupt, False if it is the key itself that is bad.
-        """
-        try:
-            del handle[key]
-            logger.warning('Removed monitoring DB key %s' % str(key))
-        except Exception, e:
-            logger.error('Tried, but unable to remove corrupt monitoring DB key %s' % str(key))
-            logger.error(exc2string2())
-                        
-    # --------------------------------------------------------------------------
-  
-    def addTextNode(self, parentNode, item):
-        try:
-            tn = self.document.createTextNode(str(item[1]).strip("'").strip())
-        except xml.dom.InvalidCharacterErr:
-            tn = self.document.createTextNode('invalid text')
-            
-        n  = self.document.createElement(str(item[0]).strip("'").strip())
-        
-        n.appendChild(tn)
-        parentNode.appendChild(n)
-
-    def addTesterInfo(self):
-        overviewNode = self.document.createElement('overview')        
-
-        els = ['build', 'originalBranch', 'site', 'platform',
-               'release', 'refRelease', 'releaseType', 'runType',
-               'rttSrcDir', 'rttLibDir']
-
-        els.sort()
-        items = [(el, self.tester.paths.__dict__[el]) for el in els]
-        [self.addTextNode(overviewNode, item) for item in items]
-
-        items = [
-            ('startTime1',       self.tester.startTime1),
-            ('startTime2',       self.tester.startTime2),
-            ('endTime1',         self.tester.endTime1),
-            ('endTime2',         self.tester.endTime2),
-            ('nicosDate1',       self.tester.nicosDateFormatted),
-            ('nicosDate2',       self.tester.nicosDate),
-            ('nPackages',        len(self.tester.testRuns) ),
-            ('statusText',       "%s %s" % (self.tester.statusText,
-                                            self.tester.statusTextTime)),
-            ('statusTextJSPath', self.tester.paths.runInfoPath),
-            ('resBaseDir',       self.tester.paths.resultsDirs['resBase/']),
-            ('resBuildDir',      self.tester.paths.resultsDirs['resBase/build']),
-            ('resBranchDir',     self.tester.paths.resultsDirs['resBase/build/branch']),
-            ('resRunTypeDir',    self.tester.paths.resultsDirs['resBase/build/branch/runType']),
-            ('resReleaseDir',    self.tester.paths.resultsDirs['resBase/build/branch/runType/release'])
-            ]
-        
-        [self.addTextNode(overviewNode, item) for item in items]
-        
-        self.docElement.appendChild(overviewNode)
-
-    def addTestRuns(self):
-
-        [self.addTestRun(tr) for tr in self.tester.testRuns]
-
-    def addTestRun(self, tr):
-
-        pNode = self.document.createElement('package')
-
-        # make a conversion table between attribute names of paths, and
-        # element names in the output xml file.
-        tagToPathsAttr = {'packageName':'package',
-                          'packageTag':'packageTag',
-                          'containerPackage':'containerPackage'}
-
-        # move information form the paths object into the document.
-        items = [(pair[0], tr.paths.__dict__[pair[1]])
-                 for pair in tagToPathsAttr.items()]
-        [self.addTextNode(pNode, item) for item in items]
-
-        # add the TestRun package status
-        self.addTextNode(pNode,('packageStatus',tr.ps.get()))
-
-        groups = tr.groupsInTestRun()
-
-        for g in groups:
-            minders  = [m for m in tr.jobMinders
-                        if m.jDescriptor.jobGroup == g]
-            nMinders = len(minders)
-            nDone    = len([m for m in minders if m.done])
-            nJobsOK  = len([m for m in minders
-                            if m.status.primary() == 'success'])
-
-
-            nTests   = reduce(lambda x, y: x+y,
-                              [len(m.tests) for m in minders],
-                              0)
-
-
-            statuses = []
-            [statuses.extend(m.testStatusList) for m in minders]
-            statuses = [s for s in statuses if s == RTTdefs.success]
-            nTestsOK = len(statuses)
-
-
-            nTimeOuts = len([m for m in minders if m.hasTimedOut()])
-            retryList = [m.errorStateCurRetry for m in minders]
-            nRetries  = reduce(lambda x, y: x+y, retryList, 0)
-                            
-            logger.debug('Creating job group node for doc: ' + str(self.document))
-            jgNode = self.document.createElement('jobgroup')
-            logger.debug('Created job group node for doc: ' + str(self.document))
-            
-            items = [
-                ('jobgroupname', g),
-                ('nJobs',        str(nMinders)),
-                ('nJobsDone',    str(nDone)),
-                ('nJobsOK',      str(nJobsOK)),
-                ('nTests',       str(nTests)),
-                ('nTestsOK',     str(nTestsOK)),
-                ('nTimeOuts',    str(nTimeOuts)),
-                ('nRetries',     str(nRetries))
-                ]
-
-            [self.addTextNode(jgNode, item) for item in items]
-
-            
-            def checkedAppendChild(jgNode,minder):
-                try:
-                    mNode = self.minderToXML(minder)
-                except :
-                    logger.debug('Error making mNode...')
-                    logger.debug(exc2string(sys.exc_info()))
-                    raise
-                
-                try:
-                    jgNode.appendChild(mNode)
-                except :
-                    logger.debug('Error appending mNode to jgNode...')
-                    logger.debug('jgNode ' + str(jgNode))
-                    logger.debug('minderNode ' + str(mNode))
-                    logger.debug(exc2string(sys.exc_info()))
-
-                    
-            [checkedAppendChild(jgNode,m) for m in minders]
-            
-            pNode.appendChild(jgNode)
-
-            self.docElement.appendChild(pNode)
-        
-    def minderToXML(self, minder):
-
-        # nested scopes are used to pass parameters to these nested functions
-        def findDataSets():
-
-            dsElement = self.document.createElement('datasets')
-            
-            try:
-                dataSets = minder.jDescriptor.datasets
-            except:
-                dataSets = {}
-
-            datasetInstances = []
-            [datasetInstances.extend(dataSets[k]) for k in dataSets]
-            for ds in datasetInstances:
-                tn = self.document.createTextNode(os.path.basename(ds.name.strip()))
-                el = self.document.createElement('dataset')
-                el.appendChild(tn)
-                dsElement.appendChild(el)
-
-            return dsElement
-
-        def findNtestSuccess():
-            l = len([s for s in minder.testStatusList
-                     if s == RTTdefs.success])
-            tn = self.document.createTextNode(str(l))
-            el = self.document.createElement('sUserTests')
-            el.appendChild(tn)
-            return el
-
-        def findKeepFiles():
-
-            resPath = minder.jDescriptor.resPath
-            
-            # minder.keepFilesPatterns is a list of dictionaries.
-            # kFilePatterns = [k['keepFileString']
-            #                 for k in minder.keepFilePatterns]
-            # minder.keepFiles is a dictionary of dictionaries.
-            kFilePatterns = []            
-            kFilePatterns.extend([minder.keepFiles[key]['keepFileString']
-                                 for key in minder.keepFiles.keys()])
-            # keep just the file names
-            kFilePatterns = [os.path.basename(k) for k in kFilePatterns]
-            
-            kFiles = []
-            [kFiles.extend(fnmatch.filter(os.listdir(resPath), p))
-             for p in kFilePatterns]
-            
-            return [os.path.join(resPath, k) for k in kFiles]
-
-        def makeKeepFileNode():
-            # logger.debug('Going to make element deliveredkeepfiles, self.document: ' +str( self.document))
-            el = self.document.createElement('deliveredkeepfiles')
-            # logger.debug('Have made element deliveredkeepfiles, self.document: ' + str(self.document))
-            keepFiles = findKeepFiles()
-            # logger.debug('After findKeepFiles method, self.document: ' + str( self.document))
-            
-            for k in keepFiles:
-                tn = self.document.createTextNode(os.path.basename(str(k).strip()))
-                el2 = self.document.createElement('keepfile')
-                # logger.debug('Have made element keepfile, self.document: ' +str( self.document))
-                el2.appendChild(tn)
-                el.appendChild(el2)
-            return el
-            
-            
-            
-        def makeTextNode((tagName, text)):
-            tn = self.document.createTextNode(str(text).strip())
-            el = self.document.createElement(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(self.document))
-        
-        mElement = self.document.createElement('minder')
-        textInfo = [
-            ('jobName',         minder.jDescriptor.name),
-            ('jobID',           minder.jDescriptor.jobSerialNumber),
-            ('status',          minder.status.primary()),
-            ('refRelease',      minder.jDescriptor.refRelease),
-            ('jobGroup',        minder.jDescriptor.jobGroup),
-            ('nUserTests',      len(minder.tests)),
-            ('resultsPath',     minder.jDescriptor.resPath),
-            ('exitStatus',      minder.exitStatus),
-            ('cpuTime',         minder.cpuTime),
-            ('mem',             minder.mem),
-            ('vmem',            minder.vmem),
-            ('wallTime',        minder.wallTime)
-            ]
-
-        [mElement.appendChild(makeTextNode(ti)) for ti in textInfo]
-
-        mElement.appendChild(findDataSets())
-        mElement.appendChild(findNtestSuccess() )
-        mElement.appendChild(makeKeepFileNode() )
-
-        def makeFlagNode((tagName, flag)):
-            if flag: return self.document.createElement(str(tagName))
-            return None
-
-        flagInfo = [
-            ('rttKitTestJob', minder.jDescriptor.rttKitTestJob),
-            ('rttPilotJob',   minder.jDescriptor.rttPilotJob),
-            ('rttATNJob',     minder.jDescriptor.rttPilotJob)
-            ]
-
-        for f in flagInfo:
-            node = makeFlagNode(f)
-            if node: mElement.appendChild(node)
-
-        logger.debug('Created, and returning, minder node for doc: ' + str(self.document))
-        
-        return mElement
-
-                
-    def docToString(self):
-        buffer = StringIO()
-        xml.dom.ext.Print(self.document, buffer)
-        return buffer.getvalue()
-        
-    def docToPrettyString(self):
-        buffer = StringIO()
-        xml.dom.ext.PrettyPrint(self.document, buffer)
-        return buffer.getvalue()
-        
-    def xmlToDB(self):
-    
-        if not self.tester.paths.moniDBenable: return
-
-        key      = self.tester.makeMoniDBKey()
-
-        dbHandle = self.openMoniDB()
-        if dbHandle==None:
-            logger.warning('Unable to store XML run data to moni DB.')
-            return
-        
-        dbHandle[key] = self.docToPrettyString()
-        dbHandle.close()
-
-        changePerm(self.moniDBpath,02775)
-
-        logger.debug('Successfully stored the XML run data in %s' % self.moniDBpath)
-
-    
-    def xmlToFile(self):
-
-        odir = self.tester.paths.resultsDirs[
-            'resBase/build/branch/runType/release']
-        of = open(os.path.join(odir, 'RTTSummary.xml'), 'w')
-        # print '------------'
-        # print 'xmlToFile ', of 
-        # print 'xmlToFile ', self.docToPrettyString()
-        # print '------------'
-        of.write(self.docToPrettyString())
-        of.close
diff --git a/Tools/RunTimeTester/src/RTTXMLConverter2.py.obsolete b/Tools/RunTimeTester/src/RTTXMLConverter2.py.obsolete
deleted file mode 100755
index 677b251ed1403f174797f3a8bc8ac8f0d9170921..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTXMLConverter2.py.obsolete
+++ /dev/null
@@ -1,431 +0,0 @@
-"""
-A class to provide an aml file summary of an RTT run
-"""
-import os, string, copy, anydbm, fnmatch
-from time              import localtime, strftime, time
-from StringIO          import StringIO
-from Tools             import changePerm
-from makeRTTDOM        import makeRTTDOM
-from NICOS             import NICOS
-from xml.dom.minidom   import parseString
-
-import xml.dom, sys
-from exc2string import exc2string
-from exc2string2 import exc2string2
-
-from Defs import RTTdefs
-
-from Evaluate import getText
-
-# -------------------------------------------------------------------------
-import Reporter
-globalReporter = Reporter.getGlobalReporter()
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-        
-class RTTXMLConverter:
-    def __init__(self, tester):
-
-
-        self.tester            = tester
-        # make the summary document easily available for other monitors.
-        self.document          = makeRTTDOM('rttSummary')
-        self.docElement        = self.document.documentElement
-        tester.summaryDocument = self.document
-        self.moniDBpath        = str(self.tester.paths.moniDB)        
-
-        self.update()
-        
-    def update(self):
-        self.enforceDBIntegrity()
-         
-        try:
-            self.document   = makeRTTDOM('RTTSummary')
-            self.docElement = self.document.documentElement
-            self.addTesterInfo()
-            self.addTestRuns()
-        except:
-            logger.error('Unable to convert this RTT run into XML.')
-            logger.error(exc2string2())
-            return
-        try:
-            self.xmlToDB()
-            self.xmlToFile()
-            # WebPageWriter(self.document)
-        except:
-            logger.error('Unable to output run XML to monitoring DB! (%s)' % str(self.moniDBpath))
-            logger.error('Consequence: cannot update web pages with latest info.')
-            logger.error(exc2string2())
-
-        self.tester.summaryDocument = self.document
-        self.enforceDBIntegrity()
-
-    # --------------------------------------------------------------------------
-
-    def openMoniDB(self):
-        handle = None
-        flag = 'n'
-        if os.path.exists(self.moniDBpath):
-            flag = 'c'
-        else:
-            logger.info('Monitoring DB (%s) does not exist, creating it.' % self.moniDBpath)
-
-        try:
-            handle = anydbm.open(self.moniDBpath, flag)
-            logger.info('Successfully opened the monitoring database.')
-        except:
-            logger.warning('Unable to open the monitoring DB!')
-            logger.warning(exc2string2())
-            
-        return handle
-    
-    # --------------------------------------------------------------------------
-                                                                                                                             
-    def corruptDBKeys(self,handle):
-        keys    = handle.keys()
-        keysThatAreCorrupt    = []
-        keysWithCorruptValues = []
-        
-        for key in keys:
-            isPrintableKey  = len([c for c in key if c not in string.printable])==0
-            if not isPrintableKey:
-                keysThatAreCorrupt.append(key)
-            else:
-                try:
-                    parseString(handle[key])
-                except:
-                    keysWithCorruptValues.append(key)
-                    
-        return (keysThatAreCorrupt,keysWithCorruptValues)
-
-    # --------------------------------------------------------------------------
-
-    def enforceDBIntegrity(self):
-        handle = self.openMoniDB()
-        if handle==None:
-            logger.warning('Unable to check moni DB integrity!')
-            return
-        
-        badKeys,keysWithBadValues = self.corruptDBKeys(handle)
-        if badKeys:
-            m = 'The monitoring DB was found to contain %d corrupt keys. These will be removed.' % len(badKeys)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in badKeys]
-        elif keysWithBadValues:
-            m = 'The monitoring DB was found to contain %d keys with corrupt values. These will be removed if not recoverable.' % len(keysWithBadValues)
-            logger.warning(m)
-            [self.recoverOrDeleteDBEntry(handle,k) for k in keysWithBadValues]
-        else:
-            m = 'Moni DB OK: no corrupt keys to remove.'
-            logger.info(m)
-
-        handle.close()
-        
-    # --------------------------------------------------------------------------
-
-    def recoverOrDeleteDBEntry(self,handle,key,badValue = True):
-        """For now, we just plain delete the corrupt entry.
-        An improvement could be to grab the file version, create a dom and stick in in the DB.
-        badValue = True if the value if corrupt, False if it is the key itself that is bad.
-        """
-        try:
-            del handle[key]
-            logger.warning('Removed monitoring DB key %s' % str(key))
-        except Exception, e:
-            logger.error('Tried, but unable to remove corrupt monitoring DB key %s' % str(key))
-            logger.error(exc2string2())
-                        
-    # --------------------------------------------------------------------------
-  
-    def addTextNode(self, parentNode, item):
-        try:
-            tn = self.document.createTextNode(str(item[1]).strip("'").strip())
-        except xml.dom.InvalidCharacterErr:
-            tn = self.document.createTextNode('invalid text')
-            
-        n  = self.document.createElement(str(item[0]).strip("'").strip())
-        
-        n.appendChild(tn)
-        parentNode.appendChild(n)
-
-    def addTesterInfo(self):
-        overviewNode = self.document.createElement('overview')        
-
-        els = ['build', 'originalBranch', 'site', 'platform',
-               'release', 'refRelease', 'releaseType', 'runType',
-               'rttSrcDir', 'rttLibDir']
-
-        els.sort()
-        items = [(el, self.tester.paths.__dict__[el]) for el in els]
-        [self.addTextNode(overviewNode, item) for item in items]
-
-        items = [
-            ('startTime1',       self.tester.startTime1),
-            ('startTime2',       self.tester.startTime2),
-            ('endTime1',         self.tester.endTime1),
-            ('endTime2',         self.tester.endTime2),
-            ('nicosDate1',       self.tester.nicosDateFormatted),
-            ('nicosDate2',       self.tester.nicosDate),
-            ('nPackages',        len(self.tester.testRuns) ),
-            ('statusText',       "%s %s" % (self.tester.statusText,
-                                            self.tester.statusTextTime)),
-            ('statusTextJSPath', self.tester.paths.runInfoPath),
-            ('resBaseDir',       self.tester.paths.resultsDirs['resBase/']),
-            ('resBuildDir',      self.tester.paths.resultsDirs['resBase/build']),
-            ('resBranchDir',     self.tester.paths.resultsDirs['resBase/build/branch']),
-            ('resRunTypeDir',    self.tester.paths.resultsDirs['resBase/build/branch/runType']),
-            ('resReleaseDir',    self.tester.paths.resultsDirs['resBase/build/branch/runType/release'])
-            ]
-        
-        [self.addTextNode(overviewNode, item) for item in items]
-        
-        self.docElement.appendChild(overviewNode)
-
-    def addTestRuns(self):
-
-        [self.addTestRun(tr) for tr in self.tester.testRuns]
-
-    def addTestRun(self, tr):
-
-        pNode = self.document.createElement('package')
-
-        # make a conversion table between attribute names of paths, and
-        # element names in the output xml file.
-        tagToPathsAttr = {'packageName':'package',
-                          'packageTag':'packageTag',
-                          'containerPackage':'containerPackage'}
-
-        # move information form the paths object into the document.
-        items = [(pair[0], tr.paths.__dict__[pair[1]])
-                 for pair in tagToPathsAttr.items()]
-        [self.addTextNode(pNode, item) for item in items]
-
-        # add the TestRun package status
-        self.addTextNode(pNode,('packageStatus',tr.ps.get()))
-
-        groups = tr.groupsInTestRun()
-
-        for g in groups:
-            minders  = [m for m in tr.jobMinders
-                        if m.jDescriptor.jobGroup == g]
-            nMinders = len(minders)
-            nDone    = len([m for m in minders if m.done])
-            nJobsOK  = len([m for m in minders
-                            if m.status.primary() == 'success'])
-
-
-            nTests   = reduce(lambda x, y: x+y,
-                              [len(m.tests) for m in minders],
-                              0)
-
-
-            statuses = []
-            [statuses.extend(m.testStatusList) for m in minders]
-            statuses = [s for s in statuses if s == RTTdefs.success]
-            nTestsOK = len(statuses)
-
-
-            nTimeOuts = len([m for m in minders if m.hasTimedOut()])
-            retryList = [m.errorStateCurRetry for m in minders]
-            nRetries  = reduce(lambda x, y: x+y, retryList, 0)
-                            
-            logger.debug('Creating job group node for doc: ' + str(self.document))
-            jgNode = self.document.createElement('jobgroup')
-            logger.debug('Created job group node for doc: ' + str(self.document))
-            
-            items = [
-                ('jobgroupname', g),
-                ('nJobs',        str(nMinders)),
-                ('nJobsDone',    str(nDone)),
-                ('nJobsOK',      str(nJobsOK)),
-                ('nTests',       str(nTests)),
-                ('nTestsOK',     str(nTestsOK)),
-                ('nTimeOuts',    str(nTimeOuts)),
-                ('nRetries',     str(nRetries))
-                ]
-
-            [self.addTextNode(jgNode, item) for item in items]
-
-            
-            def checkedAppendChild(jgNode,minder):
-                try:
-                    mNode = self.minderToXML(minder)
-                except :
-                    logger.debug('Error making mNode...')
-                    logger.debug(exc2string(sys.exc_info()))
-                    raise
-                
-                try:
-                    jgNode.appendChild(mNode)
-                except :
-                    logger.debug('Error appending mNode to jgNode...')
-                    logger.debug('jgNode ' + str(jgNode))
-                    logger.debug('minderNode ' + str(mNode))
-                    logger.debug(exc2string(sys.exc_info()))
-
-                    
-            [checkedAppendChild(jgNode,m) for m in minders]
-            
-            pNode.appendChild(jgNode)
-
-            self.docElement.appendChild(pNode)
-        
-    def minderToXML(self, minder):
-
-        # nested scopes are used to pass parameters to these nested functions
-        def findDataSets():
-
-            dsElement = self.document.createElement('datasets')
-            
-            try:
-                dataSets = minder.jDescriptor.datasets
-            except:
-                dataSets = {}
-
-            datasetInstances = []
-            [datasetInstances.extend(dataSets[k]) for k in dataSets]
-            for ds in datasetInstances:
-                tn = self.document.createTextNode(os.path.basename(ds.name.strip()))
-                el = self.document.createElement('dataset')
-                el.appendChild(tn)
-                dsElement.appendChild(el)
-
-            return dsElement
-
-        def findNtestSuccess():
-            l = len([s for s in minder.testStatusList
-                     if s == RTTdefs.success])
-            tn = self.document.createTextNode(str(l))
-            el = self.document.createElement('sUserTests')
-            el.appendChild(tn)
-            return el
-
-        def findKeepFiles():
-
-            resPath = minder.jDescriptor.resPath
-            
-            # minder.keepFilesPatterns is a list of dictionaries.
-            # kFilePatterns = [k['keepFileString']
-            #                 for k in minder.keepFilePatterns]
-            # minder.keepFiles is a dictionary of dictionaries.
-            kFilePatterns = []            
-            kFilePatterns.extend([minder.keepFiles[key]['keepFileString']
-                                 for key in minder.keepFiles.keys()])
-            # keep just the file names
-            kFilePatterns = [os.path.basename(k) for k in kFilePatterns]
-            
-            kFiles = []
-            [kFiles.extend(fnmatch.filter(os.listdir(resPath), p))
-             for p in kFilePatterns]
-            
-            return [os.path.join(resPath, k) for k in kFiles]
-
-        def makeKeepFileNode():
-            # logger.debug('Going to make element deliveredkeepfiles, self.document: ' +str( self.document))
-            el = self.document.createElement('deliveredkeepfiles')
-            # logger.debug('Have made element deliveredkeepfiles, self.document: ' + str(self.document))
-            keepFiles = findKeepFiles()
-            # logger.debug('After findKeepFiles method, self.document: ' + str( self.document))
-            
-            for k in keepFiles:
-                tn = self.document.createTextNode(os.path.basename(str(k).strip()))
-                el2 = self.document.createElement('keepfile')
-                # logger.debug('Have made element keepfile, self.document: ' +str( self.document))
-                el2.appendChild(tn)
-                el.appendChild(el2)
-            return el
-            
-            
-            
-        def makeTextNode((tagName, text)):
-            tn = self.document.createTextNode(str(text).strip())
-            el = self.document.createElement(str(tagName))
-            el.appendChild(tn)                                   
-            return el
-
-        # logger.debug('Creating minder node for doc: ' + str(self.document))
-        
-        mElement = self.document.createElement('minder')
-        textInfo = [
-            ('jobName',         minder.jDescriptor.name),
-            ('jobID',           minder.jDescriptor.jobSerialNumber),
-            ('status',          minder.status.primary()),
-            ('refRelease',      minder.jDescriptor.refRelease),
-            ('jobGroup',        minder.jDescriptor.jobGroup),
-            ('nUserTests',      len(minder.tests)),
-            ('resultsPath',     minder.jDescriptor.resPath),
-            ('exitStatus',      minder.exitStatus),
-            ('cpuTime',         minder.cpuTime),
-            ('mem',             minder.mem),
-            ('vmem',            minder.vmem),
-            ('wallTime',        minder.wallTime)
-            ]
-
-        [mElement.appendChild(makeTextNode(ti)) for ti in textInfo]
-
-        mElement.appendChild(findDataSets())
-        mElement.appendChild(findNtestSuccess() )
-        mElement.appendChild(makeKeepFileNode() )
-
-        def makeFlagNode((tagName, flag)):
-            if flag: return self.document.createElement(str(tagName))
-            return None
-
-        flagInfo = [
-            ('rttKitTestJob', minder.jDescriptor.rttKitTestJob),
-            ('rttPilotJob',   minder.jDescriptor.rttPilotJob),
-            ('rttATNJob',     minder.jDescriptor.rttPilotJob)
-            ]
-
-        for f in flagInfo:
-            node = makeFlagNode(f)
-            if node: mElement.appendChild(node)
-
-        logger.debug('Created, and returning, minder node for doc: ' + str(self.document))
-        
-        return mElement
-
-                
-    def docToString(self):
-        buffer = StringIO()
-        xml.dom.ext.Print(self.document, buffer)
-        return buffer.getvalue()
-        
-    def docToPrettyString(self):
-        buffer = StringIO()
-        xml.dom.ext.PrettyPrint(self.document, buffer)
-        return buffer.getvalue()
-        
-    def xmlToDB(self):
-    
-        if not self.tester.paths.moniDBenable: return
-
-        key      = self.tester.makeMoniDBKey()
-
-        dbHandle = self.openMoniDB()
-        if dbHandle==None:
-            logger.warning('Unable to store XML run data to moni DB.')
-            return
-        
-        dbHandle[key] = self.docToPrettyString()
-        dbHandle.close()
-
-        changePerm(self.moniDBpath,02775)
-
-        logger.debug('Successfully stored the XML run data in %s' % self.moniDBpath)
-
-    
-    def xmlToFile(self):
-
-        odir = self.tester.paths.resultsDirs[
-            'resBase/build/branch/runType/release']
-        of = open(os.path.join(odir, 'RTTSummary.xml'), 'w')
-        # print '------------'
-        # print 'xmlToFile ', of 
-        # print 'xmlToFile ', self.docToPrettyString()
-        # print '------------'
-        of.write(self.docToPrettyString())
-        of.close
diff --git a/Tools/RunTimeTester/src/RTTXMLFileSelector.py.obsolete b/Tools/RunTimeTester/src/RTTXMLFileSelector.py.obsolete
deleted file mode 100755
index b666612ffe7fda0c5e1b80a17028ed9a3a5924c4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTXMLFileSelector.py.obsolete
+++ /dev/null
@@ -1,98 +0,0 @@
-import anydbm, os.path, sys
-from xml.dom.ext     import PrettyPrint
-from Tools           import mkdirIfNew
-from StringIO        import StringIO
-from Evaluate        import getText
-
-def keyToDict(strList):
-    # ks is a list of a (string repr of a list of name val tuples) 
-    dict = {}
-    list  = eval(strList)
-    for l in list: dict[l[0]] = l[1]
-    return dict
-
-def isSubDict(selectorKey, key):
-    # checks if the all elements of the selectorKey
-    # are present in key and have the same values
-    # key can have more entries than selectorKey
-
-    
-    dict = {}
-    try:
-        for k in selectorKey.keys(): dict[k] = key[k]
-    except KeyError:
-        # dict passed  in has an element not in the dict used
-        # as the db key.
-        return False 
-         
-    
-    # check the values match
-    for k in selectorKey.keys():
-        if dict[k] != selectorKey[k]: return False
-
-    return True
-
-def  RTTXMLFileSelector(dbPath, selectorKey, logger, mustHave=['originalBranch']):
-    dbPath     = str(dbPath)
-
-    for m in mustHave:
-        if m not in selectorKey:
-            msg = 'Missing critical entry in key: %s: %s' % (m,
-                                                            str(selectorKey))
-            logger.error(msg)
-            return {}
-
-        
-    # bombs if opened without exist check and 'c' option
-    if os.path.exists(dbPath):
-        monidb = anydbm.open(dbPath, 'r')
-        logger.debug('RTTXMLFileSelector: opened db %s, has %d keys' %(dbPath, len(monidb.keys())))
-    else:
-        logger.error('RTTXMLFileSelector: cannot open db at ' %(dbPath))
-        return {}
-
-    # the db key is a stringed list of name val tuples
-    # convert it to a dictionary in preparartion for selction,
-    # and grab the data as well.
-    dbkeys = [(k, keyToDict(k))  for k in monidb.keys()]
-    logger.debug('RTTXMLFileSelector: found %d keys in database' %len(dbkeys))
-    # selectedData = [monidb[k[0]] for k in dbkeys if isSubDict(selectorKey, k[1])]
-
-    matchingKeys = [k for k in dbkeys if isSubDict(selectorKey, k[1])]
-    logger.debug('RTTXMLFileSelector: found %d matching keys in database' %len(matchingKeys))
-    logger.debug('selectorKey: %s' % str(selectorKey))
-    for k, k1 in dbkeys:
-        logger.debug('db Key: %s' % str(k1))
-
-    dict = {}
-    for keyTuple in matchingKeys:
-        key   = keyTuple[0]
-        value = monidb[key]
-        dict[key] = value
-
-    monidb.close()
-    # return a dict of matching keys and their values
-    logger.debug('returning %d files' % len(dict.keys()))
-    return dict
-        
-if __name__== '__main__':
-    # This allows a the xml files required for the web display to be
-    # run manually.
-
-
-    class Logger:
-        def error(self, text): print text
-    # The full path to the database file with the xml files for the RTT runs/
-    dbpath = '/local_disk/sherwood/RTT02/moniFile.db'
-
-    key    = {'originalBranch': u'N.X.0'}
-    keys   =  RTTXMLFileSelector(dbpath, key, Logger())
-
-    assert(len(keys) == 1)
-
-    key    = {'originalBranch': u'N.X.0'}
-    keys   = RTTXMLFileSelector(dbpath, key, Logger())
-    try:
-        assert(len(keys) == 0)
-    except AssertionError:
-        print 'OK'
diff --git a/Tools/RunTimeTester/src/RTTdict.py b/Tools/RunTimeTester/src/RTTdict.py
deleted file mode 100644
index dfc28e5e33903968bd5297ae30b8985532d95c9d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTdict.py
+++ /dev/null
@@ -1,81 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-from RTTSException   import RTTCodingError
-
-class RTTdict:
-    __module__ = os.path.splitext(os.path.basename(__file__))[0]
-
-    def __init__(self, rtt={}, user={}):
-        self.rtt  = rtt
-        self.user = user
-        
-    def __str__(self):
-        out = ''
-        for k,v in self.items():
-            out += '%s => %s\n' % (k,v)
-        return out
-    
-    def whoHasKey(self, k):
-        if k in self.rtt.keys(): return self.rtt
-        if k in self.user.keys(): return self.user
-        return None
-    
-    def __getitem__(self, k):
-        d = self.whoHasKey(k)
-        if d is not None: return d[k] 
-        m = 'KeyError: Unknown key %s' %k
-        # raise RTTCodingError(m)
-        raise KeyError(m)
-            
-    def __setitem__(self, k, v):
-        d = self.whoHasKey(k)
-        if d is not None: d[k] = v
-        else:
-            self.user[k] = v
-
-    def __delitem__(self, k):
-        if k in self.rtt.keys(): del self.rtt[k]
-        if k in self.user.keys(): del self.user[k]
-        m = 'KeyError: Unknown key %s' %k
-        raise RTTCodingError(m)
-    
-    def get(self, k, default=None):
-        try:
-            return self.__getitem__(k)
-        except KeyError:
-            return default
-
-    def __contains__(self, k):
-        return k in self.rtt or k in self.user
-
-    def has_key(self, k):
-        return self.__contains__(k)
-
-    def __len__(self):
-        return len(self.rtt)+len(self.user)
-
-    def keys(self):
-        k = self.rtt.keys()
-        k.extend(self.user.keys())
-        return k
-
-    def values(self):
-        v = self.rtt.values()
-        v.extend(self.user.values())
-        return v
-
-    def items(self):
-        i = self.rtt.items()
-        i.extend(self.user.items())
-        return [(k,v) for k,v in i]
-        
-    def update(self, d):
-        for k,v in d.items():
-            self.__setitem__(k,v)
-
-    def setdefault(self, k, v=None):
-        self.__setitem__(k,v)
-        return v
-
-        
diff --git a/Tools/RunTimeTester/src/RTTdict.py.obsolete b/Tools/RunTimeTester/src/RTTdict.py.obsolete
deleted file mode 100644
index 6ce645bde97c25c31e18953e39dbc65d810e09a5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTdict.py.obsolete
+++ /dev/null
@@ -1,78 +0,0 @@
-import os.path
-from RTTSException   import RTTCodingError
-
-class RTTdict:
-    __module__ = os.path.splitext(os.path.basename(__file__))[0]
-
-    def __init__(self, rtt={}, user={}):
-        self.rtt  = rtt
-        self.user = user
-
-    def __str__(self):
-        out = ''
-        for k,v in self.items():
-            out += '%s => %s\n' % (k,v)
-        return out
-    
-    def whoHasKey(self, k):
-        if k in self.rtt.keys(): return self.rtt
-        if k in self.user.keys(): return self.user
-        return None
-    
-    def __getitem__(self, k):
-        d = self.whoHasKey(k)
-        if d is not None: return d[k] 
-        m = 'KeyError: Unknown key %s' %k
-        raise RTTCodingError(m)
-            
-    def __setitem__(self, k, v):
-        d = self.whoHasKey(k)
-        if d is not None: d[k] = v
-        else:
-            self.user[k] = v
-
-    def __delitem__(self, k):
-        if k in self.rtt.keys(): del self.rtt[k]
-        if k in self.user.keys(): del self.user[k]
-        m = 'KeyError: Unknown key %s' %k
-        raise RTTCodingError(m)
-    
-    def get(self, k, default=None):
-        try:
-            return self.__getitem__(k)
-        except KeyError:
-            return default
-
-    def __contains__(self, k):
-        return k in self.rtt or k in self.user
-
-    def has_key(self, k):
-        return self.__contains__(k)
-
-    def __len__(self):
-        return len(self.rtt)+len(self.user)
-
-    def keys(self):
-        k = self.rtt.keys()
-        k.extend(self.user.keys())
-        return k
-
-    def values(self):
-        v = self.rtt.values()
-        v.extend(self.user.values())
-        return v
-
-    def items(self):
-        i = self.rtt.items()
-        i.extend(self.user.items())
-        return [(k,v) for k,v in i]
-        
-    def update(self, d):
-        for k,v in d.items():
-            self.__setitem__(k,v)
-
-    def setdefault(k, v=None):
-        self.__setitem__(k,v)
-        return v
-
-        
diff --git a/Tools/RunTimeTester/src/RTTmon.py.obsolete b/Tools/RunTimeTester/src/RTTmon.py.obsolete
deleted file mode 100755
index 20f9f1ae21990c7d1892827a89462bc200c8f55b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTmon.py.obsolete
+++ /dev/null
@@ -1,365 +0,0 @@
-"""
-RTTmon - a program to monitor RTT jobs.                  PS; Hopital Cantonal; 21/10/04
-
-Sends a signal to the RTT to initiate data flow from the RTT to RTTmon.
-Data is sent via xmlrpc. RTT is a server, RTTmon is a client.
-
-Data is extracted from JobMinders to NameAndStatus (a convenience only);
-A list of instances of NameAndStatus is pickeled  before transmission, and
-unpickled in RTTmon.
-
-RTTmon uses TCL/Tkinter to display the jobs names as Label widgets. The
-name of the job is shown as a Tkinter Label. The job status is shown by the
-color of the label.
-
-Updates can be obtained by hitting the update button.
-
-Auto updating is run on a separate thread to prevent the RTTmon from locking
-once the suto update loop is entered.
-"""
-import time, os, signal, pipes, pickle, popen2, sys, threading
-from Tkinter import *
-from xmlrpclib import ServerProxy
-import sys, profile
-
-#from logging import Logger,DEBUG, INFO,  WARNING, ERROR, CRITICAL
-from logging import Logger, debug, info,  warning, error, critical 
-import logging
-
-RTTcurrent="/Users/peter/RTT/RTT/Tools/RunTimeTester/RunTimeTester-00-00-18/src"
-sys.path.append(RTTcurrent)
-
-from NameAndStatus import NameAndStatus
-
-
-class RTTmon(Frame):
-    # Class variable to allow stopping of the thread running auto updates.
-
-
-    def __init__(self, master=None):
-        """Initialises:
-        - top level TCL frame,
-        - server proxy as data source.
-        - logging system
-        Directly calls run() to start RTTmon."""
-
-        
-        Frame.__init__(self, master)
-        
-        self.log = logging.getLogger('rtt')
-        formatter = logging.Formatter('%(asctime)s %(filename)s, %(lineno)s, %(levelname)s %(message)s','%d%b%y %H:%M:%S')
-        hdlr = logging.FileHandler('./RTTmon.log')
-        hdlr.setFormatter(formatter)
-        # self.log.addHandler(hdlr) 
-        
-        hdlr = logging.StreamHandler()
-        hdlr.setFormatter(formatter)
-        hdlr.setLevel(logging.DEBUG)
-        self.log.addHandler(hdlr)
-
-        self.log.setLevel(logging.DEBUG)
-        
-
-        self.log.info('Run RTTmon with pythonw if on a Mac ')
-        self.colors={'queued': 'blue',
-                    'running': 'yellow',
-                    'success': 'green',
-                    'error':   'red'
-                    }
-        self.labelDict={}
-        self.pid=0
-        self.procString='python2.3 run.py'
-        self.findRTT()
-        self.server = ServerProxy("http://localhost:8000") # local server
-        self.autoRun = False
-        self.run()
-    
-
-    def more(self):
-        "Obtain a single dose of fresh data"
-
-        try:
-            info         = self.getJobs()
-        except Exception, e:
-            self.log.info('Could not retieve info from RTT: is RTT running?')
-            return
-        jobs         = info.jobs
-        self.msg     = info.msg
-
-        # self.makeJobFrames(jobs)
-
-        # sometimes RTTmon is running before all jobs were launched.
-        # allow for these to be picked up at a later pass.
-        newJobs = []
-        for job in jobs:
-             name = job.name
-             if name not in self.labelDict.keys():
-                 msg = 'unknown job, '+str(name)+str(self.labelDict.keys())
-                 self.log.info(msg)
-                 newJobs.append(job)
-             else:
-                 
-                 label=self.labelDict[name]
-                 label['bg']=self.colors[job.status]
-
-        self.frames.append(self.makeJobFrames(newJobs))
-        self.pack(expand=YES, fill=BOTH)
-        
-
-        msgs = info.msg.split('::')
-
-        for label, text in zip(self.testerStats, msgs): label['text']=text
-             
-        self.update()
-        self.log.debug('more(), end')
-    def autoSampleOn(self):
-        """Sets global flag autoRun to run auto sampling(ie endless loop
-        that updates the diplay at regular intervals. Instantatiates
-        AutoSample to access threading library, and  starts the loop.
-        Exits only if some other thread resets autoRun"""
-        
-        lock = threading.Lock()
-        lock.acquire()
-        self.autoRun=True
-        lock.release()
-
-        self.log.debug('switching autoRun to '+str(self.autoRun))
-
-        
-    def autoSampleOff(self):
-        "callback function - switches off  auto updating flag"
-        lock =threading.Lock()
-        lock.acquire()
-        self.autoRun=False
-        lock.release()
-
-        self.log.debug('switching autoRun to '+str(self.autoRun))
-
-    def quit(self):
-        "callback function: shats down the RTT"
-        
-        Frame.quit(self)
-        logging.shutdown()
-        
-    def topLine(self):
-        "Sets up top line of the display: control buttons and color map"
-        
-        frame = Frame(self)
-        Button(frame, text='quit',   command=self.quit).pack(side='right')
-        Button(frame, text='update', command=self.more).pack(side='right')
-        
-        self.var = IntVar()
-        localAutoRun = self.autoRun
-        self.var.set(int(localAutoRun))
-        Radiobutton(frame,
-                    text='auto on',
-                    command=self.autoSampleOn,
-                    variable=self.var,
-                    value=1).pack(side='right')
-        
-        cAutoRunBar = str(int(not localAutoRun))
-        Radiobutton(frame,
-                    text='auto off',
-                    command=self.autoSampleOff,
-                    variable=self.var,
-                    value=0).pack(side='right')
-        
-        for state in self.colors.keys():
-            Label(frame, text=state, fg='black', bg=self.colors[state]).pack(side='left')
-        frame.pack(side='top', fill='x')
-        
-    def secondLine(self):
-        "Sets up top line of the display: control buttons and color map"
-
-        self.testerStats = []
-        lines = self.msg.split('::')
-        for line in lines:
-            frame = Frame(self)
-            frame.pack(fill='x')
-            l = Label(frame, text=line, fg='white',bg='black')
-            l.pack(side='left', fill='x')
-            self.testerStats.append(l)
-        # spacer for esthetics
-        frame = Frame(self)
-        frame.pack(fill='x')
-        l= Label(frame, text='', fg='white',bg='black')
-        l.pack(anchor='w',fill='x')
-        self.testerStats.append(l)
-    
-    def makeJobFrames(self, jobs):
-        """Makes, packs and returns the frames for each line. The frames
-        are returned so the can be reset during updates"""
-        
-        lineNo        =  0 
-        labelCount    =  0
-        labelsPerLine = 30
-        lines = {}
-        # split jobs into groups labelsPerLine long
-        for job in jobs:
-            
-            if labelCount != labelsPerLine:
-                lines.setdefault(lineNo,[]).append(job)
-                labelCount +=1
-            else:
-                lineNo     += 1
-                labelCount  = 0
-
-        # a function used to create labels, and pack into a frame
-        # and into a dictionary for later use (updating)
-        def layout(frame, jobs, dict):
-            frame = Frame(frame, bg='black')
-            for job in jobs:
-
-
-                l = Label(frame)
-                l["text"]=str(job.name)
-                l["bg"]=self.colors[job.status]
-                l["fg"]='black'
-                l.pack(side='left', padx=1, pady=1)
-
-                # store in dictionary to ease updating
-                dict[job.name] = l
-                frame.pack(fill='x', expand=YES)
-            return frame
-            
-        return [layout(self, lines[line], self.labelDict) for line in lines.keys()]
-            
-
-    def createWidgets(self, info):
-        """Calls topLine to the display top line. Display the job statuses
-        as color coded labels"""
-
-        jobs     = info.jobs
-        self.msg = info.msg
-        self.topLine()
-        self.secondLine()
-        self.frames = []
-        self.frames.append(self.makeJobFrames(jobs))
-        # make and pack a frame per line 
-
-        # mod   = 0
-        # modcount = 0
-        # for job in jobs:
-        #     
-        #     
-        #     if mod == 3:
-        #         mod = 0
-        #         modcount+=1
-        #         cFrame.pack(fill='x')
-        #         
-        #     if mod == 0:
-        #         self.frames.append(Frame(self,bg='black'))
-        #         cFrame=self.frames[modcount]
-        #         
-        #         
-        #     mod=mod+1
-        #     stat=job.status.primary()
-        #     l = Label(cFrame)
-        #     l["text"]=job.name
-        #     l["bg"]=self.colors[stat]
-        #     l["fg"]='black'
-        #     l.pack(side='left',padx=3, pady=3, fill='x')
-
-        # store in dictionary to ease updating
-        #     self.labelDict[job.name] = l
-        # self.frames[modcount].pack(fill='x',expand=YES)
-
-    def getJobs(self):
-        "Ask the server for a list of NameAndStatus instances."
-
-        self.log.debug( 'asking for jobs..')
-        info = pickle.loads(self.server.sendJobs())
-
-        self.log.debug('Dumping jobs:')
-        self.log.debug(str(reduce(lambda x, y:x+y,
-                                  [job.dumps() for job in info.jobs],
-                                  ' ')+'\n'))
-        return info
-
-    def findRTT(self):
-        """Find the pid number of the RTT - guess what the procs line
-        will look like"""
-        
-        self.log.debug( 'findRTT....')
-        pidFile = 'RTT.pid'
-        
-        if os.path.exists(pidFile):
-            ifile = open(pidFile,'r')
-            self.pid = int(ifile.read())
-            ifile.close()
-            stdout, stdin = popen2.popen2('ps -A')
-            procs = stdout.readlines()
-            def procId(p):
-                return p.split()[0]
-            
-            pids = [procId for p in procs if p.find(self.procString) != -1]
-            if self.pid not in pids:
-                self.log.error('RTT does not appear to be running')
-                self.log.error('perhaps pid = '+str(self.pid)+' is stale?')
-                self.pid = 0
-        else:
-            self.pid = 0
-            self.log.info('could not find the file with the RTT pid '+pidFile)
-
-        # while not  self.pid:
-        #    procsCmd = 'ps -af | grep run.py'
-        #    self.log.debug('procs cmd = '+procsCmd)
-        #    stdout, stdin = popen2.popen2(procsCmd)
-        #    procs = stdout.readlines()
-
-        #     def procId(p):
-        #        return p.split()[0]
-        #    
-        #    pids = [procId for p in procs if p.find(self.procString) != -1]
-        #    pids.sort()
-        #    self.log.debug('pids: '+str(pids))
-        #    
-        #    if len(pids) > 1:
-        #        self.pid = pids[0]
-        #        break
-        #    else:
-        #        self.log.debug( 'waiting for '+str(self.procString)+' process....')
-        #        time.sleep(2)
-
-    def autoUpdateLoop(self):
-        """Checks the auto update flag.
-        If True:asks server for new status info. Schedules itself
-        to run later."""
-
-        self.log.debug('autoRun = '+str(self.autoRun))
-        
-        self.log.debug('entering autosampling')
-        if self.autoRun:
-            self.more()
-            self.log.debug('autoSampled')
-        else:
-            self.log.debug('not auto sampling')
-
-        self.after(30000, self.autoUpdateLoop)
-
-    def run(self):
-        "run the RTT, ending up in the infinite mainloop"
-        
-        self.log.debug('--------------client ---------')
-        while 1:
-            try:
-                info = self.getJobs()
-                break
-            except Exception, v:
-                self.log.info( 'could not connect: is the server up?')
-                self.log.info( str(v))
-                time.sleep(2)
-
-        self.pack(expand=YES, fill=BOTH)
-        self.createWidgets(info)
-        #self.master.maxsize(1000,1000)
-        self.master.title("RTTmon")
-        # as = RTTmon.AutoSample(self)
-        # as.start()
-        self.autoUpdateLoop()
-        self.mainloop()
-
-
-if __name__ == '__main__':
-    print 'calling displayer'
-    profile.run('mon=RTTmon()')
diff --git a/Tools/RunTimeTester/src/RTTpath.py b/Tools/RunTimeTester/src/RTTpath.py
deleted file mode 100755
index af6771f5415be1eeade93f4a5378d4bcc3cd3a91..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RTTpath.py
+++ /dev/null
@@ -1,95 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-
-class RTTpath:
-  """class that holds various path fragments, and combines them to form
-   the most complete path possible"""
-  
-  def __init__(self, *args):
-    self.segments = []
-      
-    # sometimes an RTTpath is made with None in the arguments.
-    # this means 'no such string'. Remove these Nones
-    self.segments.extend([str(a) for a in args if a != None])
-    
-  def prepend(self, *args):
-    segments = [str(s) for s in args]
-    hold = self.segments
-    self.segments.extend(hold)
-  def postpend(self, *args):
-
-    segments = [str(s) for s in args]
-    self.segments.extend(segments)
-  def path(self):
-    "combine the segments to form a path"
-    if not len(self.segments):
-      return '' 
-    path = reduce( os.path.join, self.segments)
-    path = os.path.normpath(path)
-    return path
-  def isabs(self):
-    if not len(self.segments): return False
-    return os.path.isabs(self.path())
-  def exists(self):
-    if not len(self.segments): return False
-    return os.path.exists(self.path())
-  def __str__(self):
-    if not len(self.segments): return ''
-    return self.path()
-
-class CheckedPath(RTTpath):
-  def __init__(self, segments = []):
-    RTTpath.__init__(self, segments)
-
-  def __str__(self):
-    if not self.exists():
-      return ''
-    return RTTpath.__str__(self)
-    
-if __name__ == '__main__':
-  path = RTTpath('.', 'RTTpath.py')
-
-  test = 0
-  
-  test = 1
-  print 'Test %d' % test
-  print 'path',path
-  print 'isabs', path.isabs()
-  print 'exists', path.exists()
-  
-  here = os.getcwd()
-  print here
-  path.prepend(here)
-  
-  
-  test = 2
-  print 'Test %d' % test
-  print 'path',path
-  print 'isabs', path.isabs()
-  print 'exists', path.exists()
-  
-
-  path = RTTpath(here)
-  path.postpend('RTTpath.py')
-  
-  
-  test = 3
-  print 'Test %d' % test
-  print 'path',path
-  print 'isabs', path.isabs()
-  print 'exists', path.exists()
-  
-
-  path1 = RTTpath(here)
-  path2 = RTTpath('Path.py')
-  path2.prepend(path1)
-
-  
-  
-  test = 4
-  print 'Test %d' % test
-  print 'path',path
-  print 'isabs', path.isabs()
-  print 'exists', path.exists()
-  
diff --git a/Tools/RunTimeTester/src/RetryTester.py b/Tools/RunTimeTester/src/RetryTester.py
deleted file mode 100755
index 64f6971006575a80ab0968dd4f637a5707705498..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RetryTester.py
+++ /dev/null
@@ -1,206 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-RetryTester - see if a failed job warants a retry - an attempt to
-overcome networl connection problems
-"""
-import os
-import re
-from RTTpath                 import RTTpath
-
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-
-
-def match(patterns, lines):
-    line = None
-    for p in patterns:
-        cp = re.compile(p)
-        line = None
-
-        for l in lines:
-            if re.search(cp,l):
-                line = l
-                break
-                
-        if line: break
-
-    return line
-
-def allMatches(patterns, lines):
-    matchedLines = []
-    for p in patterns:
-        cp = re.compile(p)
-        matchedLines.extend([l for l in lines if re.search(cp,l)])
-    return matchedLines
-
-def getFromEnvLog(runPath,envVar,pathSep=':'):
-    envLogName = os.path.join(runPath, 'env.log')
-    if os.path.exists(envLogName):
-        ifile = open(envLogName, 'r')
-        content = ifile.readlines()
-        ifile.close()
-    else:
-        return []
-
-    envVar = envVar.strip() + '='
-    libstr = ''
-    for line in content:
-        if line.startswith(envVar):
-            libstr = line[len(envVar):]
-            break
-    if not libstr: return []
-        
-    return libstr.split(pathSep)
-
-class RetryTester:
-    #def __init__(self, log, elog, logger, sharePath, runPath):
-    def __init__(self, log, logger, sharePath, runPath):
-        self.logName   = log # full abs path to log file (which doesn't exist yet)
-        # self.elogName  = elog
-        self.log       = []
-        # self.elog      = []
-        self.logger    = logger
-        self.sharePath = sharePath
-        self.runPath   = runPath
-        
-    def shouldRetry(self):
-        if not self.log:            
-            if os.path.exists(self.logName):
-                ifile = open(self.logName,'r')
-                self.log = ifile.readlines()
-                ifile.close()
-
-        # if not self.elog:
-        #  if os.path.exists(self.elogName):
-        #     self.elog = open(self.elogName,'r').readlines()
-        
-        if self.commandNotFound():        return True
-        if self.missingJobOptionsExist(): return True
-        if self.missingSharedLibrary():   return True
-        
-        return False
-            
-
-    # -------------------------------------------------------------------------
-    # -------------------------------------------------------------------------
-    
-    def missingJobOptionsExist(self):
-        """
-        Looks for missing job options files in the log files. If Athena
-        complains of such a file, check to see if it really does not exist
-        """
-
-        patterns = ['^AthenaCommon.Include.IncludeErr',
-                    '^IOError: [Errno 2] No such file or directory:',
-                    '^IncludeError: include file']
-        # line = match(patterns, self.elog)
-        line = match(patterns, self.log)
-        if not line: return False
-        
-        tokens = line.split()
-        jos = match(['.py$'], tokens)
-
-        if not jos: return False
-        
-        # Note that this makes the assumption (for now, always borne out)
-        # that the 'missing' job options file is in the same project as the
-        # one trying to include it, and that these are both in the same project
-        # as the package under test.
-
-        if os.path.isabs(jos) and not os.path.exists(jos): return False
-        
-        joSearchPaths = getFromEnvLog(self.runPath,'JOBOPTSEARCHPATH',',')
-        if not joSearchPaths:
-            self.logger.warning('No search paths available in which to search for missing job options!')
-            self.logger.warning('(This is either because RTT could not find env.log file, or JOBOPTSEARCHPATH missing/empty in that file.)')
-            return False                                 
-
-        for p in joSearchPaths:
-            p = p.strip()
-            fullPath = RTTpath(p,jos)
-            if os.path.exists(str(fullPath)):
-                m  = 'Apparent connection error: Found the following job options with'
-                m += ' Athena include but which exist: %s' % str(jos)
-                self.logger.info(m)        
-                return True
-        return False
-
-
-    # -------------------------------------------------------------------------
-
-    def commandNotFound(self):
-        """
-        Checks for 'not found' commands such as athena, cmt, etc.
-        """
-
-        patterns = [
-            '^#CMT> Warning: package CMT * /afs/cern.ch/sw/contrib not found',
-            'athena.py: command not found',
-            'cmt: command not found',
-            'pool_insertFileToCatalog: command not found'
-            ]
-        
-        # lineFound = match(patterns, self.elog)
-        lineFound = match(patterns, self.log)
-        if lineFound:
-            m = 'Apparent connection error:\n%s' % lineFound
-            self.logger.info(m)
-            return True
-        return False
-
-    # -------------------------------------------------------------------------
-    
-    def missingSharedLibrary(self):
-        """
-        Looks for missing shared library files in the log files. If Athena
-        complains of such a file, check to see if it really does not exist
-        """
-        patterns = ['^System::doLoad: ERROR:']
-        # lines = allMatches(patterns, self.elog)
-        lines = allMatches(patterns, self.log)
-
-        if not lines: return False
-
-        soFiles = []
-        for l in lines:
-            #example line:
-            #System::doLoad: ERROR:libMuonAlgs.so: cannot open shared object file
-            #: No such file or directory'
-            #
-            tokens = l.split()
-            soFiles.extend([t[:-1] for t in tokens if t.endswith('.so:')])
-
-        # print soFiles
-        if not soFiles: return False
-
-        # make a list of libary paths
-        libDirs = getFromEnvLog(self.runPath,'LD_LIBRARY_PATH')
-        if not libDirs:
-            self.logger.warning('No lib paths in which to search for missing shared library!')
-            self.logger.warning('(This is either because RTT could not find env.log file, or LD_LIBRARY_PATH missing/empty in that file.)')
-            return False
-        
-        fullLibPaths = []
-        for lib in soFiles:
-            possiblePaths  = [os.path.join(libDir, lib) for libDir in libDirs]
-            foundLocations = [l for l in possiblePaths if os.path.exists(l)]
-
-            # looked everywhere on LD_LIBRARY_PAH, but did not find the
-            # library. It is genuinely not present so do not retry
-            if foundLocations: 
-
-                m  = 'Apparent afs connection error: '
-                m += 'Found the following so files reported'
-                m += ' missing but in fact present %s' % str(foundLocations)
-                self.logger.info(m)
-
-        
-                return True
-
-        return False
-
-
-
diff --git a/Tools/RunTimeTester/src/RootMacroRunner.py.obsolete b/Tools/RunTimeTester/src/RootMacroRunner.py.obsolete
deleted file mode 100755
index 6d94cd36ac463342182f72612f25cf0acb8dac76..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RootMacroRunner.py.obsolete
+++ /dev/null
@@ -1,446 +0,0 @@
-"""
-Functions to run after a jobs have completed. Organised by job group
-"""
-
-import os, time, re
-from popen2 import Popen4
-
-# from siteSpecificsFactory import siteSpecificsFactory
-from Factory_SiteSpecifics    import SiteSpecificsFactory
-from TextExtractor import TextExtractor
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class SitePostScripts:
-	
-	def setSiteSpecifics(self, site):
-		logger.debug('setting site specific env on :' + site )
-		
-	        # siteSpec = siteSpecificsFactory(site)
-	        siteSpec = SiteSpecificsFactory().create(site)
-		os.environ['LD_LIBRARY_PATH'] = siteSpec.libs()
-		os.environ['ROOTSYS']         = siteSpec.rootSys()
-	        #os.chdir(self.runPath)
-                #script(self.jDescriptor.paths.site)
-		
-	setSiteSpecifics=classmethod(setSiteSpecifics)
-
-
-class ArtemisPostScript(SitePostScripts):
-
-	psLabel = 'site'
-	
-	def run(self, site):
-		
-		"run some root macros"
-		logger.debug('running postscript  on ' +site)
-		
-		SitePostScripts.setSiteSpecifics(site)
-
-		# rootCommand = siteSpecificsFactory(site).rootCommand()
-	        rootCommand = SiteSpecificsFactory().create(site).rootCommand()
-		rootCommand = rootCommand+' artemis.C'
-		os.system(rootCommand)
-
-	run = classmethod(run)
-
-class AtlfastPostScript(SitePostScripts):
-
-	psLabel = 'site'
-	
-	def run(self, site):
-		
-		"run some root macros"
-	       
-		logger.debug('running postscript  on ' +site)
-		logger.info('Current working directory: '+ os.getcwd())
-		SitePostScripts.setSiteSpecifics(site)
-	
-		# rootCommand = siteSpecificsFactory(site).rootCommand()
-	        rootCommand = SiteSpecificsFactory().create(site).rootCommand()
-		rootCommand = rootCommand+' atlfast.C'
-
-		logger.info( "atlfastPostScript root command: "+rootCommand)
-
-		#os.system(rootCommand)
-		runMacro = Popen4(rootCommand)
-		logger.info('root command output: \n'+ runMacro.fromchild.read())
-
-	run = classmethod(run)
-
-class EgammaPostScript(SitePostScripts):
-
-	psLabel = 'site'
-	
-	def run(self,site):
-
-		"run some root macros"
-		
-		
-		logger.debug('running postscript  on ' +site)
-		logger.debug('Current working directory: '+os.getcwd())
-
-		SitePostScripts.setSiteSpecifics(site)
-	
-		# rootCommand = siteSpecificsFactory(site).rootCommand()
-	        rootCommand = SiteSpecificsFactory().create(site).rootCommand()
-	
-		macroCommand = rootCommand+' egamma_hist.C\(\\"ntuple.root\\"\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output: \n'+ runMacro.fromchild.read() )
-		runMacro.wait()
-		
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"cluster_validation.hist\\",'
-		macroCommand += '\\"refFile_cluster_validation.hist\\",'
-		macroCommand += '\\"cluster_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output: \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-	
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"softE_validation.hist\\",'
-		macroCommand += '\\"refFile_softE_validation.hist\\",'
-		macroCommand += '\\"softE_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output:\n'+runMacro.fromchild.read())
-		runMacro.wait()
-		
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"eGamma_validation.hist\\",'
-		macroCommand += '\\"refFile_eGamma_validation.hist\\",'
-		macroCommand += '\\"eGamma_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-		
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-
-
-	run = classmethod(run)
-
-class EgammaWatcherPostScript(SitePostScripts):
-
-	psLabel = 'site'
-
-	def run(self,site):
-		"run some root macros"
-	
-		logger.debug('running postscript  on ' +site)
-
-		logger.debug('Current Work Dir: ' +os.getcwd())  #watcher run directory
-
-		SitePostScripts.setSiteSpecifics(site)
-		
-		# rootCommand = siteSpecificsFactory(site).rootCommand()
-	        rootCommand = SiteSpecificsFactory().create(site).rootCommand()
-
-	
-	        macroCommand = rootCommand+' histAdder.C\(\\"cluster_validation.hist\\"\)'
-	       	logger.info("egammaWatcherPostScript root command: "+macroCommand)
-	       	#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-		macroCommand = rootCommand+' histAdder.C\(\\"softE_validation.hist\\"\)'
-		logger.info("egammaWatcherPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-		macroCommand = rootCommand+' histAdder.C\(\\"eGamma_validation.hist\\"\)'
-		logger.info("egammaWatcherPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-	
-	
-	
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"cluster_validation.hist\\",'
-		macroCommand += '\\"refFile_cluster_validation.hist\\",'
-		macroCommand += '\\"cluster_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-	
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"softE_validation.hist\\",'
-		macroCommand += '\\"refFile_softE_validation.hist\\",'
-		macroCommand += '\\"softE_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-		macroCommand  = rootCommand
-		macroCommand += ' egammaHistsToPostScript.C\('
-		macroCommand += '\\"eGamma_validation.hist\\",'
-		macroCommand += '\\"refFile_eGamma_validation.hist\\",'
-		macroCommand += '\\"eGamma_validation.ps\\"'
-		macroCommand += '\)'
-		logger.info("egammaPostScript root command: "+macroCommand)
-		#os.system(macroCommand)
-
-		runMacro =Popen4(macroCommand)
-		logger.info('root command output \n'+runMacro.fromchild.read())
-		runMacro.wait()
-	
-	
-	run =classmethod(run)
-
-class HelloWorldPostScript:
-
-	psLabel = 'site'
-	
-	def run(self, site):
-		"Nothing needs to be done"
-
-		logger.debug('running postscript  on ' +site)
-
-	run = classmethod(run)
-
-class JiveXMLPostScript:
-
-	psLabel = 'log'
-	
-	def run(self, logFile):
-		# logFile is the absolute path the the logFile in the run directory
-		"Extract lines from log and store in logExtract"
-
-                logger.info( "JiveXML PostScript")
-		logger.debug('running  postscript  on ' +logFile)
-
-		head, tail = os.path.split(logFile)
-		outFile = os.path.join(head,'logExtract.txt')
-		testStrings=['EventData2XML        INFO ',
-                     'EventData2XML     WARNING ']
-		te = TextExtractor(testStrings, logFile, outFile)
-		te.extract()
-
-	run = classmethod(run) 
-
-
-class RecExCommonPostScript:
-	
-	psLabel = 'log'
-	
-	def run(self, logFile):
-		# logFile is the absolute path the the logFile in the run directory
-		"Extract lines from log and store in logExtract"
-
-                logger.info( "RecExCommon PostScript")
-		logger.debug('running  postscript  on ' +logFile)
-
-		head, tail = os.path.split(logFile)
-		outFile = os.path.join(head,'logExtract.txt')
-
-		testStrings=["\.\.\.only second half of the job",
-			     "CBNT_Audit                       INFO End of initialization : Total Virtual memory",
-			     "CBNT_Audit                       INFO After finalise:  virtual memory VSIZE",
-			     "CBNT_Audit                       INFO Total Virtual memory",
-			     "ChronoStatSvc                    INFO Time User   : Tot"]
-		te = TextExtractor(testStrings, logFile, outFile)
-		te.extract()
-
-		outFile = os.path.join(head,'memoryLeak.txt')
-
-		testStrings=["\.\.\.only second half of the job"]
-		te = TextExtractor(testStrings, logFile, outFile)
-		te.extract()
-					
-	run = classmethod(run) 
-
-		
-class QaMetricsToHTML(SitePostScripts):
-
-	psLabel ='site'
-	
-	def run(self,site=None):
-		"""Script to convert metrics txt file to html.
-		No longer neededd: Together has an html format option
-		that works better
-		"""		
-
-		logger.debug('running postscript  on ' +site)
-
-		SitePostScripts.setSiteSpecifics(site)
-
-		import sys
-	
-	
-		lines = open('metrics.txt','r').readlines()
-	
-		if len(lines)==0: sys.exit(0)
-	
-	
-		headers=lines[0].split()
-		headerLine='      '
-		for header in headers:
-			headerLine=headerLine+'<th>'+header+'</th>'
-			headerLine = headerLine+'\n'
-
-			oline = '      <tr>'
-			for line in lines[1:]:
-				entries=line.split()
-				for entry in entries:
-					oline=oline+'<td>'+entry+'</td>'
-					oline = oline+'\n'
-					oline = oline+'</tr>'
-    
-					ofile = open('metrics.html','w')
-					ofile.write('<html>\n')
-					ofile.write('  <body>\n')
-					ofile.write('    <h2>QA metrics</h2>\n')
-					ofile.write('    <table>\n')
-
-					ofile.write('    <tr>\n')
-					ofile.write(headerLine)
-					ofile.write('    </tr>\n')
-					ofile.write('    <tr>\n')
-					ofile.write(oline)
-    
-					ofile.write('    </tr>\n')
-
-
-
-					ofile.write('    </table>\n')
-					ofile.write('  </body>\n')
-					ofile.write('</html>\n')
-    
-					ofile.close()
-			run = classmethod(run)
-
-class TrigReleasePostScript:
-
-	psLabel = 'log'
-	
-	def run(self, logFile):
-		'logFile is the absolute path the the logFile in the run directory'
-
-		logger.debug("Running TriggerRelease PostScripts")
-
-		head, tail  = os.path.split(logFile)
-		outFile1    = os.path.join(head, 'errorLog.txt')
-		testString1 = ['FATAL', 'ERROR']         
-		te         = TextExtractor(testString1, logFile, outFile1)
-		te.vetoExtract()                #error log
-
-	run = classmethod(run)
-
-class TrigReleasePostScript1:
-
-	psLabel = 'log'
-	
-	def run(self, logFile):
-		"Create log extracts needed for regression tests"
-
-		# logFile is the absolute path the the logFile in the run directory
-		logger.debug("Running TriggerRelease PostScripts")
-
-		head, tail = os.path.split(logFile)
-		outFile1   = os.path.join(head, 'idscan.reference')
-		outFile2   = os.path.join(head, 'sitrack.reference')
-		 
-		testString1 = ["^IdScan.*.REGTEST"]           #regular expressions
-		testString2 = ["^SiTrack.*.REGTEST"]          #determining lines to 
-                                                      
-		te1 = TextExtractor(testString1, logFile, outFile1)
-		te2 = TextExtractor(testString2, logFile, outFile2)
-
-	        #Files for regression tests
-		te1.extract()                 
-		te2.extract()
-
-	run = classmethod(run)
-
-class TrigReleasePostScript2:
-
-	psLabel = 'log'
-	
-	def run(self, logFile):
-		"Create log extracts needed for regression tests"
-
-		logger.debug("Running TriggerRelease PostScripts")
-
-		head, tail = os.path.split(logFile)
-		# logFile is the absolute path the the logFile in the run directory
-		outFile1   = os.path.join(head, 't2calo.reference')
-		#outFile2   = os.path.join(head, 't2calo_readBS.reference')
-		 
-		testString = ["^T2Calo.*.REGTEST"]            #regular expressions
-		                                              #determining lines to 
-                                                      
-		te1 = TextExtractor(testString, logFile, outFile1)
-		#te2 = TextExtractor(testString, logFile, outFile2)
-
-	        #Files for regression tests
-		te1.extract()                 
-		#te2.extract()
-
-	run = classmethod(run)
-
-
-class TrigReleasePostScript3:
-	
-        psLabel = 'log'
-	
-	def run(self, logFile):
-		"Create log extracts needed for regression tests"
-		
-		logger.debug("Running TriggerRelease PostScripts")
-
-		head, tail = os.path.split(logFile)
-		# logFile is the absolute path the the logFile in the run directory
-		outFile1   = os.path.join(head, 'idscan_readBS.reference')
-		outFile2   = os.path.join(head, 'sitrack_readBS.reference')
-		 
-		testString1 = ["^IdScan.*.REGTEST"]           #regular expressions
-		testString2 = ["^SiTrack.*.REGTEST"]          #determining lines to 
-                                                      
-		te1 = TextExtractor(testString1, logFile, outFile1)
-		te2 = TextExtractor(testString2, logFile, outFile2)
-
-	        #Files for regression tests
-		te1.extract()                 
-		te2.extract()
-
-	run = classmethod(run)
-		
-
-if __name__ == "__main__":
-	pass
diff --git a/Tools/RunTimeTester/src/RunTimer.py b/Tools/RunTimeTester/src/RunTimer.py
deleted file mode 100755
index d92fa63336aa4909d2385e94757a7166768539b5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/RunTimer.py
+++ /dev/null
@@ -1,109 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time
-
-class RunTimer:
-    def __init__(self,maxTime, logger):
-        "maxTime in seconds"
-        
-        self.maxProcessTime = maxTime
-        self.startTime      = time.time()
-        self.logger         = logger
-        self.logger.debug("RunTimer constructor: maxtime (secs): %d" % self.maxProcessTime)
-
-    #------------------------------------------------------
-    def __str__(self):
-        stime = time.strftime("%d %b %Y %H:%M:%S ",
-                              time.localtime(self.startTime))
-        s = 'start time: %s maxtime: %s' % (stime,  str(self.maxProcessTime))
-        return s
-
-    #------------------------------------------------------
-    
-    def remaining(self):
-        remain = self.maxProcessTime - (time.time()-self.startTime)
-        if remain<0: remain =0
-        self.logger.debug('RunTimeTimer, remaining time %d' % remain)
-        return remain
-
-    #------------------------------------------------------
-    
-    def timeOut(self):
-        now = time.time()
-        to = False
-        if (now-self.startTime)>self.maxProcessTime: to =  True
-        self.logger.debug('RunTimeTimer, timed-out =  %s' % str(to))
-        return to
-
-    #------------------------------------------------------
-    
-    def wasNotStartedToday(self):
-        """If the stamp has not come yet and it is past midnight, stop the RTT."""
-        startDate = self.extractDate(self.startTime)
-        nowDate = self.extractDate(time.time())
-        answer = False
-        if startDate != nowDate:
-            answer = True
-        return answer
-
-    #------------------------------------------------------
-    
-    def extractDate(self,theTime):
-        theList = time.gmtime(theTime)
-        date = [theList[0], theList[1], theList[2]]
-        return date
-
-    #------------------------------------------------------
-    
-    def reset(self,maxTimeFromNow=60*60*20):
-        'Resets the timer, as if it had been created now.'
-        self.startTime = time.time()
-        self.maxProcessTime = maxTimeFromNow
-        self.logger.info('Reset the RunTimer to: ' + str(maxTimeFromNow) + ' seconds.')
-        
-    #------------------------------------------------------
-    
-    def maxTimeInHours(self):
-        return int(float(self.maxProcessTime)/float(3600))
-
-    #------------------------------------------------------
-    
-    def start(self):
-        stringRT = str(self)
-        toks = stringRT.split()
-        theStart = toks[2] + ' ' + toks[3] + ' ' + toks[4] + ' ' + toks[5]
-        return theStart
-
-#=========================================================
-#=========================================================
-#=========================================================
-
-class NX0NightlyRunTimer(RunTimer):
-    def __init__(self, logger):
-        RunTimer.__init__(self, 165*60*60, logger)
-
-#=========================================================
-#=========================================================
-#=========================================================
-
-class N0XNightlyRunTimer(RunTimer):
-    def __init__(self, logger):
-        RunTimer.__init__(self, 115*60*60, logger)
-
-#=========================================================
-#=========================================================
-#=========================================================
-
-class NumberedRelRunTimer(RunTimer):
-    def __init__(self, logger):
-        RunTimer.__init__(self, 750*60*60, logger)
-
-
-
-#=========================================================
-#=========================================================
-#=========================================================
-
-if __name__ == '__main__':
-    pass
-
diff --git a/Tools/RunTimeTester/src/Runner.py.obsolete b/Tools/RunTimeTester/src/Runner.py.obsolete
deleted file mode 100755
index 871bd153f746a4199fd6b7c1bd56a68f41d86126..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Runner.py.obsolete
+++ /dev/null
@@ -1,16 +0,0 @@
-#! /usr/bin/python
-"""
-Short script that is used to start the RTT.
-Could be run from a cron job.
-User specifies the location of the RTT scripts, and the location of the
-two configuration files.
-"""
-
-from Tester import Tester
-
-if __name__ == "__main__":
-    Tester(
-        "/afs/cern.ch/user/s/simmons/ATLAS/work/RTT_latest/Tools/RunTimeTester/RunTimeTester-00-00-15/src",
-        "/afs/cern.ch/user/s/simmons/ATLAS/work/RTT_latest/Tools/RunTimeTester/RunTimeTester-00-00-15/src/cfg.cfg"
-        )
-
diff --git a/Tools/RunTimeTester/src/ScriptWriter.py b/Tools/RunTimeTester/src/ScriptWriter.py
deleted file mode 100755
index 6fe79bcd2e4763f357bc1c57b467508e4893529f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ScriptWriter.py
+++ /dev/null
@@ -1,597 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-ScriptWriter write is a base class for lineswriting shell scripts to run the jobs.
-The shell scripts are typically written to a run directory, and are submitted to a
-queue from there.
-
-The base class provides a default implementation.
-"""
-
-from Factory_CmtLinesForScripts   import CmtLinesForScriptsFactory
-from os.path                      import join
-from releaseToNum                 import releaseToNum
-from Factory_SiteSpecifics        import SiteSpecificsFactory
-from Tools                        import changePerm
-from exc2string2                  import exc2string2
-
-import CMTtools
-import os, sys, string
-
-def libLineCalculator(siteSpecifics, mode, gcc_version):
-    """
-    This function calculates the LD_LIBRARY_PATH line.
-    This line will be placed in a shell script that will be run using
-    Popen4.
-    
-    The explicit value of LD_LIBRARY_PATH is written out so the file
-    can be run by hand under the same conditions that the p4open object
-    sees.
-    """
-
-    llp = os.environ.get('LD_LIBRARY_PATH','')
-    if siteSpecifics.addToLibs():
-        libLine  = 'LD_LIBRARY_PATH='
-        libLine += siteSpecifics.addToLibs() + siteSpecifics.pathToOracleInstantClient(gcc_version) + ':$LD_LIBRARY_PATH'
-    else:
-        libLine = ''
-
-    return libLine
-    
-def writeToDir(script, logger):    
-
-    ofile = open("runScript.sh","w")
-    ofile = ofile.write(script)
-    ofile.close()
-    
-    changePerm("runScript.sh", 02775)
-    logger.debug("wrote runscript to = "+os.getcwd())
-    logger.debug("current dir: " + os.getcwd())
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class ScriptWriter:
-    
-    def __init__(self, argsForScripts, logger):
-        'Common stuff for all Scriptwriters - use Descriptor base path'
-
-        self.logger         = logger
-        self.logger.debug( 'initialising %s' % self.__class__.__name__)
-        self.RTTSrcDir      = argsForScripts.RTTSrcDir
-        self.RTTLibDir      = argsForScripts.RTTLibDir
-        self.siteSpecifics  = argsForScripts.siteSpecifics
-        self.workPath       = argsForScripts.workDirs['base']
-        self.cmtLines       = argsForScripts.cmtLines
-        self.mode           = argsForScripts.mode
-        self.runPath        = argsForScripts.runPath
-        self.castorSvcClass = '' # might be filled in from argsForScripts
-        
-        self.athenaCommand = getattr(argsForScripts,'athenaCommand', None)
-        
-        labels = ['jobGroup', 'runPath', 'castorSvcClass', 'jobTransformCmd', 'errorMessages']
-        [setattr(self, l, getattr(argsForScripts, l, None)) for l in labels]
-
-        compiler = CMTtools.decodeCMTCONFIG(argsForScripts.cmtConfig)['compiler']
-        self.libLine = libLineCalculator(self.siteSpecifics,
-                                         self.mode,
-                                         compiler)
-
-        self.logger.debug('libLine: %s' % self.libLine)
-        self.pythonPath = string.join(sys.path,':')
-        
-        self.useSQLite = True
-
-    def makeScript(self):
-        raise NotImplementedError
-    
-    def runScript(self):
-        script = self.makeScript()
-        writeToDir(script, self.logger)
-
-    def dumpEnvVars(self):
-        script  = 'echo "-----------Dumping environment variables-------------"\n'
-        script += 'env>env.log\n'
-        script += 'echo "-----------------------------------------------------"\n'
-        return script
-
-    def toJobRunDir(self):
-        return 'cd %s\n' % self.runPath
-
-    def setupLocalSQLiteReplica(self):
-        return 'setupLocalDBReplica_CERN.sh\n'
-
-    def setupCASTOR(self):
-        if self.castorSvcClass:
-            return 'export STAGE_SVCCLASS=%s\n' % self.castorSvcClass
-        return 'export STAGE_SVCCLASS=atlasgroupdisk\n'
-
-    def indicateThisIsRTT(self):
-        return 'export RTT=True\n'
-    
-    def postProcessorScript(self):
-        rpath = self.runPath
-        
-        script = 'echo "----- RTT Post Processing Starts ---------"\n'
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'cd %s\n' % rpath
-        for line in self.cmtLines: script += line+'\n'
-        script += 'which python\n'
-        script += 'PYTHONPATH=%s:%s:$PYTHONPATH\n' % (self.RTTSrcDir, self.RTTLibDir)
-        script += 'cd %s\n' % rpath
-        script += 'python  %s\n' % os.path.join(rpath, 'testRunner.py')
-        script += 'echo "----- RTT Post Processing Ends---------"\n'
-        
-        return script
-    
-        
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class NullScriptWriter(ScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        ScriptWriter.__init__(self, argsForScripts, logger)
-    
-    def runScript(self):
-        pass
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------        
-
-class MonitoringScriptWriter(ScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        ScriptWriter.__init__(self, argsForScripts, logger)
-        
-    def makeScript(self):
-
-         script = '#! /bin/bash\n'
-         script += 'set +e\n',
-         script += 'cd '
-         script += self.runPath+'\n'
-         for line in self.cmtLines: script += line+'\n'
-         script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-         script += 'cd '+self.runPath+'\n'
-         script += self.libLine+'\n'
-         script += self.dumpEnvVars()
-         script += 'echo "one: 1  \n two:2 \n three:3 " > monitoringData.dat\n'
-         script += 'echo "four: 4  \n five:5 \n six:6 " > someData.dat\n'
-         script += 'echo " ">JobCompleteMarker.txt\n'
-
-         return script
-     
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------        
-
-class AthenaScriptWriter(ScriptWriter):
-    
-    def __init__(self, argsForScripts, logger):
-        "Provides a run script for Athena jobs"
-        ScriptWriter.__init__(self,  argsForScripts, logger)
-        
-        self.releasePath            = argsForScripts.releasePath
-        self.packageInstalledCMTdir = argsForScripts.cmtDir
-        self.package                = argsForScripts.package
-
-    def grabDBfile(self):
-        return ''
-
-    def cmtShowUses(self):
-        script  = 'cd '+self.packageInstalledCMTdir+'\n'
-        script += 'cmt show uses > '+join(self.runPath,'cmtShowUses.log')+'\n'
-        return script            
-        
-    def makeScript(self):
-        """
-        Default shell script for running an Athena job.
-        """
-        
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += self.indicateThisIsRTT()
-        script += 'cd '+self.runPath+'\n'
-
-        script += 'export SITEROOT=%s\n' % self.siteSpecifics.site()
-        script += 'echo "Going to launch CMT lines"\n'
-        for line in self.cmtLines: script += line+'\n'
-        script += 'echo "Finished launching CMT lines"\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-        script += 'cd '+self.runPath+'\n'
-        script += self.grabDBfile() + '\n'
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.toJobRunDir()
-        if self.useSQLite: script += self.setupLocalSQLiteReplica()
-        script += self.lsatlF('before')
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'
-           
-        return script
-
-    # ---------------------------------------------------------------------
-    
-    def lsatlF(self,when):
-        fileName = join(self.runPath,'lsaltF.txt')+'\n'
-        message = ''
-        if when=='before':
-            message = 'echo "Issuing command ls -altF on run directory, just BEFORE invoking athena..." > ' + fileName
-        elif when=='after':
-            message = '\n echo "Issuing command ls -altF on run directory, just AFTER invoking athena..." >> ' + fileName
-            
-        message = message + 'echo "------------------------------------------------" >> ' + fileName
-        message = message + 'ls -altF >> ' + fileName
-        return message
-
-    # ---------------------------------------------------------------------
-    
-    def checkExtension(self, option):
-        ext =option[-3:]
-        if ext =='.py':
-            pass
-        else:
-            option=option+'.py'
-        return option
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------        
-
-class HLTScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-        self.runPath = argsForScripts.runPath
-        self.package = argsForScripts.package
-        
-    def makeScript(self):
-        """Default shell script for running a job."""
-        
-        script  = '#! /bin/bash\n'
-        script += 'shopt -s expand_aliases\n'
-        script += 'set +e\n'
-        script += self.indicateThisIsRTT()
-        script += 'cd '+self.runPath+'\n'
-
-        script += 'export SITEROOT=%s\n' % self.siteSpecifics.site()
-        for line in self.cmtLines: script += line+'\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-        script += 'cd '+self.runPath+'\n'
-        script += self.grabDBfile() + '\n'
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.lsatlF('before')
-        script += self.toJobRunDir()
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'
-
-        return script
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class HLTTrigConfScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-    
-    def makeScript(self):
-        """
-        Default shell script for running a job.
-        """
-        
-        script = '#! /bin/bash\n'
-        script += 'shopt -s expand_aliases\n'
-        script += self.indicateThisIsRTT()
-        script += 'set +e\n'
-        script += 'cd '+self.runPath+'\n'
-
-        script += 'export SITEROOT=%s\n' % self.siteSpecifics.site()
-        for line in self.cmtLines: script += line+'\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % os.path.join(self.runPath, self.package)
-        script += 'cd '+self.runPath+'\n'
-        script += self.grabDBfile() + '\n'
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.lsatlF('before')
-        script += self.toJobRunDir()
-        # script += 'LD_LIBRARY_PATH=/afs/cern.ch/atlas/maxidisk/d90/13.1.0/InstallArea/i686-slc4-gcc34-dbg/lib:$LD_LIBRARY_PATH\n'
-        # script += 'GAUDISVCROOT=/afs/cern.ch/atlas/maxidisk/d90/13.1.0/GaudiSvc\n'
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'
-
-        return script
-    
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class JunkScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-        self.jobOptions = argsForScripts.pop('jobOptions', 'jobOptionsUnspecified')
-        
-    def makeScript(self):
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'cd '+ self.runPath + '\n'
-        script += 'echo AthenaCommon.Include.IncludeError ' + self.jobOptions + '\n'
-
-        return script
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class AthenaChildScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-        
-    def scriptToSource(self):
-        pass
-
-    def makeScript(self):
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += self.indicateThisIsRTT()
-        script += 'cd '+self.runPath+'\n'
-        script += 'export SITEROOT="CERN"\n' #should be CERN - used by CMT
-        for line in self.cmtLines: script += line+'\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-        script += 'cd '+self.runPath+'\n'
-        script += self.scriptToSource()
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.toJobRunDir()
-
-        if self.useSQLite: script += self.setupLocalSQLiteReplica()
-        script += self.lsatlF('before')
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'
-           
-        return script
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class RecExCommonScriptWriter(AthenaChildScriptWriter):
-    """
-    Script to run RecExCommon.
-    This needs to perform various RecExCommon tasks.
-    """
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-
-    def scriptToSource(self):
-        return 'source RecExCommon_links.sh\n'
-    
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class RecExCommissionScriptWriter(AthenaChildScriptWriter):
-    """Script to run RecExCommission."""
-    
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-        self.useSQLite = False
-
-    def scriptToSource(self):
-        return 'source RecExCommissionData_links.sh\n'
-
-    # ---------------------------------------------------------------------
-
-class CaloRecExAANTRealScriptWriter(AthenaChildScriptWriter):
-    """
-    Script to run RecExCommon.
-    This needs to perform various RecExCommon tasks.
-    """
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-        self.useSQLite = False
-        
-    def scriptToSource(self):
-        return '\n' # i.e. none
-    
-    # ---------------------------------------------------------------------
-    
-class RecExCommissionSimScriptWriter(AthenaChildScriptWriter):
-    """Script to run RecExCommissionSim"""
-    
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-        self.useSQLite = False
-
-    def scriptToSource(self):
-        return 'source RecExCommission_links.sh\n'
-
-    # ---------------------------------------------------------------------
-
-class TileRecExScriptWriter(AthenaChildScriptWriter):
-    """Script to run RecExCommission."""
-    
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-
-    def scriptToSource(self):
-        return 'source TileRecEx_links.sh\n'
-
-    # ---------------------------------------------------------------------
-
-class TileSimExScriptWriter(AthenaChildScriptWriter):
-    """Script to run TileSimEx."""
-    
-    def __init__(self, argsForScripts, logger):
-        AthenaChildScriptWriter.__init__(self, argsForScripts, logger)
-
-    def scriptToSource(self):
-        return 'source TileSimEx_links.sh\n'
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class JiveXMLScriptWriter(RecExCommonScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        RecExCommonScriptWriter.__init__(self, argsForScripts, logger)
-
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class TrigReleaseIDScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-
-    def makeScript(self):
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'cd '+self.runPath+'\n'
-        script += self.indicateThisIsRTT()                 
-        script += 'export SITEROOT="CERN"\n'
-        for line in self.cmtLines: script += line+'\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-        script += 'cd '+self.runPath+'\n'
-        script += self.grabDBfile() + '\n'
-        script += 'source TriggerRelease_links.sh\n'
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.toJobRunDir()
-        script += self.setupLocalSQLiteReplica()
-        script += self.lsatlF('before')
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'      
-        
-        return script
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class TriggerReleaseScriptWriter(AthenaScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        AthenaScriptWriter.__init__(self, argsForScripts, logger)
-
-    def makeScript(self):
-         
-        script = '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += 'export SITEROOT="CERN"\n'
-        script += self.indicateThisIsRTT()                 
-        script += 'cd '+self.runPath+'\n'
-        for line in self.cmtLines: script += line+'\n'
-        script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-        script += 'cd '+self.runPath+'\n'
-        script += self.grabDBfile() + '\n'
-        #script += 'source TriggerRelease_links.sh\n'
-        script += 'source RecExCommon_links.sh\n'
-        script += self.libLine+'\n'
-        script += self.siteSpecifics.preAthena()+'\n'
-        script += self.cmtShowUses() + '\n'
-        script += self.toJobRunDir()
-        script += self.setupLocalSQLiteReplica()
-        script += self.lsatlF('before')
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += self.athenaCommand+'\n'
-        script += self.lsatlF('after')
-        script += 'echo " ">JobCompleteMarker.txt\n'
-        
-        return script
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class TrigReleaseCaloScriptWriter(TrigReleaseIDScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        TrigReleaseIDScriptWriter.__init__(self, argsForScripts, logger)
-
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-
-class JobTransformScriptWriter(ScriptWriter):
-    
-    def __init__(self, argsForScripts, logger):
-        "Provides a run script for Athena jobs"
-        ScriptWriter.__init__(self, argsForScripts, logger)
-
-        self.releasePath            = argsForScripts.releasePath
-        self.packageInstalledCMTdir = argsForScripts.cmtDir
-        self.mode                   = argsForScripts.mode
-        
-    def makeScript(self):
-        """
-        Shell script for running job trnsform python script.
-        """
-                
-        script  = 'echo "----- JobTransform shell script starts  ---------"\n'
-        script += '#! /bin/bash\n'
-        script += 'set +e\n'
-        script += self.indicateThisIsRTT()                 
-        script += 'cd %s\n' % self.runPath
-        for line in self.cmtLines: script += line+'\n'
-        script += 'cd %s\n' % self.runPath
-        script += self.setupCASTOR()
-        script += self.dumpEnvVars()
-        script += '%s\n' % self.jobTransformCmd
-        script += 'echo "----- JobTransform shell script end  ---------"\n'
-        
-        return script
-
-    
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------
-# --------------------------------------------------------------------------        
-
-class PseudoScriptWriter(ScriptWriter):
-    def __init__(self, argsForScripts, logger):
-        ScriptWriter.__init__(self, argsForScripts, logger)
-        
-    def makeScript(self):
-
-         script = '#! /bin/bash\n'
-         script += 'set +e\n'
-         script += self.indicateThisIsRTT()                 
-         script += 'cd '
-         script += self.runPath+'\n'
-         for line in self.cmtLines: script += line+'\n'
-         script += 'JOBOPTSEARCHPATH=,%s:$JOBOPTSEARCHPATH\n' % self.runPath
-         script += 'cd '+self.runPath+'\n'
-         script += self.libLine+'\n'
-         script += self.dumpEnvVars() + '\n'        
-         script += 'echo "one: 1  \n two:2 \n three:3 " > monitoringData.dat\n'
-         script += 'echo "four: 4  \n five:5 \n six:6 " > someData.dat\n'
-         script += 'sleep 60\n'
-         script += 'echo " ">JobCompleteMarker.txt\n'
-
-         return script
-     
diff --git a/Tools/RunTimeTester/src/SequentialDescriptor.py b/Tools/RunTimeTester/src/SequentialDescriptor.py
deleted file mode 100644
index a5613db033f594e15b33efb4bf5efabd7a24e713..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/SequentialDescriptor.py
+++ /dev/null
@@ -1,9 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from ContainerDescriptor import ContainerDescriptor
-
-class SequentialDescriptor(ContainerDescriptor):
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-        ContainerDescriptor.__init__(self, paths, logger, jobSerialNumber)
-        self.name = 'Sequential'
-        self.identifiedName = self.name+str(self.jobSerialNumber)
diff --git a/Tools/RunTimeTester/src/ShellCommand.py b/Tools/RunTimeTester/src/ShellCommand.py
deleted file mode 100755
index 2fe3e1cd65095840abd569b10c55843a2728236c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ShellCommand.py
+++ /dev/null
@@ -1,98 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time, os
-from formatCollection  import formatCollection
-from popen2            import popen4, Popen4
-from exc2string2       import exc2string2
-from string            import replace
-from RTTSException     import RTTTimeOut
-# -------------------------------------------------------------------------
-
-class ShellCommand:
-    def __init__(self, lines, logger, nLinesToReport=100, timeout=None, silent=False):
-
-
-        self.silent = silent
-        self.logger = logger
-        assert(nLinesToReport >=-1) # -1 - do not limit the number of lines
-        lines = [str(line.strip())+';' for line in lines]
-        
-        lastLine = lines[-1]
-        lastLine = lastLine[:-1]
-
-        self.nLinesToReport = nLinesToReport
-        try:
-            self.timeout = int(timeout)
-        except:
-            self.timeout = None
-        
-        cmd = reduce(lambda x, y: x+y, lines)
-        cmd = cmd.replace(';;',';')
-        self.cmd = cmd[:-1]
-    
-        if not self.silent: self.logger.debug('issuing command '+str(self.cmd))
-
-        self.reply = []
-        
-        try:
-            self.launch()
-        except RTTTimeOut:
-            msg = 'Timeout  %s exceeded, aborting command %s' % (str(self.timeout), str(self.cmd))
-            self.reply = [msg]
-        except:
-            msg = 'Exception on cmd %s: \nTraceback: %s' % (str(self.cmd), exc2string2())
-            self.reply = [msg]
-        
-    def launch(self):
-        self.started = time.time()
-        po = Popen4(self.cmd)
-        self.pid = po.pid
-
-        output = []
-        while po.poll() == -1:
-            time.sleep(3)
-            output.extend(po.fromchild.readlines())
-
-            if self.hasTimedOut():
-                msg  = 'Time up for ShellCommand process with pid: ' + str(self.pid) + '. It will now be killed.'
-                msg += 'The commands that were launched by this process were: \n' + str(self.cmd)
-                self.logger.debug(msg)
-                self.suicide()
-                del po
-                raise RTTTimeOut(msg)
-
-        del po
-        self.reply = output
-
-    def hasTimedOut(self):
-        if self.timeout:
-            return (time.time() - self.started) > self.timeout
-        return False
-
-    def getRawReply(self):
-        return self.reply
-        
-    def getReply(self):    
-        # remove the EOL if present
-        reply  = [replace(line, '\n', '' ) for line in self.reply]
-        lenReply = len(reply)
-        
-        if not self.silent and self.nLinesToReport>-1:
-            if lenReply < self.nLinesToReport:
-                m = 'Command: %s Reply:\n' % (str(self.cmd))
-                self.logger.debug(m)
-                [self.logger.debug('%s' % str(l)) for l in reply]
-            else:
-                msg  = 'ShellCommand - '
-                msg += 'total number of lines in reply  %d ' % lenReply
-                msg += 'Truncating to %d lines\n' % self.nLinesToReport
-                msg += formatCollection(reply[:self.nLinesToReport]
-                                        )
-                self.logger.debug(msg)
-
-        return reply
-            
-    def suicide(self):
-        self.logger.warning('Received request to terminate child process: ' + str(self.pid))
-        os.system('kill -9 ' + str(self.pid))
-        self.logger.warning('Child process ' + str(self.pid) + ' killed!')
diff --git a/Tools/RunTimeTester/src/Singleton.py.obsolete b/Tools/RunTimeTester/src/Singleton.py.obsolete
deleted file mode 100644
index 57e0f1e1afdf247cd95740553daf1965886d5f2c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Singleton.py.obsolete
+++ /dev/null
@@ -1,26 +0,0 @@
-class Singleton(object):
-    def __new__(cls, *args, **kwargs):
-        if '_inst' not in vars(cls):
-            cls._inst = object.__new__(cls, *args, **kwargs)
-        return cls._inst
-
-
-
-## class MyClass(Singleton):
-##     def __init__(self):
-##         pass
-##     def testFunc(self):
-##         print 17.25
-
-
-
-## if __name__ == '__main__':
-##     x=MyClass()
-##     y=MyClass()
-##     print id(x)
-##     print id(y)
-##     z=MyClass()
-##     print id(z)
-##     x.testFunc()
-##     y.testFunc()
-##     z.testFunc()
diff --git a/Tools/RunTimeTester/src/SiteSpecifics.py b/Tools/RunTimeTester/src/SiteSpecifics.py
deleted file mode 100755
index 08545db4ef9f61513a35d6836a3e293f1b082986..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/SiteSpecifics.py
+++ /dev/null
@@ -1,154 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-from Factory_HardCoded import HardCodedFactory
-from RTTSException     import RTTInputError
-
-"""
-Classes that specify actions needed to be taken according to the site.
-"""
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class SiteSpecifics:     
-    """ Base class """
-    def __init__(self):
-        hardCoded = HardCodedFactory(logger).create('base', None)
-    
-    def rootCommand(self):
-        """How root is called at a specific site."""
-        pass
-
-    def pathToOracleInstantClient(self,gcc_version):
-        return ''
-    
-    def addToLibs(self):
-        return ''
-
-    def addToPath(self):
-        return ''
-
-    def rootSys(self):
-        """How a root env var is called at a specific site."""
-        return ''
-
-    def preAthena(self):
-        """Actions required before invoking athena at a specific site."""
-        return ''
-
-    def preTogether(self): return ''
-    
-    def postTogether(self): return ''
-    
-    def togetherCommand(self):
-        return '/usr/local/Borland/Together/bin/Together.sh'
-
-    def site(self): pass
-
-# ***********************************************************************
-# ***********************************************************************
-# ***********************************************************************
-
-class UCLslc3SiteSpecifics(SiteSpecifics):
-    """
-    Running on SLC3 machines at UCL
-    """
-
-    def __init__(self):
-        SiteSpecifics.__init__(self)
-
-    # -----------------------------------------------------------------------
-    
-    def rootCommand(self):
-        return '/afs/cern.ch/sw/root/v4.03.02/slc3_gcc3.2.3/root/bin/root -b -q '
-
-    # -----------------------------------------------------------------------
-    
-    def rootSys(self):
-        return '/afs/cern.ch/sw/root/v4.03.02/slc3_gcc3.2.3/root'
-
-    # -----------------------------------------------------------------------
-    
-    def site(self): return 'UCLslc3'
-    
-    # -----------------------------------------------------------------------
-
-    def preAthena(self):
-        "Set up env variables required by Gaudi, but not usually set at UCL"
-        envVarsToSet = 'export HOST=$HOSTNAME\n'
-        envVarsToSet = envVarsToSet+'export OS=Linux'
-        return envVarsToSet
-
-    # -----------------------------------------------------------------------
-    
-    def pathToOracleInstantClient(self,gcc_version):
-        paths = {'gcc344':'/unix/atlasrtt/Oracle/instantclient/10.2.0.2/slc3_ia32_gcc344/lib',
-                 'gcc323':'/unix/atlas1/Oracle/instantclient/10.1.0.3-1/slc3_ia32_gcc323/lib'}
-
-        path = ''
-        try:
-            path = paths[gcc_version]
-        except KeyError:
-            m  = 'KeyError: Unable to return a path to an Oracle instant client at UCL\n'
-            m += 'because the compiler version (upon which the location depends), namely: ' + str(gcc_version) + '\n'
-            raise RTTInputError(m)
-        else:
-            return path            
-
-    # -----------------------------------------------------------------------
-    
-    def addToLibs(self):
-        r = '/usr/lib:'        
-        return r
-
-    def addToPath(self):
-        return '/usr/local/bin'
-
-# ***********************************************************************
-# ***********************************************************************
-# ***********************************************************************
-
-class CERNSiteSpecifics(SiteSpecifics):
-    """
-    For running at CERN
-    """
-    def __init__(self):
-        SiteSpecifics.__init__(self)
-        
-    def rootCommand(self):
-        return "/afs/cern.ch/sw/root/v4.03.02/slc3_gcc3.2.3/root/bin/root -b -q "
-    
-    def rootSys(self):
-        return '/afs/cern.ch/sw/root/v4.03.02/slc3_gcc3.2.3/root'
-
-    def site(self): return 'CERN'
-
-class CERNslc4SiteSpecifics(CERNSiteSpecifics):
-    """
-    For running at CERN on SLC4
-    """
-    def __init__(self):
-        CERNSiteSpecifics.__init__(self)
-
-    def rootCommand(self): ''
-    def rootSys(self): ''
-    def site(self): return 'CERNslc4'
-        
-    
-class PetersMacSiteSpecifics(SiteSpecifics):
-    "For running on Peter Sherwood's laptop"
-
-    def __init__(self):
-        SiteSpecifics.__init__(self)
-        
-    def rootCommand(self):
-        return '/sw/bin/root -b -q'
-
-class LancasterSiteSpecifics(SiteSpecifics):
-    def __init__(self):
-        SiteSpecifics.__init__(self)
-
-    def site(self): return 'Lancaster'
-
diff --git a/Tools/RunTimeTester/src/StampObject.py b/Tools/RunTimeTester/src/StampObject.py
deleted file mode 100755
index 32f89ae0be9736c28482e83171ccc75c98e7a3d1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/StampObject.py
+++ /dev/null
@@ -1,41 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Tools     import isDateStampedToday
-from Tools2    import extractReleaseFromNightlyLink
-import os.path
-
-# ---------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# ---------------------------------------------------
-
-class StampObject:
-    'Responsible for returning the nightly release'
-
-    def __init__(self, stampFile):
-        self.stampFile = stampFile
-        self.nicosNewDayHour = '19:00' # midnight "is" 19:00
-        
-    def giveUpWaitingForNightly(self):
-        return False # for now
-
-    def findNightlyRelease(self): 
-         found = False
-         release = ''
-
-         if not os.path.exists(self.stampFile):
-             logger.warning('Non-existent stamp file %s' % str(self.stampFile))
-             return (found,release)
-
-         logger.debug("Stamp file: %s" %str(self.stampFile))
-
-         if isDateStampedToday(self.stampFile, self.nicosNewDayHour):
-             release = extractReleaseFromNightlyLink(self.stampFile)
-             found = True
-         else:
-             msg = 'StampFile is not from today %s' % str(self.stampFile)
-             logger.warning(msg)
-             
-         return (found,str(release))
-    
-    
diff --git a/Tools/RunTimeTester/src/State.py b/Tools/RunTimeTester/src/State.py
deleted file mode 100755
index 36608b353d6e2903200fdb1862a1313bcaf60cb2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/State.py
+++ /dev/null
@@ -1,24 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Timer import timer
-"""
-A class to hold status information.
-Contains two statuses: primay and secondary.
-Primary status    = job monitor status
-Secondary status  = check status (for, eg regression tests).
-A time of every set operation is noted.
-"""
-
-class State:
-    def __init__(self, state):
-        self.history = []
-        self.state = ''
-        self.set(state)
-    def set(self, state):
-        if state != self.state:
-            self.state = state
-            self.timeStamp = timer()
-            self.history.append( (self.state, self.timeStamp) )
-
-    def isDone(self):
-        return self.state == 'done'
diff --git a/Tools/RunTimeTester/src/Status.py.obsolete b/Tools/RunTimeTester/src/Status.py.obsolete
deleted file mode 100755
index 7c77135bfe89322e810876e5afb7b7918f6518cf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Status.py.obsolete
+++ /dev/null
@@ -1,111 +0,0 @@
-from Timer import timer
-"""
-A class to hold status information.
-Contains two statuses: primay and secondary.
-Primary status    = job monitor status
-Secondary status  = check status (for, eg regression tests).
-A time of every set operation is noted.
-"""
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-class Status:
-    def __init__(self, status):
-        self.primaryLegals = [
-            'unborn',
-            'queued',
-            'running',
-            'error',
-            'success',
-            'timedOut',
-            'PBSfatalError',
-            'operatorCloseDown',
-            'rttTimeOutCloseDown'
-            ]
-        self.secondaryLegals = ['error', 'success','n/a']
-        self.statuses= [status,'']
-        self.timeStamp = timer()
-
-    def status(self):
-        totStatus=''
-        for status in self.statuses:
-            totStatus=totStatus+' '+status
-
-        return totStatus
-
-    def set(self, status):
-
-        if not status in self.primaryLegals:
-            print 'Cannot set Status object with illegal state %s! Ignoring request...' % str(status)
-            return
-
-        if(self.statuses[0]!=status):
-            self.statuses[0] = status
-            self.timeStamp = timer()
-
-    def setPrimary(self, status):
-
-        self.set(status)
-        
-    def setSecondary(self, status):
-        
-        if not status in self.secondaryLegals:
-            print 'Status object, illegal set, balking '+status
-            return
-        
-        if(self.statuses[1] != status):
-            self.statuses[1] = status
-            self.timeStamp = timer()
-
-    def setList(self, statusList):
-        if len(statusList)>2:
-            print 'Status object, too many inputs, balking '+len(statusList)
-            return
-
-        if len(statusList)>0: self.setPrimary(statusList[0])
-        if len(statusList)>1: self.setSecondary(statusList[1])
-            
-
-    def isFailedPrimaryStatus(self):
-        return self.primary() in ['error','timedOut']
-
-    def isFailedSecondaryStatus(self):
-        return self.secondary()=='error'
-
-    def primary(self):
-        stat = self.statuses[0]
-        # logger.debug('Primary status is: ' + str(stat))
-        return stat
-    
-    def secondary(self):
-        return self.statuses[1]
-
-    def time(self):
-        return self.timeStamp
-    
-
-#class PackageStatus:
-#    def __init__(self):
-#        self.initState              = 'initialised'
-#        self.activeState            = 'processing'
-#        self.doneState              = 'completed'
-#        self.failureState           = 'abandoned'
-#        self.current                = self.initState
-#
-#    def init(self):
-#        self.current = self.initState
-#
-#    def ongoing(self):
-#        self.current = self.activeState
-#
-#    def failure(self):
-#        self.current = self.failureState
-#
-#    def done(self):
-#        self.current = self.doneState
-#
-#    def get(self):
-#        return self.current
diff --git a/Tools/RunTimeTester/src/StatusLogger.py.obsolete b/Tools/RunTimeTester/src/StatusLogger.py.obsolete
deleted file mode 100755
index 49841493a3fb90b45bfc91cd01f28944bcbc0e03..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/StatusLogger.py.obsolete
+++ /dev/null
@@ -1,30 +0,0 @@
-from os.path import join
-from Status import Status
-
-class StatusLogger:
-    "Write the job status and time stamps to a file for debugging."
-
-    def __init__(self, path, file = 'status.txt'):
-        "store the full status file name"
-        self.file = join(path, file)
-        # resPath=='': resPath=self.paths.resPath
-    def updateStatusLog(self,
-                        job,
-                        statusObj,
-                        platform,
-                        kitOrRelease,
-                        nicosDate):
-        """  Write a file with a line per job, plus an 'Overall' line
-        giving the status. The file is written to the directory continaining
-        the results for the release being processed"""
-
-        
-        status     = statusObj.status()
-        ts         = statusObj.time()
-
-        open(self.file,"a").write("%s %s %s %s %s %s\n"  % (job,
-                                                            status,
-                                                            ts,
-                                                            platform,
-                                                            kitOrRelease,
-                                                            nicosDate))
diff --git a/Tools/RunTimeTester/src/SuperParser.py.obsolete b/Tools/RunTimeTester/src/SuperParser.py.obsolete
deleted file mode 100755
index a59e1fdb98dbd74f1ed8bb3745084de841189fc2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/SuperParser.py.obsolete
+++ /dev/null
@@ -1,92 +0,0 @@
-"""
-SuperParser parses the RTT part of the unfied test configuration file for
-job groups and jobs. The two actions are delegated to other hander modules.
-"""
-
-from JobGroupKits           import JobGroupKits
-from exc2string2            import exc2string2
-from XMLTools               import getText
-from Tools                  import sendMail
-from Factory_JobsXMLReader  import JobsXMLReaderFactory
-from JobMaker               import JobMaker
-
-import sys
-
-#from MethodTracer        import MetaTracer
-#__metaclass__ = MetaTracer
-
-# -------------------------------------------------------------------------
-
-class SuperParser:
-
-    def __init__(self, packageXMLfile, logger):
-        self.packageXMLfile = packageXMLfile
-        self.logger = logger
-        
-    def parse(self, paths):
-        """Returns a list of all minders (RTT, RTTpilot and ATN)"""
-
-        rttElement = self.packageXMLfile['rtt[0]']
-
-        try:
-            paths.mailTo   = self.packageXMLfile['rtt/mailto[0]/text()']
-        except:
-            # no mailto tag
-            paths.mailTo = ''
-
-        jobListElement  = self.packageXMLfile['rtt/jobList[0]']
-        try:
-            atnElement  = self.packageXMLfile['atn[0]']
-        except IndexError:
-            atnElement = None
-            
-        factory  = JobsXMLReaderFactory(self.logger)
-        if not factory:
-            self.logger.error('Could not make a JobsXMLReader factory')
-            return []
-
-        def makeMinders((mode, element), paths):
-            try:
-                xmlReader = factory.create(mode, element, paths)
-            except Exception, e:
-                msg  = 'Exception while parsing package xml file'
-                msg += 'mode: %s package %s \n' % (mode, paths.package)
-                msg += exc2string2() 
-                self.logger.error(msg)
-                return []
-
-            try:
-                jobMaker = JobMaker(xmlReader, paths, self.logger)
-                # the jobMaker object makes the minders in its __init__
-                # instantiate the jobMaker class
-                jobMinders  = jobMaker.jobMinders
-                self.logger.debug('Made %d jobs' % len(jobMinders))
-                return jobMinders
-            except:
-                msg  = 'Exception while parsing package xml file '
-                msg += 'mode: %s package %s \n' % (mode, paths.package)
-                msg += exc2string2() 
-                self.logger.error(msg)
-                return []
-
-        # Always prcocess RTT jobs
-        modes = {paths.mode: jobListElement} # to make RTT and RTTpilot jobs
-
-        # process ATN if requested
-        if paths.atnPilotJobs:
-            if atnElement: modes['ATN'] = atnElement 
-        
-        jobMinders = []
-        [jobMinders.extend(makeMinders(m, paths)) for m in modes.items()]
-
-        # remove pilot jobs if they are NOT requested
-        if not paths.rttPilotJobs:
-            jobMinders = [j for j in jobMinders if not j.jDescriptor.rttPilotJob]
-
-        self.logger.debug('No of kits %d' % JobGroupKits().nKits())
-        self.logger.debug('No of jobs %d' % len(jobMinders))
-        return jobMinders
-
-
-if __name__ == "__main__":
-    pass
diff --git a/Tools/RunTimeTester/src/TestRun.py b/Tools/RunTimeTester/src/TestRun.py
deleted file mode 100755
index 1b4b5e717077f60a913f8a615a8354c3924e4cd7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestRun.py
+++ /dev/null
@@ -1,153 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-TestRun subclasses threading.Thread to allow concurrent running of instances.
-
-
-Does roughly the following:
-
-1. Reads the RTT configuration file, which is used to set up a large number
-   of paths and other variables
-
-2. Uses factory method to set up a Commander (mode specific commands)
-
-3. Deletes existing  run directories
-
-4. Reads the Jobs configuration file - this sets up a list of Job Descriptors
-   (one JobDescriptor per job)
-   JobDescriptors contain info abouta job. They also create a run directory for
-   thir job.
-
-5. Submit the jobs. One JobMinder per job is created. JobMinders incoporate the
-   JobDescriptor, and then monitor the progress of the job.class 
-
-6. A poll command is periodically sent to the JobMinders. These check their job status.
-   This status is interogated, and html pages are updated with new status info.
-
-7. After all jobs have completed perform cleanup activities
-"""
-#----- RTT modules -----
-
-
-from RTTpath                  import RTTpath
-from MethodTimer              import MethodTimer
-from RTTCloseDownKeyMaker     import dictToDocument
-from makeRTTDOM               import makeRTTDOM
-import threading
-import os
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-# -------------------------------------------------------------------------
-# -------------------------------------------------------------------------
-
-
-class TestRun(threading.Thread, MethodTimer):
-    def __init__(self,
-                 jobMinders,
-                 packageName,
-                 closeDownKey,
-                 closeDownKeyMaker,
-                 logger,
-                 statusText,
-                 vetoFlag = False):
-
-        MethodTimer.__init__(self)
-
-        self.logger            = logger
-        self.closeDownKeyMaker = closeDownKeyMaker
-        self.name              = packageName 
-        self.xmlConverter      = None # set later
-        self.statusText        = statusText
-        self.vetoFlag          = vetoFlag
-        
-        m = self.__class__.__name__+' '+self.name+' ' + '__init__'
-        self.logger.debug(m)
-
-        self.emailStats        = {}
-        self.jobMinders        = jobMinders
-        self.phase             = 'Initial'
-
-        self.nActiveJobs = 0
-        self.doneJobs       = 0
-        self.successfulJobs = 0
-
-
-        self.nActiveJobs    = len(self.jobMinders)
-        self.totalJobs      = len(self.jobMinders)
-        self.closeDownKey   = closeDownKey
-        # For convenience, a copy of the key to be used to close down this
-        # job will be output to the run directory. To make the xml file
-        # start by creating a document. The output is done elsewhere.
-        self.closeDownDom = dictToDocument(self.closeDownKey.__dict__, makeRTTDOM('RTTCloseDownKey'))
-
-        self.logger.info('TestRun %d active jobs' % self.nActiveJobs)
-
-        self.removeBadMinders()
-        self.logger.info('TestRun made run directories')
-
-        self.logger.info('TestRun initialising thread')
-        threading.Thread.__init__(self)
-
-# ------------------------------------------------------------------------
-
-    def operatorCloseDownTest(self):
-        return self.closeDownKeyMaker.laterMatch(self.closeDownKey.__dict__)
-
-# -------------------------------------------------------------------------
-
-    def removeBadMinders(self):
-        """Override in derived classes."""
-        pass
-    
-#-------------------------------------------------------------------------
-
-    #  clean up operations on html pages: parent class placeholder
-    def closeDown(self):
-        pass    
-                
-# -------------------------------------------------------------------------
-
-    def run(self): pass
-
-# -------------------------------------------------------------------------
-
-    def __str__(self):
-        text = 'TestRun: no of minders: '+str(len(self.jobMinders))+'\n'
-        for job in self.jobMinders:
-            text += str(job)
-            text += '\n'
-
-        return text
-
-# -------------------------------------------------------------------------
-
-    def printMethodTimes(self):
-        m = ' -- %s --\n%s\n-- xml converter--\n%s' % (
-            self.name,
-            self.formatMethodTimes(),
-            self.xmlConverter.formatMethodTimes()
-            )
-        self.logger.debug(m)
-
-    def dumpMinders(self):
-        [m.dump() for m in self.jobMinders]
-
-
-    def dataForMonitoring(self):
-        dict = {
-            'name': self.name,
-            'nSucc': self.successfulJobs,
-            'nDone': self.doneJobs,
-            'nJobs': self.totalJobs
-            }
-        return dict
-
-    def setXMLConverter(self, xmlConverter):
-
-        self.xmlConverter = xmlConverter
-        self.xmlConverter.setTestRun(self)
-        
-        # now pass to all job minders        
-        [m.setSummaryCallBack(self.xmlConverter) for m in self.jobMinders]
-        
diff --git a/Tools/RunTimeTester/src/TestRunXMLConverter2.py b/Tools/RunTimeTester/src/TestRunXMLConverter2.py
deleted file mode 100755
index f39e35f35a5f3ec50b8bb2d60d04de2f700de3f1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestRunXMLConverter2.py
+++ /dev/null
@@ -1,210 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an aml file summary of an RTT run
-"""
-from BaseXMLConverter2 import BaseXMLConverter
-from StringIO          import StringIO
-from xml.dom.ext       import PrettyPrint
-import os.path, time, copy
-from exc2string2       import exc2string2
-from Evaluate          import Evaluate
-from sortValues        import sortValues
-
-class TestRunXMLConverter(BaseXMLConverter):
-    def __init__(self, outputFile, package, packageTag, containerPackage, elementCreator, textNodeCreator,
-                 callBackConverter, logger):
-        BaseXMLConverter.__init__(self, elementCreator, textNodeCreator, logger)
-        
-        self.nJobs          = 0     # set in setTestRun
-        self.testRun        = None  # set in setTestRun
-        self.outputFile     = outputFile
-        self.callBack       = callBackConverter
-        
-        self.minderNodes   = {}
-        self.minderDicts   = {}
-        
-        # temp = StringIO()
-        # PrettyPrint (self.node, temp)
-
-        self.overviewInfo = {
-        'nJobsTotalPackage':      0, # set in setTestRun
-        'nJobsInPackage':         0, # number that have reported
-        'nJobsDoneInPackage':     0,
-        'nJobsOpCloseInPackage':  0,
-        'nJobsSuccessInPackage':  0,
-        'nJobsFailureInPackage':  0,
-        'nPPSuccessInPackage':    0,
-        'nPPFailureInPackage':    0,
-        'nTestsInPackage':        0,
-        'nTestsSuccessInPackage': 0,
-        'nTestsFailureInPackage': 0,
-        'nRetriesInPackage':      0,
-        'nTimeOutsInPackage':     0,
-        'nBatchErrorsInPackage':  0,
-        'packageName':            package,
-        'packageTag':             packageTag,
-        'containerPackage':       containerPackage,
-        'phase':                  '', # set in setTesteRun
-        'packageStatus':          ''  # set in setTesteRun
-        }
-
-
-        # self.makePackageNode()
-        self.updatedMinders = []
-        #m = 'TestRunXMLConverter: __init__, self.node:\n%s' % (temp.getvalue())
-        #self.logger.debug(m)
-        self.hasData = False
-        
-    def setTestRun(self, tr):
-        self.testRun = tr
-        self.nJobs                             = len(tr.jobMinders)
-        self.overviewInfo['nJobsTotalPackage'] = tr.totalJobs
-        self.overviewInfo['phase']             = tr.phase
-        self.overviewInfo['packageStatus']     = tr.statusText
-        
-        self.makePackageNode()
-        
-    def makePackageNode(self):
-        # make a conversion table between attribute names of paths, and
-        # element names in the output xml file.
-        self.node      = self.elementCreator('package')
-        [self.addTextNode(self.node, item) for item in self.overviewInfo.items()]
-        if self.testRun.vetoFlag: self.node.appendChild(self.elementCreator('vetoed'))
-
-    def update(self, dict, node):
-        name = dict['name']
-        self.logger.info('TestRunXMLConverter2::update called by minder %s START' % name)
-        self.updatedMinders.append(name)
-        
-        # self.logger.error('Start TestRunXML update for minder: %s' % name)
-        
-        if name not in self.minderNodes.keys():
-            self.overviewInfo['nJobsInPackage']  += 1    
-            self.overviewInfo['nTestsInPackage'] += dict['nTests']
-
-        self.minderNodes[name] = node
-        self.minderDicts[name] = dict
-
-        nDone        = len( [1 for d in self.minderDicts.values() if d['done']])
-        nOpClose     = len( [1 for d in self.minderDicts.values() if d['operatorCloseDown']])
-        nJobsSuccess = len( [1 for d in self.minderDicts.values() if d['jobSuccess']])
-        nJobsFailure = len( [1 for d in self.minderDicts.values() if d['jobFailure']])
-        nPPSuccess   = len( [1 for d in self.minderDicts.values() if d['ppSuccess']])
-        nPPFailure =   len( [1 for d in self.minderDicts.values() if d['ppFailure']])
-
-        temp = [d['nTestsSuccess'] for d in self.minderDicts.values()]
-        nTestsSuccess = reduce(lambda x,y: x+y, temp, 0)
-        temp = [d['nTestsFailure'] for d in self.minderDicts.values()]
-        nTestsFailure = reduce(lambda x,y: x+y, temp, 0)
-
-        self.overviewInfo['nJobsDoneInPackage']      = nDone
-        self.overviewInfo['nJobsOpCloseInPackage']   = nOpClose
-        self.overviewInfo['nJobsSuccessInPackage']   = nJobsSuccess
-        self.overviewInfo['nJobsFailureInPackage']   = nJobsFailure
-        self.overviewInfo['nPPSuccessInPackage']     = nPPSuccess
-        self.overviewInfo['nPPFailureInPackage']     = nPPFailure
-        self.overviewInfo['nTestsSuccessInPackage']  = nTestsSuccess
-        self.overviewInfo['nTestsFailureInPackage']  = nTestsFailure
-
-        nto = 0
-        for d in  self.minderDicts.values():
-            try: # only in the dict for BatchJobMinders
-                if d['timedOut']:
-                    nto += 1
-            except:
-                pass
-            
-        self.overviewInfo['nTimeOutsInPackage']  = nto
-
-
-        nbe = 0
-        for d in  self.minderDicts.values():
-            try: # only in the dict for BatchJobMinders
-                if d['batchError']:
-                    nbe += 1
-            except:
-                pass
-            
-        self.overviewInfo['nBatchErrorsInPackage']  = nbe
-
-
-        mretries = []
-        [mretries.append(d['nRetries']) for d in self.minderDicts.values()]
-        # print 'TestRunXMLCOnverter2: mretries ' + str(mretries)
-        nretries = reduce(lambda x,y: x+y, mretries, 0)
-
-        self.overviewInfo['nRetriesInPackage']  = nretries
-
-        self.overviewInfo['phase']              = self.testRun.phase
-
-        self.overviewInfo['packageStatus']      = self.testRun.statusText
-
-        # maintain the node order in the dom. Required for RTT self tests
-        self.makePackageNode()
-        sortedNodes = sortValues(self.minderNodes)
-        # [self.node.appendChild(node) for node in self.minderNodes.values()]
-        [self.node.appendChild(node) for node in sortedNodes]
-
-        self.hasData = True
-        self.logger.info('TestRunXMLConverter2::update called by minder %s END' % name)
-
-    def outputToFile(self):
-        space = ' '
-        h = None
-        try:
-            h = open(self.outputFile, 'w')
-            h.write(self.node.toprettyxml(3*space, '\n'))
-            h.close()
-        except:
-            self.logger.error('Unable to output package info XML file. Traceback: \n%s' % exc2string2())
-        if h: h.close()
-    
-    def finalise(self):
-        self.logger.debug('Entering TestRunXMLConverter.finalise')
-        self.overviewInfo['packageStatus'] = self.testRun.statusText
-        self.makePackageNode()
-        [self.node.appendChild(node) for node in self.minderNodes.values()]
-        self.logger.debug('Package node updated with final testrun status text')
-        updatedMinders = []
-        infoTuple = (self.testRun.dataForMonitoring(), self.node.cloneNode(True), updatedMinders)
-        self.logger.debug('Calling back...')
-        self.callBack.update(infoTuple)
-        self.logger.debug('Callback complete. Exiting TestRunXMLConverter.finalise method.')
-
-        
-    def pushData(self):
-        self.logger.debug('begin %s:pushData() for package: %s' % (self.__class__.__name__, self.testRun.name))
-        dict = self.testRun.dataForMonitoring()
-        # -------------- debug -------------
-
-        #nMindersFromPackage = int(Evaluate('nJobsInPackage/text()', self.node))
-        #nMindersFromMinders = len(Evaluate('minder', self.node))
-        # print str(self.node)
-        #temp = StringIO()
-        #PrettyPrint (self.node, temp)
-        #assert nMindersFromPackage == nMindersFromMinders
-                                  
-        # ----------------------------------
-        self.logger.debug('TestRunXMLConverter::callBack: %s, TestRunXMLConverter::hasData: %s' % (str(self.callBack), str(self.hasData)))
-        if self.callBack and self.hasData:
-            # startTime = time.time()
-            updatedMinders = copy.copy(self.updatedMinders)
-            self.logger.debug('Pushing TestRun data for package: %s' % self.testRun.name)
-            self.callBack.update( (dict, self.node.cloneNode(True), updatedMinders) )
-            self.hasData = False
-            self.updatedMinders = []
-            self.logger.debug('Pushed TestRun data for package: %s' % self.testRun.name)
-            self.outputToFile() # write out the node to file
-
-    def __del__(self):
-        if self.node: self.node.unlink()
-        
-class FailureTestRunXMLConverter(TestRunXMLConverter):
-    def __init__(self, outputFile, package, packageTag, containerPackage, elementCreator, textNodeCreator,
-                 callBackConverter, logger):
-        TestRunXMLConverter.__init__(self, outputFile, package, packageTag, containerPackage, elementCreator, textNodeCreator,
-                                     callBackConverter, logger)
-    def pushData(self):
-        self.hasData = True
-        TestRunXMLConverter.pushData(self)
diff --git a/Tools/RunTimeTester/src/TestSuite2.log b/Tools/RunTimeTester/src/TestSuite2.log
deleted file mode 100755
index fba6bdf7e4ba954d82c1d0d6feebf052a584919d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuite2.log
+++ /dev/null
@@ -1 +0,0 @@
-TestSuite2 results:
\ No newline at end of file
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Readme.txt
deleted file mode 100755
index 5fdd073e6a8f14244f02af607cc9ac9ac43ad5fe..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/Readme.txt
+++ /dev/null
@@ -1,4 +0,0 @@
-project builds
-RTT provides conf files
-numbered release 11.3.0
-
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/cfg/rttUnitTests.xml
deleted file mode 100755
index cf1402887cd7db43e0ca124c4695c233499c2ecb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,70 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd"> 
-
-<!-- <!DOCTYPE rttconfig SYSTEM "file:///Users/peter/RTTtop/DTD/RTT.dtd"> -->
-
-<rttconfig>
-   <mode>PBS</mode>
-   <release>11.3.0</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>TESTSUITESCRATCH/rttUnitTestMoniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-
-   <confFileSrc>
-      <rttProvides>
-        <packageConfFile>
-           <package>Simulation/Atlfast/TestAtlfast</package>
-           <pathToFile>TESTSUITESCRATCH/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasReconstruction/1.3.0</cmtpath>
-           <version>TestAtlfast-00-00-61</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/AtlasReconstruction/1.3.0/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/AtlasReconstruction/1.3.0/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-        <packageConfFile>
-           <package>Tools/RunTimeTester</package>
-           <pathToFile>TESTSUITESCRATCH/xml/RunTimeTester_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/1.3.0</cmtpath>
-           <version>RunTimeTester-00-00-82</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/1.3.0/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/1.3.0/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-      </rttProvides>
-      <installedConfFiles/>
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <!-- <rttKitTest/> -->
-   <releaseType>project</releaseType>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 84c311021ced0b7d38413fe7bc631eabccb431bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,139 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-                  
-        <!-- <rulechecker> -->
-        <!--    <package>Database/NovaBlob</package>-->
-        <!--    <queue>short</queue>-->
-        <!-- </rulechecker> -->
-
-           <rulechecker>  
-              <package>Database</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-
-           <rulechecker>  
-              <package>Simulation/Atlfast</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-         
-         <!-- <qametrics> -->
-             <!--  <package>Simulation/Atlfast</package> -->
-            <!--   <queue>short</queue> -->
-          <!--  </qametrics> -->
-         
-           <kitvalidation> 
-              <queue>medium</queue> 
-           </kitvalidation>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <action>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </action>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <keepFilePattern>badViolFiles.*</keepFilePattern>
-
-            <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>WatcherFilesWithStringsFinder</testname>
-               <arg>
-                  <argname>outputFile</argname>
-                  <argvalue>badViolFiles</argvalue>
-               </arg>
-               <arg>
-                  <argname>dirWithFiles</argname>
-                  <argvalue>resultsDir</argvalue>
-               </arg>
-               <listarg>
-                  <listargname>searchPatterns</listargname>
-                  <listargvalue>expecting</listargvalue>
-               </listarg>
-               <listarg>
-                  <listargname>inputFilePatterns</listargname>
-                  <listargvalue>*.viol</listargvalue>
-               </listarg>
-
-            </action>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures1/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Readme.txt
deleted file mode 100755
index 9c6d34092b3b5685505b62fa787bdf1f180384b2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/Readme.txt
+++ /dev/null
@@ -1,3 +0,0 @@
-project builds
-package conf files from release
-numbered release 11.3.0
\ No newline at end of file
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/cfg/rttUnitTests.xml
deleted file mode 100755
index 06b8fdca40111113bb4d890d1d733a3145022969..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,43 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-
-
-<!-- <!DOCTYPE rttconfig SYSTEM "RTT.dtd"> -->
-<rttconfig>
-   <mode>PBS</mode>
-   <release>11.3.0</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>TESTSUITESCRATCH/rttUnitTestMoniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-
-   <confFileSrc>
-      <installedConfFiles/>
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <releaseType>project</releaseType>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 84c311021ced0b7d38413fe7bc631eabccb431bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,139 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-                  
-        <!-- <rulechecker> -->
-        <!--    <package>Database/NovaBlob</package>-->
-        <!--    <queue>short</queue>-->
-        <!-- </rulechecker> -->
-
-           <rulechecker>  
-              <package>Database</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-
-           <rulechecker>  
-              <package>Simulation/Atlfast</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-         
-         <!-- <qametrics> -->
-             <!--  <package>Simulation/Atlfast</package> -->
-            <!--   <queue>short</queue> -->
-          <!--  </qametrics> -->
-         
-           <kitvalidation> 
-              <queue>medium</queue> 
-           </kitvalidation>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <action>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </action>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <keepFilePattern>badViolFiles.*</keepFilePattern>
-
-            <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>WatcherFilesWithStringsFinder</testname>
-               <arg>
-                  <argname>outputFile</argname>
-                  <argvalue>badViolFiles</argvalue>
-               </arg>
-               <arg>
-                  <argname>dirWithFiles</argname>
-                  <argvalue>resultsDir</argvalue>
-               </arg>
-               <listarg>
-                  <listargname>searchPatterns</listargname>
-                  <listargvalue>expecting</listargvalue>
-               </listarg>
-               <listarg>
-                  <listargname>inputFilePatterns</listargname>
-                  <listargvalue>*.viol</listargvalue>
-               </listarg>
-
-            </action>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures2/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Readme.txt
deleted file mode 100755
index fad8d2ba99f5d8ab773af879e1d82efddc7b5b9c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/Readme.txt
+++ /dev/null
@@ -1,3 +0,0 @@
-project builds
-RTT provides conf files
-nightly
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/cfg/rttUnitTests.xml
deleted file mode 100755
index 6bfd05883a672f18863208f6283dd247bfc6d302..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,68 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-<!-- <!DOCTYPE rttconfig SYSTEM "file:///Users/peter/RTTtop/DTD/RTTpeter.dtd"> -->
-
-<rttconfig>
-   <mode>PBS</mode>
-   <release>atlrel_4</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>TESTSUITESCRATCH/rttUnitTestMoniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-
-   <confFileSrc>
-      <rttProvides>
-        <packageConfFile>
-           <package>Simulation/Atlfast/TestAtlfast</package>
-           <pathToFile>TESTSUITESCRATCH/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4</cmtpath>
-           <version>TestAtlfast-00-00-61</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-        <packageConfFile>
-           <package>Tools/RunTimeTester</package>
-           <pathToFile>TESTSUITESCRATCH/xml/RunTimeTester_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4</cmtpath>
-           <version>RunTimeTester-00-00-82</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-      </rttProvides>
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <!-- <rttKitTest/> -->
-   <releaseType>project</releaseType>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 84c311021ced0b7d38413fe7bc631eabccb431bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,139 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-                  
-        <!-- <rulechecker> -->
-        <!--    <package>Database/NovaBlob</package>-->
-        <!--    <queue>short</queue>-->
-        <!-- </rulechecker> -->
-
-           <rulechecker>  
-              <package>Database</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-
-           <rulechecker>  
-              <package>Simulation/Atlfast</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-         
-         <!-- <qametrics> -->
-             <!--  <package>Simulation/Atlfast</package> -->
-            <!--   <queue>short</queue> -->
-          <!--  </qametrics> -->
-         
-           <kitvalidation> 
-              <queue>medium</queue> 
-           </kitvalidation>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <action>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </action>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <keepFilePattern>badViolFiles.*</keepFilePattern>
-
-            <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>WatcherFilesWithStringsFinder</testname>
-               <arg>
-                  <argname>outputFile</argname>
-                  <argvalue>badViolFiles</argvalue>
-               </arg>
-               <arg>
-                  <argname>dirWithFiles</argname>
-                  <argvalue>resultsDir</argvalue>
-               </arg>
-               <listarg>
-                  <listargname>searchPatterns</listargname>
-                  <listargvalue>expecting</listargvalue>
-               </listarg>
-               <listarg>
-                  <listargname>inputFilePatterns</listargname>
-                  <listargvalue>*.viol</listargvalue>
-               </listarg>
-
-            </action>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures3/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Readme.txt
deleted file mode 100755
index c545ead86d4153c51a70d2d8a26ebf53dbb24192..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/Readme.txt
+++ /dev/null
@@ -1,4 +0,0 @@
-project builds
-package conf files from release
-release run 
-12.0.0
\ No newline at end of file
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/cfg/rttUnitTests.xml
deleted file mode 100755
index 813d47fb3ce864caef0f2166f66937f0d09018c8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,58 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-
-
-<!-- <!DOCTYPE rttconfig SYSTEM "RTT.dtd"> -->
-<rttconfig>
-   <mode>PBS</mode>
-   <release>12.0.0</release>
-   <refRelease>12.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>TESTSUITESCRATCH/rttUnitTestMoniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-
-   <confFileSrc> 
-     <rttProvides>
-       <packageConfFile>
-           <package>Tools/RunTimeTester</package>
-           <pathToFile>TESTSUITESCRATCH/xml/RunTimeTester_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/releases/AtlasCore/1.2.0</cmtpath>
-           <version>RunTimeTester-00-01-01</version>
-           <installArea>/afs/cern.ch/atlas/software/releases/AtlasCore/1.2.0/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/releases/AtlasCore/1.2.0/InstallArea/share</shareArea>
-        </packageConfFile>
-
-      </rttProvides>
-
-       <installedConfFiles/>
-   </confFileSrc>
-
-    <packageAccept>JiveXML</packageAccept> 
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <distArea>/afs/cern.ch/atlas/software/releases</distArea>
-   <releaseType>project</releaseType>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 84c311021ced0b7d38413fe7bc631eabccb431bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,139 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-                  
-        <!-- <rulechecker> -->
-        <!--    <package>Database/NovaBlob</package>-->
-        <!--    <queue>short</queue>-->
-        <!-- </rulechecker> -->
-
-           <rulechecker>  
-              <package>Database</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-
-           <rulechecker>  
-              <package>Simulation/Atlfast</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-         
-         <!-- <qametrics> -->
-             <!--  <package>Simulation/Atlfast</package> -->
-            <!--   <queue>short</queue> -->
-          <!--  </qametrics> -->
-         
-           <kitvalidation> 
-              <queue>medium</queue> 
-           </kitvalidation>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <action>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </action>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <keepFilePattern>badViolFiles.*</keepFilePattern>
-
-            <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>WatcherFilesWithStringsFinder</testname>
-               <arg>
-                  <argname>outputFile</argname>
-                  <argvalue>badViolFiles</argvalue>
-               </arg>
-               <arg>
-                  <argname>dirWithFiles</argname>
-                  <argvalue>resultsDir</argvalue>
-               </arg>
-               <listarg>
-                  <listargname>searchPatterns</listargname>
-                  <listargvalue>expecting</listargvalue>
-               </listarg>
-               <listarg>
-                  <listargname>inputFilePatterns</listargname>
-                  <listargvalue>*.viol</listargvalue>
-               </listarg>
-
-            </action>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures4/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Readme.txt
deleted file mode 100755
index 20319cda2292e185b34619b78b6cec3e90684724..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/Readme.txt
+++ /dev/null
@@ -1,4 +0,0 @@
-project builds
-RTT provides conf files
-nightly
-N.0.X
\ No newline at end of file
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/cfg/rttUnitTests.xml
deleted file mode 100755
index a402b4328bfb4e36827633cbc72a68dc9eaa4a4f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,68 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-<!-- <!DOCTYPE rttconfig SYSTEM "file:///Users/peter/RTTtop/DTD/RTTpeter.dtd"> -->
-
-<rttconfig>
-   <mode>PBS</mode>
-   <release>atlrel_3</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.0.X</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>TESTSUITESCRATCH/rttUnitTestMoniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-
-   <confFileSrc>
-      <rttProvides>
-        <packageConfFile>
-           <package>Simulation/Atlfast/TestAtlfast</package>
-           <pathToFile>TESTSUITESCRATCH/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4</cmtpath>
-           <version>TestAtlfast-00-00-61</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasReconstruction/rel_4/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-        <packageConfFile>
-           <package>Tools/RunTimeTester</package>
-           <pathToFile>TESTSUITESCRATCH/xml/RunTimeTester_TestConfiguration.xml</pathToFile>
-
-           <cmtpath>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4</cmtpath>
-           <version>RunTimeTester-00-00-82</version>
-           <installArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/InstallArea</installArea>
-
-           <shareArea>/afs/cern.ch/atlas/software/builds/nightlies/dev/AtlasCore/rel_4/InstallArea/share</shareArea>
-        </packageConfFile>
-
-
-      </rttProvides>
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <!-- <rttKitTest/> -->
-   <releaseType>project</releaseType>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/RunTimeTester_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/RunTimeTester_TestConfiguration.xml
deleted file mode 100755
index 84c311021ced0b7d38413fe7bc631eabccb431bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/RunTimeTester_TestConfiguration.xml
+++ /dev/null
@@ -1,139 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-      <refRelease>10.0.0</refRelease>
-      
-      <jobList>
-                  
-        <!-- <rulechecker> -->
-        <!--    <package>Database/NovaBlob</package>-->
-        <!--    <queue>short</queue>-->
-        <!-- </rulechecker> -->
-
-           <rulechecker>  
-              <package>Database</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-
-           <rulechecker>  
-              <package>Simulation/Atlfast</package> 
-              <queue>medium</queue>  
-           </rulechecker> 
-         
-         <!-- <qametrics> -->
-             <!--  <package>Simulation/Atlfast</package> -->
-            <!--   <queue>short</queue> -->
-          <!--  </qametrics> -->
-         
-           <kitvalidation> 
-              <queue>medium</queue> 
-           </kitvalidation>
-
-      </jobList>
-      
-      <jobGroups>
-         
-         <jobGroup name="RuleChecker" parent="Top">
-            
-            <keepFilePattern>*.viol</keepFilePattern>
-            <keepFilePattern>*.i</keepFilePattern>
-            <keepFilePattern>srcFileInfo.txt</keepFilePattern>
-            <keepFilePattern>*.xml</keepFilePattern>
-            <keepFilePattern>*.html</keepFilePattern>
-            
-            <storeInDatabasePattern>*.xml</storeInDatabasePattern>
-            <action>
-               <modulename>RuleCheckerLibraryTools</modulename>
-               <testname>Viol2XMLLocal</testname>
-            </action>
-
-            <logChecker class="SimpleLogChecker">
-               <searchString>rulechecker ok.</searchString>
-            </logChecker>
-
-         </jobGroup>
-
-         <jobGroup name="RuleCheckerWatcher" parent="Top">
-
-            <keepFilePattern>*.html</keepFilePattern>
-            <keepFilePattern>badViolFiles.*</keepFilePattern>
-
-            <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>WatcherFilesWithStringsFinder</testname>
-               <arg>
-                  <argname>outputFile</argname>
-                  <argvalue>badViolFiles</argvalue>
-               </arg>
-               <arg>
-                  <argname>dirWithFiles</argname>
-                  <argvalue>resultsDir</argvalue>
-               </arg>
-               <listarg>
-                  <listargname>searchPatterns</listargname>
-                  <listargvalue>expecting</listargvalue>
-               </listarg>
-               <listarg>
-                  <listargname>inputFilePatterns</listargname>
-                  <listargvalue>*.viol</listargvalue>
-               </listarg>
-
-            </action>            
-            <summary>RuleCheckerSummary.html</summary>
-
-         </jobGroup>
-         
-         
-         <jobGroup name="PBSTest" parent="Top"/>
-         
-         <jobGroup name="KitValidation" parent="Top">
-            
-            <keepFilePattern>Summary.results</keepFilePattern>
-
-            <checker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-               <vetoString>[FAILED]</vetoString>
-            </checker>
-            
-            <test>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-               
-               <arg>
-                  <argname>inputFile</argname>
-                  <argvalue>*_log</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>outputFile</argname>
-                  <argvalue>Summary.results</argvalue>
-               </arg>
-
-               <arg>               
-                  <argname>searchList</argname>
-                  <argvalue>PASSED,FAILED</argvalue>
-               </arg>
-               
-               <arg>
-                  <argname>vetoList</argname>
-                  <argvalue>DEBUG</argvalue>
-               </arg>
-               
-               
-            </test>
-            
-            
-            
-            <logChecker class="SimpleLogChecker">
-               <searchString>Validation [</searchString>
-            </logChecker>
-         </jobGroup>
-                  
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures5/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/Readme.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/Readme.txt
deleted file mode 100755
index 11fde62b3cf0270dca9a685b3f9b0c4c555174b7..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/Readme.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-project builds
-package conf files from release
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/status.txt b/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/status.txt
deleted file mode 100755
index d3cc53b22df37fe75c8cc579fe22a2951b1a7b9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Fixtures6/status.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-Overall  Started  06/01/16  15:39:32 slc3 release 2006/1/16
-Overall  completed  06/01/18  04:40:00 slc3 release 2006/1/16
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py b/Tools/RunTimeTester/src/TestSuiteFixtures/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
deleted file mode 100755
index e1ef31b0043c2a31afc7e512d988db138a55ba26..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-junk
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTests.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTests.xml
deleted file mode 100755
index 1a67f0efcbaefc58da2ebea844cfed8ab68d4754..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTests.xml
+++ /dev/null
@@ -1,51 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-
-
-<!-- <!DOCTYPE rttconfig SYSTEM "RTT.dtd"> -->
-<rttconfig>
-   <mode>PBS</mode>
-   <release>11.0.0</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-
-
-   <confFileSrc>
-      <rttProvides>
-        <!-- <packageConfFile> -->
-           <!-- <package>HelloWorld</package> -->
-           <!-- <pathToFile>/local_disk/sherwood/RTT/HelloWorld_TestConfiguration.xml</pathToFile> -->
-        <!-- </packageConfFile> -->
-        <packageConfFile>
-           <package>Simulation/Atlfast/TestAtlfast</package>
-           <!-- <pathToFile>/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/test/TestAtlfast_TestConfiguration.xml</pathToFile> -->
-           <pathToFile>TESTSUITESCRATCH/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-           <cmtdir>/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/</cmtdir>
-        </packageConfFile>
-      </rttProvides>
-      <!-- <localConfFiles> -->
-         <!-- <package>Simulation/Atlfast/TestAtlfast</package> -->
-      <!-- </localConfFiles> -->
-      <!-- <installedConfFiles/> -->
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-   <!-- <rttKitTest/> -->
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTestsWithRelease.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTestsWithRelease.xml
deleted file mode 100755
index f7095aee1029cbc577ee90ff35104238e51bee8d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/cfg/rttUnitTestsWithRelease.xml
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-
-<!DOCTYPE rttconfig SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/RTT.dtd">
-
-
-<!-- <!DOCTYPE rttconfig SYSTEM "RTT.dtd"> -->
-<rttconfig>
-   <mode>PBS</mode>
-   <release>11.0.0</release>
-   <refRelease>11.0.0</refRelease>
-   <build>opt</build>
-   <branch>N.X.0</branch>
-   <workBasePath>TESTSUITESCRATCH/Work</workBasePath>
-   <resultsBasePath>TESTSUITESCRATCH/Results</resultsBasePath>
-   <site>CERN</site>
-   <dataSetCatalog>TESTSUITESCRATCH/xml/rttUnitTestDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>TESTSUITESCRATCH/xml/rttUnitTestTopLevelJobGroup.xml</jobGroupConfig>
-   <refDB>
-      <file>TESTSUITESCRATCH/rttUnitTestRefFile.db</file>
-      <enable>1</enable>
-   </refDB>
-
-
-   <confFileSrc>
-      <rttProvides>
-        <!-- <packageConfFile> -->
-           <!-- <package>HelloWorld</package> -->
-           <!-- <pathToFile>/local_disk/sherwood/RTT/HelloWorld_TestConfiguration.xml</pathToFile> -->
-        <!-- </packageConfFile> -->
-        <packageConfFile>
-           <package>AthenaAtlfast</package>
-           <!-- <pathToFile>/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/test/TestAtlfast_TestConfiguration.xml</pathToFile> -->
-           <pathToFile>TESTSUITESCRATCH/xml/TestAtlfast_TestConfiguration.xml</pathToFile>
-           <cmtdir>/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/</cmtdir>
-        </packageConfFile>
-      </rttProvides>
-      <!-- <localConfFiles> -->
-         <!-- <package>Simulation/Atlfast/TestAtlfast</package> -->
-      <!-- </localConfFiles> -->
-      <installedConfFiles/>
-   </confFileSrc>
-
-   <!-- <packageAccept>HelloWorld</packageAccept> -->
-   <!-- <packageVeto>TestAtlfast</packageVeto> -->
-<!--   <testMode>NoNewRunDirs</testMode> -->
-   <maxTime>
-      <hours>1</hours>
-   </maxTime>
-</rttconfig>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/TestAtlfast_TestConfiguration.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/xml/TestAtlfast_TestConfiguration.xml
deleted file mode 100755
index 6979396d64b0ced5c150e14b61787cd9090cd24a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/TestAtlfast_TestConfiguration.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                             
-                                                                                                                                                             
-<unifiedTestConfiguration>
-<atn/>
-<kv/>
-<rtt  xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-   <refRelease>10.0.0</refRelease>
-   <jobList>
-       <athena>
-          <!-- <rttKitTestJob/> -->
-          <options>Atlfast_POOL_False_low_cone_cells.py</options>
-          <package>Simulation/Atlfast/TestAtlfast</package>
-          <group>Atlfast</group>
-          <queue>short</queue>
-          <dataset>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</dataset>
-          <asciiRefFile>outFile_simpleTest.txt</asciiRefFile>
-       </athena>
-   </jobList>
-
-<jobGroups>
-      <jobGroup name="AthenaAtlfast" parent="Athena">
-          
-         <keepFilePattern>histo.hbook</keepFilePattern>
-         <keepFilePattern>ntuple.root</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron.ps</keepFilePattern>
-         <keepFilePattern>isolatedElectron_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton.ps</keepFilePattern>
-         <keepFilePattern>isolatedPhoton_regression.ps</keepFilePattern>
-         <keepFilePattern>nonIsolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>isolatedMuon.ps</keepFilePattern>
-         <keepFilePattern>cells.ps</keepFilePattern>
-         <keepFilePattern>clusters.ps</keepFilePattern>
-         <keepFilePattern>jets.ps</keepFilePattern>
-         <keepFilePattern>bjets.ps</keepFilePattern>
-         <keepFilePattern>eventHeader1.ps</keepFilePattern>
-         <keepFilePattern>eventHeader2.ps</keepFilePattern>
-         <keepFilePattern>*_simpleTest.txt</keepFilePattern>
- 
- 
-         <!--<asciiRefFile>outFile_simpleTest.txt</asciiRefFile> -->
-          
-         <binRefFile>histo.hbook</binRefFile>
-          
-         <checker class="AtlfastRegressionChecker"/>
-          
-         <auxFilePattern>Atlfast*.dat</auxFilePattern>
-         <auxFilePattern>*.tab</auxFilePattern>
-         <auxFilePattern>PDG*</auxFilePattern>
-         <auxFilePattern>atlfast.C</auxFilePattern>
-     
-      </jobGroup>
-       
-   </jobGroups>
-</rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestDataSetCatalog.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestDataSetCatalog.xml
deleted file mode 100755
index 40f76ec74598f7a164e72863feb2377ccef66a95..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestDataSetCatalog.xml
+++ /dev/null
@@ -1,154 +0,0 @@
-<dsns>
-   <!-- EGAMMA TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0000.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0000.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0001.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0001.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0002.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0002.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0003.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0003.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0004.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0004.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0005.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0005.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0006.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0006.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0007.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0007.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0008.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0008.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0009.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0009.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0010.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0010.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0011.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0011.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0012.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0012.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0013.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0013.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0014.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0014.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0015.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0015.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0016.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0016.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0017.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0017.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0018.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0018.root</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/user/c/costanzo/validation/data/dc2val.e_e100_eta25.006005.g4dig/dc2val.e_e100_eta25.006005.g4dig.0019.root</logical>
-      <physical>/local_disk/data/dc2val.e_e100_eta25.006005.g4dig.0019.root</physical>
-   </dsn>
-   
-   <!-- ATLFAST TESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001002/dc1.001002.simul.0001.test.h130_4e.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001002.simul.0001.test.h130_4e.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/dc1.001003.simul.redig/data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001003.simul.redig._00001.test.h130_2e2mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001004/dc1.001004.simul.0001.test.h130_2gam.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001004.simul.0001.test.h130_2gam.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/simul/data/001005/dc1.001005.simul.0001.test.h130_4mu.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.001005.simul.0001.test.h130_4mu.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>jet_25</logical> <!-- do not know the castor name -->
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002001.simul.0001.hlt.pythia_jet_25.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc1/lumi02/002046/data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</logical>
-      <physical>/unix/atlas/AtlasTests/Data/dc1.002046.lumi02.00001.hlt.Z_ee.zebra</physical>
-   </dsn>
-   
-   <!-- JIVEXMLTESTS -->
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/g4dig805/dc2.002885.pyt_z_ee.g4dig805/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002885.pyt_z_ee.g4dig805._0001.pool.root</physical>
-   </dsn>
-   <!-- TRIGGER TESTS -->
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.zebra</physical>
-   </dsn>
-   
-   <dsn>
-      <logical>dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</logical><!-- Simon George gives this with $ATLFASTTEST stuck in front -->
-      <physical>/local_disk/data/dc1.001003.simul.redig.00001.test.h130_2e2mu_extract.re</physical>
-   </dsn>
-   <dsn>
-      <logical>/castor/cern.ch/atlas/project/dc2/preprod/evgen804/dc2.002898.pyt_h120_gamgam.evgen804/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</logical>
-      <physical>/local_disk/data/dc2.002898.pyt_h120_gamgam.evgen804._0001.pool.root</physical>
-   </dsn>
-   
-</dsns>
diff --git a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestTopLevelJobGroup.xml b/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestTopLevelJobGroup.xml
deleted file mode 100755
index 1f6e940fc1dd00fe091c042c1a4d3475998568ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteFixtures/xml/rttUnitTestTopLevelJobGroup.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-         
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         
-         </jobGroup>
-   
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/src/TestSuiteRefDicts.py.obsolete b/Tools/RunTimeTester/src/TestSuiteRefDicts.py.obsolete
deleted file mode 100755
index b2a498b995a31c6967214fe1e62adcfe2a70e99f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteRefDicts.py.obsolete
+++ /dev/null
@@ -1,178 +0,0 @@
-refPathsDict = {
-    'libToolsLoc': 'releaseSharePath',
-    'rttTagBeingRun':'RunTimeTester-00-00-78',
-    'kitID': '',
-    'installPath': None,
-    'refRelease': u'11.0.0',
-    'distArea': '/afs/cern.ch/atlas/software/dist',
-    'site': u'CERN',
-    'localConfFiles': False,
-    'installedConfFiles': False,
-    'testMode': 'Off',
-    'refDBenable': u'1',
-    'branchesNotToDisplay': [],
-    'packageTag': 'TestAtlfast-00-00-55',
-    'versionDir': None,
-    'releaseType': 'monolithic',
-    'workDirs': {'workBase/build/branch': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered',
-                 'workBase/build/branch/runType': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered/release',
-                 'workBase/build/branch/runType/release': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered/release/11.0.0',
-                 'workBase/': u'/Users/peter/RTTtop/RTTunittest/Work',
-                 'workBase/build': u'/Users/peter/RTTtop/RTTunittest/Work/opt'},
-    'fullPackageName': u'Simulation/Atlfast/TestAtlfast',
-    'dataSetCatalog': u'/Users/peter/RTTtop/RTTunittest/xml/rttUnitTestDataSetCatalog.xml',
-    'platform': 'slc3',
-    'refDB': u'/Users/peter/RTTtop/RTTunittest/rttUnitTestRefFile.db',
-    'resultsDirs': {'resBase/build': u'/Users/peter/RTTtop/RTTunittest/Results/opt',
-                    'resBase/build/branch/runType/release': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered/release/11.0.0',
-                    'resBase/': u'/Users/peter/RTTtop/RTTunittest/Results',
-                    'resBase/build/branch/runType': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered/release',
-                    'resBase/build/branch': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered'},
-    'releasePath': u'/afs/cern.ch/atlas/software/dist/11.0.0',
-    'build': u'opt',
-    'branch': 'Numbered',
-    'originalBranch':  u'N.X.0',
-    'runType': 'release',
-    'local':     False,
-    'sharePath': None,
-    'mailTo': '',
-    'libToolsSharePath': u'/afs/cern.ch/atlas/software/dist/11.0.0/InstallArea/share',
-    'rttProvidesConfFiles': True,
-    'containerPackage': u'Simulation/Atlfast/TestAtlfast',
-    'confFile': u'/Users/peter/RTTtop/RTTunittest/xml/TestAtlfast_TestConfiguration.xml',
-    'rttKitTest': False, 
-    'logDir': '/Users/peter/RTTtop/RTTunittest/logs',
-    'package': u'TestAtlfast',
-    #    'basePath': '/Users/peter/RTTtop/RTT/',
-    'mode': u'PBS',
-    'localPackages': [],
-    'release': u'11.0.0',
-    'cmtDir': '/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/',
-    'dirFinder': '<FindInstalledDirsMonolithic.FindInstalledDirsMonolithic instance at 0xb6ef514c>',
-    'isNightly': False
-    }
-
-refCVSPackageDict = {
-    'local': False,
-    'packageOffset': u'Simulation/Atlfast/TestAtlfast',
-    'errorMessage': '',
-    'name': u'TestAtlfast',
-    'pathToXMLfile': u'/unix/atlas/RTTunittest/xml/TestAtlfast_TestConfiguration.xml',
-    'packageVersion': 'TestAtlfast-00-00-55',
-    'packageCmtPath': '/afs/cern.ch/atlas/software/dist/11.0.0/Simulation/Atlfast/TestAtlfast/TestAtlfast-00-00-55/'}
-
-refUserConfigData = {
-    'UnifiedConfiguration': ['<CVSPackage.CVSPackage instance at 0x748d78>']
-    }
-refConfigDict = {
-    'libToolsLoc': 'releaseSharePath',    
-    'packagesVeto': [],
-    'refRelease': u'11.0.0',
-    'site': u'CERN',
-    'localConfFiles': False,
-    'installedConfFiles': False,
-    'jobGroupConfig': u'/Users/peter/RTTtop/RTTunittest/xml/rttUnitTestTopLevelJobGroup.xml',
-    'testMode': 'Off',
-    'refDBenable': u'1',
-    'versionDir': None,
-    'resultsBasePath': u'/Users/peter/RTTtop/RTTunittest/Results',
-    'dataSetCatalog': u'/Users/peter/RTTtop/RTTunittest/xml/rttUnitTestDataSetCatalog.xml',
-    'refDB': u'/Users/peter/RTTtop/RTTunittest/rttUnitTestRefFile.db',
-    'build': u'opt',
-    'branch': u'N.X.0',
-    'maxTime': 3600,
-    'rttProvidesConfFiles': True,
-    'rttKitTest': False,
-    'workBasePath': u'/Users/peter/RTTtop/RTTunittest/Work',
-    'webpageBranchVeto': [],
-    'confFiles': {u'AthenaAtlfast': u'/Users/peter/RTTtop/RTTunittest/xml/TestAtlfast_TestConfiguration.xml'},
-    'packagesAccept': [],
-    'mode': u'PBS',
-    'localPackages': [],
-    'release': u'11.0.0',
-    'releaseType': 'monolithic',
-    'confFileSuffix': '_TestConfiguration'  
-    }
-refAthenaDescriptorDict ={
-    'DC1dataSet': '',
-    'installPath': None,
-    'refRelease': u'11.0.0',
-    'fullPackagePath': '',
-    'actions': [],
-    'modeFromXML': False,
-    'binRefFiles': [],
-    'elog': 'Job2_elog',
-    'furtherJobOptions': [],
-    'paths': '<Paths.Paths instance at 0x72da80>',
-    'keepFilePatterns': [],
-    'log': 'Job2_log',
-    'pScripts': [],
-    'jobOptions': '',
-    'jobGroup': 'AthenaAtlfast',
-    'hasDataSet': False,
-    'identifiedName': '',
-    'suggestedQueue': '',
-    'tests': [],
-    'reporter': '<Reporter.LocalReporter instance at 0x733cd8>',
-    'storeInDBPatterns': [],
-    'jobSerialNumber': 2,
-    'auxFilePatterns': [],
-    'DC2dataSet': [],
-    'runPath': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered/release/11.0.0/Job/0',
-    'BSdataSet': '',
-    'isDC1data': False,
-    'name': '',
-    'package': '',
-    'asciiRefFiles': [],
-    'resPath': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered/release/11.0.0/Job/0',
-    'rttKitTestJob': False,
-    'mode': u'PBS',
-    'noNightly':     False,
-    'isBSdata': False,
-    'confederation': None,
-    'logger':  '<logging.Logger instance at 0xb71b09ac>',
-    'commandLineFlags': ''
-    }
-
-
-refNoPackagePathsDict= {
-    'libToolsLoc': 'releaseSharePath',
-    'rttTagBeingRun':'RunTimeTester-00-00-78',
-    'kitID': '',
-    'refRelease': u'11.0.0',
-    'distArea': '/afs/cern.ch/atlas/software/dist',
-    'site': u'CERN',
-    'localConfFiles': False,
-    'installedConfFiles': False,
-    'testMode': 'Off',
-    'refDBenable': u'1',
-    'branchesNotToDisplay': [],
-    'versionDir': None,
-    'releaseType': 'monolithic',
-    'workDirs': {'workBase/build/branch': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered',
-                 'workBase/build/branch/runType': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered/release',
-                 'workBase/build/branch/runType/release': u'/Users/peter/RTTtop/RTTunittest/Work/opt/Numbered/release/11.0.0',
-                 'workBase/': u'/Users/peter/RTTtop/RTTunittest/Work',
-                 'workBase/build': u'/Users/peter/RTTtop/RTTunittest/Work/opt'},
-    'dataSetCatalog': u'/Users/peter/RTTtop/RTTunittest/xml/rttUnitTestDataSetCatalog.xml',
-    'platform': 'slc3',
-    'refDB': u'/Users/peter/RTTtop/RTTunittest/rttUnitTestRefFile.db',
-    'resultsDirs': {'resBase/build': u'/Users/peter/RTTtop/RTTunittest/Results/opt',
-                    'resBase/build/branch/runType/release': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered/release/11.0.0',
-                    'resBase/': u'/Users/peter/RTTtop/RTTunittest/Results',
-                    'resBase/build/branch/runType': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered/release',
-                    'resBase/build/branch': u'/Users/peter/RTTtop/RTTunittest/Results/opt/Numbered'
-                    },
-    'releasePath': u'/afs/cern.ch/atlas/software/dist/11.0.0',
-    'build': u'opt',
-    'branch': 'Numbered',
-    'originalBranch':  u'N.X.0',
-    'runType': 'release',
-    'rttProvidesConfFiles': True,
-    'rttKitTest': False,
-    'logDir': '/Users/peter/RTTtop/RTTunittest/logs',
-    'mode': u'PBS',
-    'localPackages': [],
-    'release': u'11.0.0',
-    'isNightly': False
-}
diff --git a/Tools/RunTimeTester/src/TestSuiteRun.py.obsolete b/Tools/RunTimeTester/src/TestSuiteRun.py.obsolete
deleted file mode 100755
index f675fd2faf24f80ead453a516ceeb23b4c4200b5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteRun.py.obsolete
+++ /dev/null
@@ -1,238 +0,0 @@
-"""
-PS 31/10/05
-
-Script to test the creation of RTT objects.
-At the moment the only test is whether the creation succeeds.
-Possible extension is to add checks on the objects.
-
-Object creation is performed through zero argument factory fucntions,
-so the module can be used as a convenient way of getting hold of an
-RTT object that would otherwise be complex to contruct.
-
-The configuration information is hardwired and global, but can be overridden
-with data read from a configuration file.
-"""
-import os, time, string
-from xml.sax.handler import ContentHandler
-import xml.sax, sys
-
-from Tools   import mkdirIfNew
-
-import string, shutil
-
-import logging
-from LoggerSetup import setupLoggerForTestSuite
-
-from TestSuiteTests    import TestSuiteTests
-from TestSuiteGlobals  import getGlobalStats
-
-
-class TestSuiteRun:
-
-    def __init__(self, startPath, fixtureDir):
-    
-        print '\n\n ---TestSuiteRun starts %s ----\n' % time.ctime(time.time())
-        testSuiteScratchPath = os.path.join(startPath, '../TestSuiteScratch')
-    
-
-        if not os.path.exists(testSuiteScratchPath):
-            os.mkdir(testSuiteScratchPath)
-        
-        destPath = os.path.join(testSuiteScratchPath, fixtureDir)
-        srcPath  = os.path.join(startPath, fixtureDir)
-        if os.path.exists(destPath):shutil.rmtree(destPath)
-        os.makedirs(destPath)
-        # print 'made destPath',destPath,os.path.exists(destPath)
-        self.fixtures =  self.setupFixtures(srcPath, destPath)
-        logDir = os.path.join(destPath, self.fixtures['testLogDir'])
-        
-
-        self.handler = setupLoggerForTestSuite(logDir)
-        logger = logging.getLogger('rttTestSuite')
-        
-        # # print 'handlers %s' % str(logger.handlers)
-        # if len(logger.handlers) ==2:
-        #    # for some reason this is a logging.StreamHandler
-        #    logger.removeHandler(logger.handlers[0]) 
-        #    handler = logger.handlers[0]
-        #    # print 'TestSuite: handler = ',handler
-
-
-        self.logger = logger
-
-        self.passed = self.execute()
-        # self.passed = True # for debug !!
-        # self.passed = False # for debug !!
-        print '\n\n ---TestSuiteRun ends %s ----\n' % time.ctime(time.time())
-        
-    def setupFixtures(self, srcPath, destPath):
-
-        mkdirIfNew(os.path.join(destPath, 'logs'))
-        mkdirIfNew(os.path.join(destPath, 'logs/Tester'))
-        mkdirIfNew(os.path.join(destPath, 'Install'))
-        mkdirIfNew(os.path.join(destPath, 'Install/jobOptions'))
-        mkdirIfNew(os.path.join(destPath,
-                            'Install/jobOptions/TestAtlfast'))
-        mkdirIfNew(os.path.join(destPath, 'Work'))
-        mkdirIfNew(os.path.join(destPath, 'Work/cmtHome'))
-        mkdirIfNew(os.path.join(destPath, 'Work/cmtProject'))
-        mkdirIfNew(os.path.join(destPath,
-                                'Results/opt/N.0.X/release/atlrel_0'))
-        mkdirIfNew(os.path.join(destPath,
-                                'Results/opt/N.0.X/release/atlrel_1'))
-        mkdirIfNew(os.path.join(destPath, 'xml'))
-        mkdirIfNew(os.path.join(destPath, 'cfg'))
-        mkdirIfNew(os.path.join(destPath,
-                                'opt/N.0.X/release/atlrel_2'))
-        mkdirIfNew(os.path.join(destPath, 'opt/N.0.X/kit'))
- 
-        toPreprocess = [
-            'cfg/rttUnitTests.xml',
-            ]
-
-        toCopy =[
-            'status.txt',
-            'Readme.txt',
-            'xml/rttUnitTestTopLevelJobGroup.xml',
-            'xml/TestAtlfast_TestConfiguration.xml',
-            'xml/RunTimeTester_TestConfiguration.xml',
-            'xml/rttUnitTestDataSetCatalog.xml',
-            'Install/jobOptions/TestAtlfast/Atlfast_POOL_False_low_cone_cells.py'
-            ]
-        toCopy2 ={
-            'status.txt': ['Results/opt/N.0.X/release/atlrel_0',
-                           'Results/opt/N.0.X/release/atlrel_1']
-            }
-        [self.preprocessConfigFile(srcPath, destPath, f) for f in toPreprocess]
-        [shutil.copy(
-            os.path.join(srcPath,file),
-            os.path.join(destPath, file)
-            ) for file in toCopy]
-
-
-        for k in toCopy2.keys():
-            destDirs = toCopy2[k]
-            [shutil.copy(os.path.join(srcPath, k),
-                         os.path.join(destPath, dir, k))
-             for dir in destDirs]
-            
-            fixtures = self.getFixtures(destPath)
-            fixtures['testSuiteScratchPath'] = destPath
-            return fixtures
-    def writeline(self, file, line): file.write(line+'\n')
-    # -------------------------------------------------------------------------
-    def getFixtures(self, basePath):
-
-        testTopLevelRTTFile  = os.path.join(basePath,'cfg/rttUnitTests.xml')
-        testTopLevelJobGroup = os.path.join(basePath,'xml/rttUnitTestTopLevelJobGroup.xml')
-        
-        testRunTimerTime     = 5
-        testRTTtagBeingTested= 'NoTag' 
-        testInstallArea      = os.path.join(basePath, 'Install')
-        testPackage          = 'TestAtlfast'
-        
-        testPackageConfFile  = os.path.join(basePath,'xml/TestAtlfast_TestConfiguration.xml')
-        testLogDir           = os.path.join(basePath, 'logs')
-        testTesterLogDir     = os.path.join(basePath, 'logs/Tester')
-        testFixtureDoc       = os.path.join(basePath, 'Readme.txt')
-        
-        fixtures = {
-            'testTopLevelRTTFile':   testTopLevelRTTFile,
-            'testTopLevelJobGroup':  testTopLevelJobGroup,
-            'testRunTimerTime':      testRunTimerTime,
-            'testRTTtagBeingTested': testRTTtagBeingTested,
-            'testInstallArea':       testInstallArea,
-            'testPackage':           testPackage,
-            'testPackageConfFile':   testPackageConfFile,
-            'testLogDir':            testLogDir,
-            'testTesterLogDir':      testTesterLogDir,
-            'testFixtureDoc':        testFixtureDoc
-            }
-        self.fixturesAsStr = self.fixturesToStr(basePath, fixtures)
-        print self.fixturesAsStr
-
-        tlcf = fixtures['testTopLevelRTTFile']
-        print 'Contents of top level configuration file %s:' % tlcf
-        #print open(fixtures['testTopLevelRTTFile'], 'r').read()
-        self.stripXML(tlcf)
-        print
-        
-        return fixtures
-    # ----------------------------------------------------------------
-    def fixturesToStr(self, basePath, fixtures):
-        text =  '\nAll paths relative to %s\n' % basePath
-        
-        for i in fixtures.items():
-            var = i[0]
-            val = str(i[1])
-            # if val.find(basePath) != -1: val = val[val.rfind(basePath), -1]
-            if val.find(basePath) != -1: val = val[len(basePath):]
-            text += '%30s          %40s\n' % (string.ljust(var, 30), val)
-            
-            
-
-        docFile = fixtures['testFixtureDoc']
-        text +=  '\n\nFixture Documentation:\n'
-    
-        if os.path.exists(docFile):
-            text += open(docFile, 'r').read()
-            text += '\n\n'
-        else:
-            text +=  'No documentation file for these fixtures\n\n'
-
-        return text
-    # ----------------------------------------------------------------
-
-
-    def preprocessConfigFile(self, srcDir, destDir, file):
-
-        ifile = open(os.path.join(srcDir, file), 'r')
-        srcLines  = ifile.readlines()
-        ifile.close()
-        newSrcLines= [string.replace(l, 'TESTSUITESCRATCH',
-                                     destDir) for l in srcLines] 
-
-        dest      = open(os.path.join(destDir, file), 'w')
-        [self.writeline(dest, l) for l in newSrcLines]
-    
-    def execute(self):
-        # functionRunner = FunctionRunner(TestSuiteCoreTests.getTests())
-        # functionRunner.runNewFunctions(TestSuiteUserTests.getTests())
-        # print functionRunner
-        tests = TestSuiteTests(self.fixtures,
-                               self.logger,
-                               self.handler)
-        tests.runNonCoreTests()
-        self.testResults = str(tests)
-        self.summary     = str(getGlobalStats())
-        print self.testResults
-        print self.summary
-        return tests.allTestsPassed()
-
-    def stripXML(self, file):
-        
-        class TextHandler(ContentHandler):
-            def __init__(self):
-                self.text   = ''
-            def characters(self, ch):
-                ch = ch.strip()
-                if len(ch)==0: return
-                if len(ch)>60: ch=ch[len(ch)-60:]
-                self.text += '                 '+ch+'\n'
-                
-            def startElement(self, name, attr):
-                tag = name.strip()+':\n'
-                self.text   += tag
-                # self.indent += self.delta
-                
-                
-        parser = xml.sax.make_parser()
-        handler = TextHandler()
-        parser.setContentHandler(handler)
-        parser.parse(file)
-        print handler.text
-# -------------------------------------------------------------------------
-if __name__ == '__main__':
-    here = os.getcwd()
-    tsr = TestSuiteRun(here, 'TestSuiteFixtures/Fixtures3')
-    print 'result = ', tsr.passed
diff --git a/Tools/RunTimeTester/src/TestSuiteTests.py.obsolete b/Tools/RunTimeTester/src/TestSuiteTests.py.obsolete
deleted file mode 100755
index 042c86f70d55631ef616568af4d111e1b91f9fff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TestSuiteTests.py.obsolete
+++ /dev/null
@@ -1,914 +0,0 @@
-from TestSuiteChecker     import Checker
-from TestSuiteGlobals     import *
-
-from RunTimer             import RunTimer
-from RTTConfigParser      import RTTConfiguration
-from Paths                import NoPackagePaths, Paths
-from Paths                import PathsForTestRun, PathsForDescriptors
-from GetUserConfigData    import GetUserConfigData
-from DataSetCatalog       import DataSetCatalog
-from JobGroupKits         import JobGroupKits
-from Tools                import unique
-from exc2string           import exc2string
-
-
-
-# ----- RTT SRC IMPORTS -----
-
-from TestRun                    import TestRun, UnifiedConfigurationTestRun
-from SuperParser                import SuperParser
-from Factory_CmtLines           import CmtLinesFactory
-from Factory_Commander          import CommanderFactory
-from Factory_DirectoryMaker     import DirectoryMakerFactory
-from Factory_JobDescriptor      import DescriptorFactory
-from Factory_JobMaker           import JobMakerFactory
-from Factory_Minder             import MinderFactory
-from Factory_ScriptWriter       import ScriptWriterFactory
-from Factory_TestRun            import TestRunFactory
-from Factory_SiteSpecifics      import SiteSpecificsFactory
-from Factory_RunTimer           import RunTimerFactory
-from Factory_UserStuffRetriever import UserStuffRetrieverFactory
-# from Factory_FindInstalledDirs  import FindInstalledDirsFactory
-from Factory_HardCoded          import HardCodedFactory
-from Factory_Legals             import LegalsFactory
-from Factory_PathNameBuilder    import PathNameBuilderFactory
-from Tester                     import Tester
-from validateXMLFile            import validateXMLFile    
-from NICOS                      import NICOS
-from RTTDateTime                import RTTDateTime
-from requirementsWriter         import requirementsWriter
-from ModuleLoader               import ModuleLoader
-
-import Geneology
-import Tools
-import Tools2
-import JobGroupsParser
-import HTMLWriter
-
-# ----- TEST SUITE IMPORTS -----
-from TestSuiteRefDicts       import *
-from TestSuiteGlobals        import *
-from TestSuiteChecker        import Checker
-# from TestSuiteLogging        import loggerOff, loggerOn
-# ----- PYTHON IMPORTS -----
-
-import time, sys, os, copy
-import sys, xml.dom.minidom
-
-# ==================================================================
-
-from TestSuiteGlobals import getGlobalStats
-globalStats = getGlobalStats()
-
-class TestSuiteTests:
-    def  __init__(self, fixtures, logger, handler):
-
-        self.testTopLevelRTTFile   = fixtures['testTopLevelRTTFile']
-        self.testTopLevelJobGroup  = fixtures['testTopLevelJobGroup']
-        
-        self.testRunTimerTime      = fixtures['testRunTimerTime']
-        self.testRTTtagBeingTested = fixtures['testRTTtagBeingTested']
-        self.testInstallArea       = fixtures['testInstallArea']
-        self.testPackage           = fixtures['testPackage']
-        
-        self.testPackageConfFile   = fixtures['testPackageConfFile']
-        self.testLogDir            = fixtures['testLogDir']
-        self.testTesterLogDir      = fixtures['testTesterLogDir']
-        self.testSuiteScratchPath  = fixtures['testSuiteScratchPath']
-        
-        coreTests = [
-            ]
-
-
-
-        nonCoreTests = [
-            self.testTimer,
-            self.testRTTConfiguration,
-            self.testNoPackagePaths,
-            self.testGetUserConfigData,
-            self.testCVSPackage,
-            self.testPaths,
-            self.testDataSetCatalog,
-            self.testJobGroupKits,
-            self.testPathsForTestRun,
-            self.testPathsForDescriptors,
-            self.testAthenaJobDescriptor,
-            self.testJobGroupsParser,
-            self.testSuperParser,
-            self.testDescriptorFactory,
-            self.testDirectoryMakerFactory,
-            self.testScriptWriterFactory,
-            self.testTestRunFactory,
-            self.testJobMakerFactory,
-            self.testCmtLinesFactory,
-            self.testCommanderFactory,
-            self.testMinderFactory,
-            self.testSiteSpecificsFactory,
-            self.testRunTimerFactory,
-            self.testUserStuffRetrieverFactory,
-            self.testUserStuffRetriever,
-            # self.testFindInstalledDirsFactory,
-            self.testHardCodedFactory,
-            self.testHardCoded,
-            self.testLegalsFactory,
-            self.testLegals,
-            self.testNICOS,
-            self.testMindersBySuperParser,
-            self.testValidateXMLFile,
-            self.testTools,
-            self.testTools2,
-            self.testHTMLWriter,
-            self.testPage2Writer,
-            self.testRequirementsWriter,
-            self.testUserStuffRetriever,
-            self.testTester,
-            self.testPathNameBuilder,
-            self.testGeneology,
-            self.testModuleLoader
-            ]
-
-
-        nTests = len(coreTests)+len(nonCoreTests)
-        print 'Number of functions to test: %d' % nTests
-        # self.coreTests    = unique(coreTests)
-        self.coreTests    = coreTests
-        self.nonCoreTests = nonCoreTests
-        self.runOrder     = copy.copy(self.coreTests)
-        
-        self.runOrder.extend(self.nonCoreTests)
-
-
-        self.status = {}
-
-        self.runTests(self.coreTests)
-
-        self.cache = {}
-
-        self.logger      = logger
-        self.handler     = handler
-        self.loggerBlock = 0
-
-        globalStats.reset()
-        
-        print 'End TestSuite initialise'
-        
-    def loggerOff(self):
-        self.loggerBlock += 1
-        self.logger.removeHandler(self.handler)
-    def loggerOn(self):
-        self.loggerBlock -= 1
-        if self.loggerBlock == 0:
-            self.logger.addHandler(self.handler)    
-
-    def runNonCoreTests(self):
-        self.runTests(self.nonCoreTests)
-
-    def runTests(self, tests):
-        for t in tests:
-            try:
-                self.logger.debug('\n\n spacer line')
-                self.logger.debug( '%s start ---------------------' % t.__name__)
-                msg = t.__name__
-                status = t()
-            except Exception, e:
-                self.logger.debug( str(e) )
-                self.logger.debug( exc2string(sys.exc_info()))
-                print str(e) 
-                print exc2string(sys.exc_info())
-                status = 'exploded'
-            
-            self.status[t] = status
-            self.logger.info(status)
-            print '%45s %10s ' %(t.__name__, status)
-            self.logger.debug( '%s end   --------------------' % t.__name__)
-            # print msg, status
-
-    def allTestsPassed(self):
-        return not [f for f in self.status.keys() if (self.status[f]) == 'exploded']
-    
-    def __str__(self):
-
-        badTests = [f for f in self.status.keys() if (self.status[f])[1]< (self.status[f])[0]]
-        explodedTests = [f for f in self.status.keys() if (self.status[f]) == 'exploded']
-        str = '\n\n Test Results\n============\n\n'
-        for f in self.runOrder:
-            str+= '%35s %s\n' % (f.__name__, self.status.setdefault(f, 'notRun'))
-
-        if badTests:
-            str += '\n\n Functions that failed some tests:\n'
-            for s in badTests:
-                str += '%s\n' % s.__name__
-        if explodedTests:
-            str += '\n\n Functions that raised exceptions:\n'
-            for s in explodedTests:
-                str += '%s\n' % s.__name__
-        if not badTests:
-            if not explodedTests:
-                str += '\n\n All tests passed\n'
-        return str
-    # ==================================================================
-    
-    def fromCache(self, f):
-        result = self.cache.get(f, None)
-        if result: return result
-        result = self.cache.setdefault(f, f())
-        return result
-
-    
-    # ==================================================================
-    
-    def makeTimer(self):
-        timer = RunTimer(self.testRunTimerTime, self.logger)
-        return timer
-
-    # ==================================================================
-    
-    def testTimer(self):
-        timer =  self.fromCache(self.makeTimer)
-        checker = Checker()
-        checker('timerMax', timer.maxProcessTime)
-        checker.inRange(timer.remaining(), (0, timer.maxProcessTime))
-        time.sleep(timer.maxProcessTime+1)
-        checker.inRange(timer.remaining(), (0, timer.maxProcessTime))    
-        return checker.results()
-
-    # ==================================================================
-    
-    def makeRTTConfiguration(self):
-        rttConfig = RTTConfiguration(self.testTopLevelRTTFile)
-        return rttConfig
-
-    # ==================================================================
-    
-    def testRTTConfiguration(self):
-        rttConfig =  self.fromCache(self.makeRTTConfiguration)
-        # print '****', configDict
-        configDict = rttConfig.config
-        ignoreKeys = ['jobGroupConfig', 'resultsBasePath', 'dataSetCatalog']
-        ignoreKeys.extend(['refDB', 'workBasePath', 'confFiles'])
-        checker = Checker()
-        checker.dictCmp('refConfigDict', configDict, ignoreKeys)
-        return checker.results()
-    
-    # ==================================================================
-
-    def makeNoPackagePaths(self):
-        self.loggerOff()
-        rttConfig = self.fromCache(self.makeRTTConfiguration)
-        timer      = self.makeTimer()
-        self.loggerOn()
-        paths      = NoPackagePaths(self.testRTTtagBeingTested,
-                                    rttConfig, timer, self.testLogDir,
-                                    self.logger)
-        return paths
-    def testNoPackagePaths(self):
-        paths =  self.fromCache(self.makeNoPackagePaths)
-        checker = Checker()
-        ignore = ['workDirs', 'dataSetCatalog', 'refDB', 'resultsDirs']
-        ignore.extend(['logDir', 'rttTagBeingRun', 'hardCoded'])
-        checker.dictCmp('refNoPackagePathsDict', paths.__dict__, ignore)
-        return checker.results()
-    # ==================================================================
-    def makeGetUserConfigData(self):
-        self.loggerOff()
-        rttConfig = self.fromCache(self.makeRTTConfiguration)
-        paths      = self.fromCache(self.makeNoPackagePaths)
-        self.loggerOn()
-        return GetUserConfigData(rttConfig, paths)
-    def testGetUserConfigData(self):
-        dict = self.fromCache(self.makeGetUserConfigData).userConfigData()
-        checker = Checker()
-        checker.dictCmp('refUserConfigData', dict, ['UnifiedConfiguration'])
-        return checker.results()
-    # ==================================================================
-    def makeCVSPackage(self):
-        self.loggerOff()
-        ucd = self.fromCache(self.makeGetUserConfigData)
-        packages = ucd.userConfigData()
-        self.loggerOn()
-        package = packages[0]
-        return package
-    def testCVSPackage(self):
-        package = self.fromCache(self.makeCVSPackage)
-        checker = Checker()
-        ignore = ['pathToXMLfile']
-        checker.dictCmp('refCVSPackageDict', package.__dict__, ignore)
-        return checker.results()
-    # ==================================================================
-    def makePaths(self):
-        self.loggerOff()
-        rttConfig  = self.fromCache(self.makeRTTConfiguration)
-        timer      = self.fromCache(self.makeTimer)
-        package    = self.fromCache(self.makeCVSPackage)
-        userConfig = self.fromCache(self.makeGetUserConfigData)
-        self.loggerOn()
-        paths =  Paths(rttConfig,
-                       self.testRTTtagBeingTested,
-                       timer,
-                       self.testLogDir,
-                       package,
-                       userConfig.packageInfo('RunTimeTester'),
-                       self.logger)
-        # start kludge: these areas are not set in Paths because
-        # the real existence of the shareArea and installArea is tested
-        # before setting. Force them to junk directories for testing.
-        paths.shareArea   = self.testInstallArea
-        paths.installArea = self.testInstallArea
-        # end kludge
-        return paths
-    def testPaths(self):
-        paths =  self.fromCache(self.makePaths)
-        ignore = ['distArea', 'workDirs', 'dataSetCatalog', 'refDB',
-                  'resultsDirts', 'resultsBasePath', 'confFile', 'logDir',
-                  'installArea', 'shareArea', 'resultsDirs', 'hardCoded',
-                  'rttLibToolsPath','rttTagBeingRun', 'dirFinder']
-        checker = Checker()
-        checker.dictCmp('refPathsDict', paths.__dict__, ignore)
-        return checker.results()
-    # ==================================================================
-    def makePathsForTestRun(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        nPaths = PathsForTestRun(paths)
-        return nPaths
-    # ==================================================================
-    def testPathsForTestRun(self):
-        self.fromCache(self.makePathsForTestRun)
-        return Checker().results()
-    # ==================================================================
-    def makePathsForDescriptors(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        nPaths = PathsForDescriptors(paths)
-        return nPaths
-    # ==================================================================
-    def testPathsForDescriptors(self):
-        self.fromCache(self.makePathsForDescriptors)
-        return Checker().results()
-    # ==================================================================
-    def makeDataSetCatalog(self):
-        paths = self.fromCache(self.makePaths)
-        return DataSetCatalog(self.logger, paths)
-    # ==================================================================
-    def testDataSetCatalog(self):
-        catalog = self.fromCache(self.makeDataSetCatalog)
-        checker = Checker()
-        checker('refDataSetLen', len(catalog.catalog.keys()))
-        return checker.results()
-    # ==================================================================
-    def makeJobGroupKits(self):
-        self.loggerOff()
-        rttKits  = JobGroupsParser.parseJobGroups(self.testTopLevelJobGroup)
-        userKits = JobGroupsParser.parseJobGroups(self.testPackageConfFile)
-        self.loggerOn()
-        JobGroupKits(rttKits, userKits)
-    def testJobGroupKits(self):
-        self.fromCache(self.makeJobGroupKits)
-        jobGroupKits = JobGroupKits()
-        checker = Checker()
-        checker('refNKits', len(jobGroupKits.getAllKits()))
-        return checker.results()
-    # ==================================================================
-    def makeAthenaJobDescriptor(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        desc = self.fromCache(
-            self.makeDescriptorFactory).create('Athena',paths)
-        desc.setGroup('AthenaAtlfast')
-        desc.package = 'TestAtlfast'
-        desc.setJobOptions('TestAtlfastOptions.py')
-        return desc
-    def testAthenaJobDescriptor(self):
-        desc =  self.fromCache(self.makeAthenaJobDescriptor)
-        checker = Checker()
-        dict = desc.__dict__
-        # remove paths (platform dependant) and logs (job creation order dependant)
-        ignoreKeys = ['paths', 'reporter', 'runPath', 'installArea', 'resPath',
-                      'elog', 'log', 'jobSerialNumber', 'logger']
-        checker.dictCmp ('refAthenaDescriptorDict',dict, ignoreKeys)
-        return checker.results()
-    # ==================================================================
-    def makeJobGroupsParser(self):
-        kits = JobGroupsParser.parseJobGroups(
-            self.testPackageConfFile)
-        return kits
-    def testJobGroupsParser(self):
-        kits = self.fromCache(self.makeJobGroupsParser)
-        return Checker().results()
-    # ==================================================================
-    def makeSuperParser(self):
-        sp = SuperParser(self.testPackageConfFile, self.logger)
-        return sp
-    def testSuperParser(self):
-        superParser = self.fromCache(self.makeSuperParser)
-        return Checker().results()
-    # ==================================================================
-    def makeMindersBySuperParser(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        sp    = self.fromCache(self.makeSuperParser)
-        self.loggerOn()
-        minders = sp.parse(paths)
-        return minders
-    def testMindersBySuperParser(self):
-        minders = self.fromCache(self.makeMindersBySuperParser)
-        checker = Checker()
-        checker('nMindersBySuperParser',          len(minders))
-        if minders:
-            m = minders[0]
-            checker('nActionsBySuperParser',          len(m.actions))
-            checker('nUserDefinedTestsBySuperParser', len(m.userDefinedTests))
-            checker('nStringsToFilesBySuperParser',   len(m.stringsToFiles))
-            checker('nAsciiRefFilesBySuperParser',    len(m.asciiRefFiles))
-            checker('nBinRefFilesBySuperParser',      len(m.binRefFiles))
-            checker('nChecksBySuperParser',           len(m.checks))
-            checker('nKeepFilePatternsBySuperParser', len(m.keepFilePatterns))
-            checker('nPostScriptsBySuperParser',      len(m.postScripts))
-            checker('nAuxFilePatternsBySuperParser',  len(m.auxFilePatterns))
-            
-            checker('jobWeightBySuperParser',         m.jobWeight)
-            
-        return checker.results()
-    # ==================================================================
-    def makeDescriptorFactory(self):
-        factory = DescriptorFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testDescriptorFactory(self):
-        factory = self.fromCache(self.makeDescriptorFactory)
-        self.loggerOff()
-        paths = self.fromCache(self.makePathsForDescriptors)
-        self.loggerOn()
-        [factory.create(name, paths) for name in factory.keys()] 
-        
-        return Checker().results()
-    # ==================================================================
-    def makeDirectoryMakerFactory(self):
-        factory = DirectoryMakerFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testDirectoryMakerFactory(self):
-        factory = self.fromCache(self.makeDirectoryMakerFactory)
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        dirMaker = factory.create(paths)
-        minder = self.fromCache(self.makeMinder)
-        dirMaker.makeJobGroupDirs(minder)
-            
-        
-        return Checker().results()
-    # ==================================================================
-    def makeUserStuffRetrieverFactory(self):
-        factory = UserStuffRetrieverFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testUserStuffRetrieverFactory(self):
-        factory = self.fromCache(self.makeUserStuffRetrieverFactory)
-        return Checker().results()
-    # ----------------------------------------------------------------
-    def makeUserStuffRetriever(self):
-        self.loggerOff()
-        factory = self.fromCache(self.makeUserStuffRetrieverFactory)
-        paths      = self.fromCache(self.makeNoPackagePaths)
-        rttConfig  = self.fromCache(self.makeRTTConfiguration)
-        substring  = '_TestConfiguration'
-        legals = self.fromCache(self.makeLegals)
-        self.loggerOn()
-
-        obj = factory.create(paths.releaseType,
-                             paths,
-                             substring,
-                             rttConfig)
-        
-        return obj
-    # ----------------------------------------------------------------
-    def testUserStuffRetriever(self):
-        usr = self.fromCache(self.makeUserStuffRetriever)
-        usr.userStuff()
-                
-        return Checker().results()
-    # ==================================================================
-    # def makeFindInstalledDirsFactory(self):
-    #     factory = FindInstalledDirsFactory(logger)
-    #     return factory
-    # ----------------------------------------------------------------
-    # def testFindInstalledDirsFactory(self):
-    #    factory = self.fromCache(self.makeFindInstalledDirsFactory)
-    #    self.loggerOff()
-    #    paths      = self.fromCache(self.makeNoPackagePaths)
-    #    self.loggerOn()
-    #    releasePath = paths.releasePath 
-    #    runType     = paths.runType 
-    #    
-    #
-    #   locals = [True, False]
-    #    findInstalleds = []
-    #    for mode in factory.keys():
-    #        for local in locals:
-    #            findInstalleds.append(factory.create(mode,
-    #                                                 local,
-    #                                                 releasePath,
-    #                                                 runType))   
-    #        [f.findOptionsPaths() for f in findInstalleds]
-    #    [f.findSharePaths()   for f in findInstalleds]
-    #   
-    #    return Checker().results()
-    # ==================================================================
-    def makeHardCodedFactory(self):
-        factory = HardCodedFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testHardCodedFactory(self):
-        factory = self.fromCache(self.makeHardCodedFactory)                
-        return Checker().results()
-    # ==================================================================
-    def makeHardCoded(self):
-        factory = HardCodedFactory(self.logger)
-        paths   = self.fromCache(self.makePaths)
-        return factory.create(paths.releaseType, paths.runType)
-    # ----------------------------------------------------------------
-    def testHardCoded(self):
-        checker = Checker()
-        hardCoded = self.fromCache(self.makeHardCoded)
-        hardCoded.platform()
-        hardCoded.installAreaDirName()
-        hardCoded.AFS_release_string()
-        hardCoded.AFS_kit_string()
-        hardCoded.platform()
-
-        hardCoded.distArea()
-        checker.assertTrue(os.path.exists(hardCoded.distArea()))
-        
-        legals   = self.fromCache(self.makeLegals)
-        branches = legals.vals('branch')
-        dirs = [hardCoded.nightlyStamp(b) for b in branches]
-        [checker.assertTrue(os.path.exists(d)) for d in dirs]
-
-        [hardCoded.branchToWebPageText(b) for b in branches]
-
-        
-        for build in legals.vals('build'):
-            [hardCoded.nicosNightlyWebPageURL((b, build)) for b in branches]
-                
-        for release in legals.vals('releases'):
-            for build in legals.vals('build'):
-                hardCoded.nicosNumberedWebPageURL(release, build)
-        
-                
-        return checker.results()
-    # ==================================================================
-    def makeLegalsFactory(self):
-        factory = LegalsFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testLegalsFactory(self):
-        factory = self.fromCache(self.makeLegalsFactory)
-                
-        return Checker().results()
-    # ==================================================================
-    def makeLegals(self):
-        self.loggerOff()
-        factory     = LegalsFactory(self.logger)
-        paths       = self.fromCache(self.makePaths)
-        hardCoded   = self.fromCache(self.makeHardCoded)
-        self.loggerOn()
-        return factory.create(paths.releaseType,
-                              paths.originalBranch,
-                              hardCoded)
-    # ----------------------------------------------------------------
-    def testLegals(self):
-        self.fromCache(self.makeLegals)
-        return Checker().results()
-    # ==================================================================
-    def makeNICOS(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        nicos = NICOS(paths)
-        return nicos
-    # ----------------------------------------------------------------
-    def testNICOS(self):
-        nicos = self.fromCache(self.makeNICOS)
-        nicosDate = nicos.getReleaseDate()
-        nicos.isLatestNightly()
-
-        date = RTTDateTime(time.time())
-        nicos.toNICOSdateLink(date)
-
-        return Checker().results()
-    # ==================================================================
-    def makeScriptWriterFactory(self):
-        factory = ScriptWriterFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testScriptWriterFactory(self):
-        factory = self.fromCache(self.makeScriptWriterFactory)
-        self.loggerOff()
-        descFactory  = self.fromCache(self.makeDescriptorFactory)
-        self.loggerOn()
-        desc = descFactory.create(
-            'Athena',
-            self.fromCache(self.makePathsForDescriptors))
-        for name in factory.keys():
-            desc.setGroup(name)
-            factory.create(desc)
-        
-        return Checker().results()
-    # ==================================================================
-    def makeTestRunFactory(self):
-        factory = TestRunFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testTestRunFactory(self):
-        factory = self.fromCache(self.makeTestRunFactory)
-        self.loggerOff()
-        paths           = self.fromCache(self.makePaths)
-        pathsForTestRun = self.fromCache(self.makePathsForTestRun)
-        # dm              = self.fromCache(self.makeDirectoryMakerFactory
-        #                                  ).create(paths)
-        timer           = self.fromCache(self.makeTimer)
-        p2Writer        = self.fromCache(self.makePage2Writer)
-        self.loggerOn()
-        for name in factory.keys():
-            tr = factory.create(name,
-                                pathsForTestRun,
-        #                        dm,
-                                timer,
-                                p2Writer)
-
-            # if tr == None:
-            #    raise "TestRunFactory returned NoneType"
-        
-        
-        return Checker().results()
-    # ==================================================================
-    def makeJobMakerFactory(self):
-        factory = JobMakerFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testJobMakerFactory(self):
-        factory = self.fromCache(self.makeJobMakerFactory)
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        dom = xml.dom.minidom.parse(self.testPackageConfFile)
-        self.loggerOn()
-        jobListElement = dom.getElementsByTagName('jobList')[0]
-
-        jobMaker = factory.create('UnifiedConfiguration',
-                                  jobListElement,
-                                  paths)
-
-        # return jobMaker
-        return Checker().results()
-    # ==================================================================
-    def makeCmtLinesFactory(self):
-        factory = CmtLinesFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testCmtLinesFactory(self):
-        fctry = self.fromCache(self.makeCmtLinesFactory)
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        desc = self.fromCache(self.makeAthenaJobDescriptor)
-        legals = self.fromCache(self.makeLegals)
-        self.loggerOn()
-        
-        cmtLineObj = fctry.create(paths.releaseType)
-        cmtLineObj(desc)
-        
-        return Checker().results()
-    
-    # ==================================================================
-    def makeCommanderFactory(self):
-        factory = CommanderFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testCommanderFactory(self):
-        factory = self.fromCache(self.makeCommanderFactory)
-        legals = self.fromCache(self.makeLegals)
-
-        [factory.create(mode) for mode in factory.keys()]
-        commanders = [factory.create(mode) for mode in legals.vals('mode')]
-
-        desc = self.fromCache(self.makeAthenaJobDescriptor)
-        [c.submitCommand(desc) for c in commanders]
-        return Checker().results()
-    
-    # ==================================================================
-    def makeMinderFactory(self):
-        factory = MinderFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def makeMinder(self):
-        factory = self.fromCache(self.makeMinderFactory)
-        self.loggerOff()
-        
-        desc = self.fromCache(self.makeAthenaJobDescriptor)
-        paths = self.fromCache(self.makePaths)
-        commander = self.fromCache(
-            self.makeCommanderFactory).create(paths.mode)
-        self.loggerOn()
-        cmd = commander.submitCommand(desc)
-
-        minder = factory.create(desc, cmd)
-
-        return minder
-    # ----------------------------------------------------------------
-    def testMinderFactory(self):
-        minder = self.fromCache(self.makeMinder)
-        return Checker().results()
-    
-    # ==================================================================
-    def makeSiteSpecificsFactory(self):
-        factory = SiteSpecificsFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testSiteSpecificsFactory(self):
-        factory = self.fromCache(self.makeSiteSpecificsFactory)
-        legals = self.fromCache(self.makeLegals)
-        [factory.create(name) for name in factory.keys()]
-        [factory.create(name) for name in legals.vals('site')]
-        
-        return Checker().results()
-    # ==================================================================
-    def makeRunTimerFactory(self):
-        factory = RunTimerFactory(self.logger)
-        return factory
-    # ----------------------------------------------------------------
-    def testRunTimerFactory(self):
-        factory = self.fromCache(self.makeRunTimerFactory)
-        legals = self.fromCache(self.makeLegals)
-
-        isNightly = True
-        [factory.create(isNightly, name) for name in factory.keys()]
-        [factory.create(isNightly, name) for name in legals.vals('branch')]
-        
-        isNightly = False
-        [factory.create(isNightly, name) for name in factory.keys()]
-        [factory.create(isNightly, name) for name in legals.vals('branch')]
-        
-        return Checker().results()
-    # ==================================================================
-    def testListDir(self):
-        pass
-    # ==================================================================
-    def makeTester(self):
-        obj = Tester(self.testRTTtagBeingTested,
-                     self.testTopLevelRTTFile,
-                     os.path.dirname(self.testTesterLogDir),
-                     self.logger,
-                     'Test')
-        return obj
-    # ----------------------------------------------------------------
-    def testTester(self):
-        self.fromCache(self.makeTester)
-        return Checker().results()
-    # ==================================================================
-    def makePathNameBuilder(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        obj = PathNameBuilderFactory(self.logger).create(paths.releaseType,
-                                                    paths.rttConfig,
-                                                    paths.hardCoded,
-                                                    paths.legals,
-                                                    paths.siteSpecificsObj)
-        return obj
-    # ----------------------------------------------------------------
-    def testPathNameBuilder(self):
-        self.fromCache(self.makePathNameBuilder)
-        return Checker().results()
-    # ==================================================================
-    def makePage2Writer(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makeNoPackagePaths)
-        self.loggerOn()
-        obj = HTMLWriter.Page2Writer(paths, self.testRTTtagBeingTested, True)
-        return obj
-    # ----------------------------------------------------------------
-    def testPage2Writer(self):
-        self.fromCache(self.makePage2Writer)
-        return Checker().results()
-    # ==================================================================
-    def makeModuleLoader(self):
-        self.loggerOff()
-        paths = self.fromCache(self.makePaths)
-        self.loggerOn()
-        pathToModule = str(paths.rttLibToolsPath)
-        obj = ModuleLoader('RttLibraryTools', pathToModule, self.logger)
-        return obj
-    # ----------------------------------------------------------------
-    def testModuleLoader(self):
-        self.fromCache(self.makeModuleLoader)
-        return Checker().results()
-    # ==================================================================
-    def testValidateXMLFile(self):
-        validateXMLFile(self.testPackageConfFile)
-        return Checker().results()
-    # ==================================================================
-    def testGeneology(self):
-        ok  = Geneology.TreeTest(self.logger)
-        checker = Checker()
-        checker.assertTrue(ok)
-        self.fromCache(self.makeJobGroupKits)
-        jobGroupKits = JobGroupKits().getAllKits()
-        Geneology.KitTest(self.logger, jobGroupKits)
-        return checker.results()
-    # ==================================================================
-    def testTools(self):
-        # pass
-        # checker = Checker()
-
-        # version = Tools.getPackageTag('./TestSuiteFixtures/Package')
-        # checker('packageVersion', version)
-
-        # dirSnapOK = Tools.dirSnapShot('.')
-        # checker('dirSnapOK', dirSnapOK)
-
-        # path = './junk.txt'
-        # f = open(path, 'w')
-        # f.write('test\n')
-        # f.close
-        # iOwn = Tools.pathIsOwnedByMe(path)
-        # checker.assertTrue(iOwn)
-        # os.remove(path)
-        
-        # path = './junk.txt'
-        # f = open(path, 'w')
-        # f.write('test\n')
-        # f.close
-        # permToSet = '0777'
-        # permSet = Tools.changePerm(path, permToSet)
-        # checker.assertTrue(permSet == permToSet)
-        # os.remove(path)
-
-        
-        
-        return Checker().results()
-    # ==================================================================
-    def testTools2(self):
-
-        self.loggerOff()
-        legals = self.fromCache(self.makeLegals)
-        self.loggerOn()
-
-        releases = legals.vals('releases')
-        [Tools2.isValidReleaseNumber(rel) for rel in releases]
-        
-        Tools2.releaseNumberComponents('11.0.0')
-        
-        Tools2.fixSortBranches(['N.0.X','10.0.X','N.X.0','Numbered'])
-        Tools2.fixSortBranches(['10.0.X','N.0.X','N.X.0','Numbered'])
-
-        Tools2.getFileOrDirTimeStamp(self.testSuiteScratchPath)
-        Tools2.getFileOrDirTimeStamp(
-            os.path.join(self.testSuiteScratchPath, 'status.txt'))
-
-        paths = self.fromCache(self.makePaths)
-        resBase = paths.resultsDirs['resBase/']
-        Tools2.getAllResultDirs(resBase, legals)
-
-        resPath = self.testSuiteScratchPath
-        job = 'TGCDigitValidationOptions_RTT'
-        Tools2.jobStatusFromFile(job, resPath)
-
-        Tools2.makeReleaseDictionary(resPath, legals)
-        
-        return Checker().results()
-
-    # ==================================================================
-
-    def testHTMLWriter(self):
-        HTMLWriter.writePage1(self.fromCache(self.makePaths))
-        htmlFile = os.path.join(self.testSuiteScratchPath,
-                             'Results',
-                             'RTTpage1.html')
-        size = -1
-        if os.path.exists(htmlFile):
-            size = os.stat(htmlFile).st_size
-
-        checker = Checker()
-        checker('RTTpage1Size', size)
-        return checker.results()
-        
-    # ==================================================================
-
-    def testRequirementsWriter(self):
-        self.loggerOff()
-        desc = self.fromCache(self.makeAthenaJobDescriptor)
-        self.loggerOn()
-        distArea = desc.paths.hardCoded.distArea()
-        release  = desc.paths.release
-        runPath  = desc.runPath
-        runType  = desc.paths.runType
-        
-        requirementsWriter(distArea,
-                           release,
-                           runPath,
-                           runType)
-
-        return Checker().results()
diff --git a/Tools/RunTimeTester/src/Tester.py b/Tools/RunTimeTester/src/Tester.py
deleted file mode 100755
index 3c887367d66d7bc5f59e2d4e7e20af2cfd71b7d6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Tester.py
+++ /dev/null
@@ -1,204 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A Giant class which does everything in its constructor !??!
-Does roughly the following:
-
-1. Reads the RTT configuration file, which is used to set up a large number
-   of paths and other variables
-
-2. Uses factory method to set up a Commander (mode specific commands)
-
-3. Deletes existing  run directories
-
-4. Reads the Jobs configuration file - this sets up a list of Job Descriptors
-   (one JobDescriptor per job)
-   JobDescriptors contain info about a job. They also create a run directory for
-   thir job.
-
-5. Submit the jobs. One JobMinder per job is created. JobMinders incoporate the
-   JobDescriptor, and then monitor the progress of the job.
-
-6. A poll command is periodically sent to the JobMinders. These check their job status.
-   This status is interrogated, and html pages are updated with new status info.
-
-"""
-#----- Python modules -----
-
-import os, os.path, time, sys
-import threading
-import logging
-
-#----- RTT modules -----
-
-from exc2string2                import exc2string2
-
-from setMaxMem                  import setMaxMem
-from getTimes                   import getTimes
-# -------------------------------------------------------------------------
-# from MethodTracer        import MetaTracer
-# __metaclass__ = MetaTracer
-# -------------------------------------------------------------------------
-
-
-def getHost():    
-    try:
-        host =  os.environ['HOST']
-    except:
-        try:
-            host =  os.environ['HOSTNAME']
-        except:
-            host = 'UNKNOWN'
-    return host
-
-class Tester:
-    
-    def __init__(self, argBag):
-
-        self.xmlConverter = argBag.pop('xmlConverter')
-        self.logger       = argBag.pop('logger')
-        self.startTime2   = argBag.pop('startTime2')
-        self.testRuns     = argBag.pop('testRuns')
-
-
-        self.__dict__.update(argBag.__dict__)
-         
-        self.endTime1           = 'n/a'
-        self.endTime2           = 'n/a'
-        self.statusText         = 'Initialising RTT run'
-        self.statusTextTime, j  = getTimes()
-
-
-        self.xmlConverter.setTester(self)
-
-        self.logger.info('maximum process memory set to %s' % setMaxMem)
-        self.logger.info('Running on host %s' % getHost())
-        
-        self.logger.info('---RTT initialisation phase---')
-        self.configureMonitorMinders()
-        
-
-        self.statusText = '---End of RTT initialisation phase---'
-        self.logger.info(self.statusText)
-        
-# ------------------------------------------------------------------------
-# End of __init__
-# -------------------------------------------------------------------------
-
-    def configureMonitorMinders(self):
-        """
-        Pass references to all non-RTTmonitor job minders to each  monitor
-        minder.
-        """
-        
-        # pass a list of all non-rttMonitor jobs to the monitor descriptors
-        # for monitoring
-        minders = []
-        [minders.extend(tr.jobMinders) for tr in self.testRuns]
-        monis  = [m for m in minders if m.jobGroup == "RTTMonitor"]
-        others = [m for m in minders if m.jobGroup != "RTTMonitor"]
-
-        m  = 'Number of monitoring minders '
-        m += '%d number non moni %d' %(len(monis), len(others))
-        self.logger.debug(m)
-        [moni.setJobs(others) for moni in monis]
-        
-
-    
-#-------------------------------------------------------------------------
-#----  EXECUTE -----------------------------------------------------------
-#-------------------------------------------------------------------------
-
-    def execute(self):
-        #-----------------------------------------------
-        # Loop through and execute the run method for
-        # each test run object
-        #-----------------------------------------------
-        statusText =  '---Tester  execute phase starts ---'
-        self.logger.info(statusText)
-
-        # start the xmlConverter looking for testrun reports
-        self.xmlConverter.start() # start the thread
-
-        
-        self.executeTestRuns()
-
-        # close down the XMLCOnverter
-        
-        self.logger.info('Waiting for xml converter thread to terminate')
-        self.xmlConverter.terminateThread()
-        self.xmlConverter.join()
-        self.xmlConverter.finalise()
-        statusText =  '---Tester  execute phase ends ---'
-        self.logger.info(statusText)
-        
-        #------------------------------------------
-        # Output run summariser files
-        #------------------------------------------
-        statusText =  '---Tester  monitoring output starts ---'
-        self.logger.info(statusText)
-        self.doMonitoringOutput()
-        statusText =  '---Tester  monitoring output ends ---'
-        self.logger.info(statusText)
-        
-# -------------------------------------------------------------------------
-
-    def doMonitoringOutput(self):
-        
-        # self.doRunSummariserOutput()
-
-        #------------------------------------------
-        # create Monitoring Plots
-        #------------------------------------------
-        self.doTimingOutput()
-
-# -------------------------------------------------------------------------
-
-    def executeTestRuns(self):
-
-        ntr = 0
-        for testRun in self.testRuns:
-            try:
-                
-                m = 'Starting test run %s with %d jobs' % (
-                    testRun.name,
-                    len(testRun.jobMinders)
-                    )
-                self.logger.debug(m)
-                
-                testRun.start() # start the run method in a new thread
-                self.logger.debug("Started  testrun thread %i %s" %
-                             (ntr, str(testRun.name)))
-                ntr+=1
-            except Exception, e:
-                msg  = 'TestRun %s raised exception on thread start. Traceback:\n %s\n%s' % (str(testRun.name),
-                                                                                             str(e),
-                                                                                             exc2string2())
-                self.logger.warning(msg)
-
-        # wait for al the test runs to complete
-        [self.logger.debug('Thread running for TestRun %s' % str(tr.name)) for tr in self.testRuns]
-        [tr.join() for tr in self.testRuns]
-        [self.logger.debug('Thread terminated for TestRun %s' % str(tr.name)) for tr in self.testRuns]
-
-#-------------------------------------------------------------------------
-#-------------------------------------------------------------------------
-
-    def doTimingOutput(self): [tr.printMethodTimes() for tr in self.testRuns]
-    # def doTimingOutput(self): pass
-    
-# -------------------------------------------------------------------------
-# -------------------------------------------------------------------------
-# ----  FINALISE ----------------------------------------------------------
-# -------------------------------------------------------------------------
-
-    def finalise(self):
-        self.endTime1, self.endTime2 = getTimes()
-        statusText = '---finalise - starts  ---'
-        self.logger.info(statusText)
-        statusText = '---finalise - ends  ---'
-        self.logger.info(statusText)
-        self.logger.info('RTT Tester close down...')
-        logging.shutdown()
-        # sys.exit()
-
diff --git a/Tools/RunTimeTester/src/TesterXMLConverter2.py b/Tools/RunTimeTester/src/TesterXMLConverter2.py
deleted file mode 100755
index 8ca490937723928a28ab74824832d0f4d48c587c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TesterXMLConverter2.py
+++ /dev/null
@@ -1,407 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-A class to provide an xml file summary of an RTT run
-"""
-from BaseXMLConverter2     import BaseXMLConverter
-from exc2string2           import exc2string2
-from Tools                 import changePerm
-from Evaluate              import Evaluate
-from MemoryUsageDiagnostic import MemoryUsageDiagnostic
-from memstat               import memstat
-from RTTSummaryToText      import RTTSummaryToText
-from getTimes              import getTimes
-from StringIO              import StringIO
-from xml.dom.ext           import PrettyPrint
-
-
-import os, time, commands, gc
-
-
-import threading
-
-summaryDocumentLock = threading.Lock()
-
-# from MethodTracer        import MetaTracer
-# __metaclass__ = MetaTracer
-
-
-def countOpenFiles():
-    basedir = '/proc/self/fd'
-    nfiles = len(os.listdir(basedir))
-    listfiles =  commands.getoutput('ls -o %s' %basedir)
-    return (nfiles, listfiles)
-
-
-
-class TesterXMLConverter(threading.Thread, BaseXMLConverter):
-    def __init__(self, argBag, createElement, createTextNode):
-        
-        logger = argBag.pop('logger')
-        BaseXMLConverter.__init__(self, createElement, createTextNode, logger)
-
-        threading.Thread.__init__(self)
-
-        self.node               = argBag.pop('node') # node must be passed in for this subclass
-        self.devRun             = argBag.pop('devRun')
-        self.startTime1         = argBag.pop('startTime1')
-        self.startTime2         = argBag.pop('startTime2')
-        self.nicosDateFormatted = argBag.pop('nicosDateFormatted')
-        self.nicosDate          = argBag.pop('nicosDate')
-        self.resultsDirs        = argBag.pop('resultsDirs')
-        self.RTTSrcDir          = argBag.pop('RTTSrcDir')
-        self.RTTLibDir          = argBag.pop('RTTLibDir')
-        self.topProject         = argBag.pop('topProject')
-        self.otherProject       = argBag.pop('otherProject')
-        self.cmtConfig          = argBag.pop('cmtConfig')
-        self.originalBranch     = argBag.pop('originalBranch')
-        self.site               = argBag.pop('site')
-        self.release            = argBag.pop('release')
-        self.releaseType        = argBag.pop('releaseType')
-        self.runType            = argBag.pop('runType')
-        self.legals             = argBag.pop('legals')
-        self.RTTSummary         = argBag.pop('RTTSummary')
-
-        self.__dict__.update(argBag.__dict__)
-        
-        self.testRunNodes  = {}
-        self.testRunDicts  = {}
-
-        self.nlock = 0
-
-        self.packagesToProcess = {}
-        self.packageLocks      = {}
-        self.exit              = False
-
-        self.testRunsAwaitingLock = []
-
-        self.pollSleepTime = 60
-        if self.devRun: self.pollSleepTime = 3
-        
-    def setTester(self, tester):
-        self.tester = tester
-        
-        self.initOverviewData()
-        overviewNode = self.makeOverviewNode() #None->node
-        self.node.appendChild(overviewNode)
-        
-
-        
-    def makeOverviewNode(self):
-        node = self.elementCreator('overview')
-        [self.addTextNode(node, item) for item in self.overviewData.items()]
-        return node
-    
-    def initOverviewData(self):
-
-        def getVersion(path):
-            version = 'unknown'
-            vpath = os.path.join(str(path), '../cmt/version.cmt')
-            try:
-                f = open(vpath, 'r')
-                version = f.read()
-                f.close()
-            except:
-                m='Error reading version from %s\n%s'%(vpath, exc2string2())
-                self.logger.error(m)
-            return version
-
-
-
-        self.overviewData = {
-            'startTime1':         self.startTime1,
-            'startTime2':         self.startTime2,
-            'latestUpdateTime1':  self.startTime1,
-            'latestUpdateTime2':  self.startTime2,
-            'nicosDate1':         self.nicosDateFormatted,
-            'nicosDate2':         self.nicosDate,
-            'resBaseDir':         self.resultsDirs['base'],
-            'resReleaseDir':      self.resultsDirs['packages'],
-            'RTTSrcVersion':      getVersion(self.RTTSrcDir),
-            'RTTLibVersion':      getVersion(self.RTTLibDir),
-            'endTime1':           self.tester.endTime1,
-            'endTime2':           self.tester.endTime2,
-            'nPackages':          len(self.tester.testRuns),
-            'statusText':      "%s %s" % (self.tester.statusText,
-                                          self.tester.statusTextTime),
-            'topProject':         self.topProject,
-            'otherProject':       self.otherProject,
-            'cmtConfig':          self.cmtConfig,
-            'originalBranch':     self.originalBranch,
-            'site':               self.site,
-            'release':            self.release,
-            'releaseType':        self.releaseType,
-            'runType':            self.runType,
-            'nPackagesAllJobsOK': 0
-
-            }        
-
-
-        # add the release name to the overview data
-        if self.legals.isNightlyInRelease(self.overviewData['release']):
-            self.overviewData['releaseName'] = os.path.join(self.overviewData['release'],
-                                                            self.overviewData['originalBranch'])
-        else:
-            self.overviewData['releaseName'] = self.overviewData['release']
-
-    def dynamicOverviewData(self):        
-        self.overviewData['nPackages']       = len(self.tester.testRuns)
-        self.overviewData['statusText']      = self.getOverviewStatusText()
-        self.overviewData['latestUpdateTime1'], self.overviewData['latestUpdateTime2'] = getTimes()
-
-        dicts = self.testRunDicts.values()
-
-        def getJobCounts(dict):
-            nSucc = dict['nSucc']
-            nJobs = dict['nJobs']
-            if nJobs == nSucc: return 1
-            return 0
-
-        nAllJobsOK = 0
-        for dict in dicts: nAllJobsOK += getJobCounts(dict)
-
-        self.overviewData['nPackagesAllJobsOK'] = nAllJobsOK
-
-    def getOverviewStatusText(self):        
-        dicts = self.testRunDicts.values()
-
-        def getJobCounts(dict):
-            nSucc = dict['nSucc']
-            nDone = dict['nDone']
-            nJobs = dict['nJobs']
-            return (nSucc, nDone, nJobs)
-
-        counts = [getJobCounts(dict) for dict in dicts]
-        sDone = sJobs = sSucc = 0
-        for nSucc,nDone,nJobs in counts:
-            sDone += nDone
-            sJobs += nJobs
-            sSucc += nSucc
-        
-        ftime, utime = getTimes()
-        return  "%s: jobs succ/done/tot: %i/%i/%i" % (ftime, sSucc, sDone, sJobs)
-
-    def run(self):
-
-        while True:
-
-            self.processPackages()
-
-            if self.exit:
-                self.processPackages()
-                self.logger.debug('TesterXMLConverter2 exit')
-                return
-
-            time.sleep(self.pollSleepTime)
-            
-    def processPackages(self):
-        
-        kkeys = self.packagesToProcess.keys()
-        self.logger.debug('start %d packages to process %s' % (len(kkeys),
-                                                                   str(kkeys)))
-
-        for package in kkeys:
-            try:
-                self.processPackage(package)
-                self.logger.info('Processed package %s' % (package))
-            except:
-                self.logger.error('Error processing package %s\nTraceback:\n %s' % (package, exc2string2()))
-
-                        
-        self.logger.debug('end %d packages to process %s' % (len(kkeys),
-                                                             str(kkeys)))
-
-    def processPackage(self, package):
-            
-        updatedMinders = []
-        lock = self.packageLocks[package]
-        # testruns add info to a list - need to know all updated minders, but only
-        # the last node and dictionary supersede the earlier ones.
-        
-        
-        ##################locked code###################
-        lock.acquire(True)
-        infoList = self.packagesToProcess[package]
-        self.logger.debug('package %s inflList length %d' % (package, len(infoList)))
-        [updatedMinders.extend(pUpdatedMinders) for (dict, node, pUpdatedMinders) in infoList]
-        dict, node, unusedM = infoList[-1]
-        #[node.unlink() for (dict, node, pUpdatedMinders) in infoList[:-1]]
-        del self.packagesToProcess[package]
-        lock.release()
-        ##################locked code###################
-            
-        self.updateTestRunDict(dict, node)
-        #self.updateOverview()
-        self.storeXMLData()
-
-        # ensure each minder is named only once
-        temp = {}
-        for m in updatedMinders: temp[m] = True
-        updatedMinders = temp.keys()
-        
-        # self.tester.updateMonitors(package, updatedMinders)
-
-    def update(self, info):
-        # receives information as a dict and as an XML node
-        # keep track of the current package values in dictionaries
-
-
-        # self.logger.debug('memory usage from TesterXMLConverter.update\n %s' % memstat())
-        package = info[0]['name']
-        self.logger.debug('TesterXMLConverter update called by %s START' % package )
-        
-        lock = self.packageLocks.setdefault(package, threading.Lock())
-
-
-        self.testRunsAwaitingLock.append(package)
-        self.logger.debug('start %d packages awaiting locks %s' % (len(self.testRunsAwaitingLock), str(self.testRunsAwaitingLock)))
-
-
-        # lock the code that updates the dictionaries that hold the data to process for each package.
-        # the different packages write to different dictionary entries.
-        # the lock prevents a package from updating a dictionary entry while the entry is being processed.
-        lock.acquire(True)
-        self.packagesToProcess.setdefault(package, []).append(info)
-        
-        self.testRunsAwaitingLock.remove(package)
-        self.logger.debug('end %d packages awaiting locks %s' % (len(self.testRunsAwaitingLock), str(self.testRunsAwaitingLock)))
-        
-        lock.release()
-        self.logger.debug('TesterXMLConverter update called by %s END' % package )
-        
-
-    def updateTestRunDict(self, dict, node):
-        self.logger.debug('updating for package %s ' % dict['name'])
-        pName = dict['name']
-        self.testRunNodes[pName] = node
-        self.testRunDicts[pName] = dict
-
-    def updateOverview(self):
-        self.dynamicOverviewData()
-        return self.makeOverviewNode()
-
-    def storeXMLData(self):
-        """ updates the summaryDocument (in memory)
-        updates the XML files, updatesd the monitors
-        """
-
-        #self.logger.error('xml before clear:\n%s'%self.tester.summaryDocument.toxml())
-        nover = Evaluate('overview', self.node)
-        npack = Evaluate('package', self.node)
-        self.logger.debug('No of overview, package nodes before clearDOM(): %d %d' % (len(nover),len(npack)))
-        self.clearDOM()
-        nover = Evaluate('overview', self.node)
-        npack = Evaluate('package', self.node)
-        self.logger.debug('No of overview, package nodes after  clearDOM(): %d %d' % (len(nover),len(npack)))
-        #self.logger.debug('xml after clear:\n%s'%self.tester.summaryDocument.toxml())
-
-        nodes = [self.updateOverview()]
-        nodes.extend(self.testRunNodes.values())
-        [self.node.appendChild(n) for n in nodes]
-        
-        nover = Evaluate('overview', self.node)
-        npack = Evaluate('package', self.node)
-        self.logger.debug('No of overview, package nodes after readding: %d %d' % (len(nover),len(npack)))
-
-        md = MemoryUsageDiagnostic(self.logger)
-        md.execute()
-        #  nopen, listopen = countOpenFiles()
-        # self.logger.debug('no of open files %d ' % nopen)
-        # self.logger.debug('no of open files %d\nopen files:\n%s' % countOpenFiles())
-        
-        self.doXMLDataStorage()
-
-
-
-    def doXMLDataStorage(self):
-
-
-    
-        def storeIt(fn):
-            name = fn.__name__
-            self.logger.debug('about to call %s' % name)
-            try:
-                fn()
-                return ''
-            except:
-                return exc2string2()
-
-        fns = [self.xmlToFile, self.overviewNodeToFile, self.xmlToText]
-
-        errs = [(f.__name__, storeIt(f)) for f in fns]
-
-        [self.logger.error('Error executing %s\n%s: ' % (fname,  tb))
-         for fname, tb in errs if tb != '']
-
-        
-    def finalise(self):
-        """Called from Tester.finalise(). Updates the summary document overview data with the
-        endTimes newly created in that method. Stores the new XML and calls Tester.updateMonitors().
-        """
-
-        self.processPackages()
-        endTime1, endTime2 = getTimes()
-
-        self.overviewData['endTime1']          = endTime1
-        self.overviewData['endTime2']          = endTime2
-        self.overviewData['latestUpdateTime1'] = endTime1
-        self.overviewData['latestUpdateTime2'] = endTime2
-        self.storeXMLData()
-
-    def terminateThread(self): self.exit = True
-
-    # --------------------------------------------------------------------------
-
-    def xmlToText(self):
-
-        of = open(os.path.join(self.resultsDirs['packages'], 'RTTSummary.txt'), 'w')
-        of.write( RTTSummaryToText(self.tester.summaryDocument))
-        of.close
-
-
-    def xmlToFile(self):
-        temp = StringIO() 
-        PrettyPrint(self.tester.summaryDocument.documentElement, temp)
-
-        of = open(self.RTTSummary, 'w')        
-        of.write(temp.getvalue())
-        # of.write(self.tester.summaryDocument.toprettyxml('   '))
-        # of.write(self.tester.summaryDocument.toprettyxml('   ','\n'))
-        # of.write(self.tester.summaryDocument.toxml('utf-8'))
-        of.close
-
-    def overviewNodeToFile(self):
-        of = open(os.path.join(os.path.dirname(self.RTTSummary), 'overview.xml'), 'w')
-        of.write(self.updateOverview().toprettyxml('   ','\n'))
-        of.close
-
-    def clearDOM(self):
-        
-        nodesToRemove = Evaluate('overview', self.node) # do not remove the RTTSummary/ATNSummary/overview
-        nodesToRemove.extend(self.node.getElementsByTagName('package'))
-
-        try:
-            xmlstr = self.node.toxml()
-        except:
-            self.logger.error('Clear dom 1, could not convert doc to string. self.node exists: %s' % self.node)
-        [self.node.removeChild(n) for n in nodesToRemove]
-        try:
-            xmlstr = self.node.toxml()
-        except:
-            self.logger.error('Clear dom 2, could not convert doc to string')
-        # [n.unlink() for n in nodesToRemove]
-        self.logger.debug('Garbage collection is enabled?: %s' %  str(gc.isenabled()))
-        gc.collect()
-        self.logger.debug('Garbage from clearDom: %s' % str(gc.garbage))
-        #for g in gc.garbage:
-        #    m = '\nreferents:\n%s\n\nreferrers:\n%s' % (str(gc.get_referents(g)), str(gc.get_referrers(g)))
-        #    self.logger.error(m)
-        
-        try:
-            xmlstr = self.node.toxml()
-        except:
-            self.logger.error('Clear dom 3, could not convert doc to string')
-        
-    def printMethodTimes(self):
-        print '--TesterXMLConverter--\n%s' % self.formatMethodTimes()
-                                                  
diff --git a/Tools/RunTimeTester/src/TimedDataHist.py.obsolete b/Tools/RunTimeTester/src/TimedDataHist.py.obsolete
deleted file mode 100755
index 25efebbd80788546c80423dd64709770601562e0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/TimedDataHist.py.obsolete
+++ /dev/null
@@ -1,93 +0,0 @@
-import time
-
-class TimeHist:
-    def __init__(self, nchan=10):
-        self.data = []
-        self.nchan = nchan
-        self.dist = {}
-        self.tdist = []
-        self.minVal = 0.
-        self.maxVal = 0.
-        self.step   = 1.
-        
-        [self.dist.setdefault(i,0) for i in range(self.nchan)]
-        
-    def newPoint(self, data):
-        self.data.append( (data, time.time()) )
-
-    def histVals(self):
-        self.setBinWidth()
-        self.zeroDist()
-        if len(self.data) < 2: return str
-        
-        for d in self.data:
-            ichan  = int((d[0]-self.minVal)/self.step)
-            self.dist[ichan] =  self.dist[ichan]+1
-
-    def setBinWidth(self):
-        data = [d[0] for d in self.data]        
-        self.minVal = min(data)
-        self.maxVal = max(data)
-        width = self.maxVal-self.minVal
-        self.step = width/float(self.nchan-1)
-        if self.step == 0.: self.step = 1.
-
-        print self.minVal, self.maxVal, self.step
-
-    def histTimes(self):
-        self.setBinWidth()
-        self.tdist = []
-        if len(self.data) < 2: return str
-
-        for d in self.data:
-            ichan  = int((d[0]-self.minVal)/self.step)
-            t =  time.strftime("%d %b %Y %H:%M:%S", time.gmtime(d[1])),
-            self.tdist.append( (t, ichan*'x') )
-
-    def header(self):
-        str = '\n\nmin: %s max: %s bin width: %s\n\n' % (self.minVal,
-                                                       self.maxVal,
-                                                       self.step )
-        return str
-        
-    def valueDistToString(self):
-        str = self.header()
-        if len(self.data) < 2: return str
-        
-        for i in range(self.nchan):
-            line = int(self.dist[i])*'x'
-            str +=  '%f5: %s\n' % ((self.minVal+float(i)*self.step),line)
-
-        return str
-    
-    def timeDistToString(self):
-        str = self.header()
-        if len(self.data) < 2: return str
-
-        for point in self.data:
-            ichan  = int((point[0]-self.minVal)/self.step)
-            line = ichan*'x'
-            t = time.strftime("%d %b %Y %H:%M:%S", time.gmtime(point[1]))
-            str += '%s: %s\n' % (t,line)
-
-        return str
-    def zeroDist(self):
-        for key in self.dist.keys(): self.dist[key]=0.
-        
-    def valuesPlot(self):
-        self.histVals()
-        return self.valueDistToString()
-    def timesPlot(self):
-        self.histTimes()
-        return self.timeDistToString()
-if __name__ == '__main__':
-    hist = TimeHist()
-    data = [1,2,2,3,4,5,6,7,8, 8, 8, 5, 2]
-    for d in data:
-        
-        hist.newPoint(d)
-        time.sleep(0.5)
-        
-    print hist.valuesPlot()
-    print hist.timesPlot()
-    
diff --git a/Tools/RunTimeTester/src/Timer.py b/Tools/RunTimeTester/src/Timer.py
deleted file mode 100755
index e67250c6a1b52b5b0b7c113a758c055e85d58080..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Timer.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time
-"""
-Prints the time with a particular format
-"""
-
-def timer():
-    format = "%y/%m/%d %H:%M"
-    return time.strftime(format, time.localtime())
-
diff --git a/Tools/RunTimeTester/src/ToVariableLoggingHandler.py b/Tools/RunTimeTester/src/ToVariableLoggingHandler.py
deleted file mode 100755
index 424b159f8068c69ff915f72fac2bff2abf5f4c6c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ToVariableLoggingHandler.py
+++ /dev/null
@@ -1,19 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"A logging handler to get a message from memory to a variable"
-
-import logging.handlers, copy
-
-class ToVariableLoggingHandler(logging.handlers.MemoryHandler):
-    def __init__(self):
-        logging.handlers.MemoryHandler.__init__(self, 2048)
-        self.copyTo=''
-    def handle(self, record):
-        "copy the message"
-        
-        logging.handlers.MemoryHandler.handle(self, record)
-        self.copyTo = copy.deepcopy(self.format(record))
-    def getMsg(self):
-        " return the message in demand"
-        
-        return self.copyTo
diff --git a/Tools/RunTimeTester/src/ToVariableLoggingHandler.py.obsolete b/Tools/RunTimeTester/src/ToVariableLoggingHandler.py.obsolete
deleted file mode 100755
index ccb0fa2f1efe25e9cfcdf4317bdcf82077077111..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/ToVariableLoggingHandler.py.obsolete
+++ /dev/null
@@ -1,17 +0,0 @@
-"A logging handler to get a message from memory to a variable"
-
-import logging.handlers, copy
-
-class ToVariableLoggingHandler(logging.handlers.MemoryHandler):
-    def __init__(self):
-        logging.handlers.MemoryHandler.__init__(self, 2048)
-        self.copyTo=''
-    def handle(self, record):
-        "copy the message"
-        
-        logging.handlers.MemoryHandler.handle(self, record)
-        self.copyTo = copy.deepcopy(self.format(record))
-    def getMsg(self):
-        " return the message in demand"
-        
-        return self.copyTo
diff --git a/Tools/RunTimeTester/src/Tools.py b/Tools/RunTimeTester/src/Tools.py
deleted file mode 100755
index e185831d4f419d3baae8476360ae732ed88e913f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Tools.py
+++ /dev/null
@@ -1,284 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from exc2string2       import exc2string2
-#from Borg              import Borg
-from Tools2            import isValidReleaseNumber, fixSBit
-from stat              import ST_ATIME, ST_CTIME
-from RTTpath           import RTTpath
-from os.path           import join, isdir, isfile, islink, isabs
-from popen2            import Popen4
-from RTTSException     import RTTCodingError, RTTTimeOut
-import os, sys, shutil, time, string, types
-from ShellCommand import ShellCommand
-
-#----------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-smallLogger = logging.getLogger('rtt.standalone')
-
-#----------------------------------------------------
-
-def timed_cmd(c,secs=-1, execute=1):
-    def isTimedOut(start,secs):
-        if secs == -1: return False
-        return (time()-start)>secs
-    from time import time
-    from popen2 import Popen4
-
-    po   = Popen4(c)
-    proc = po.pid
-    out  = []
-    start = time()
-    while po.poll() == -1:
-        if isTimedOut(start, secs): 
-            m = 'Command timed out(%d secs)!\n%s' % (secs, c)
-            raise RTTTimeOut(m)
-        out.extend(po.fromchild.readlines())
-    del po
-    return out
-
-#----------------------------------------------------
-
-def pathIsOwnedByMe(path):
-    if os.path.exists(path):
-        return os.getuid() == os.stat(path).st_uid
-    return False
-        
-#----------------------------------------------------
-
-def changePerm(path,perm):
-    if pathIsOwnedByMe(path):
-        os.chmod(path,fixSBit(path, perm))
-    return os.stat(path).st_mode
-
-#----------------------------------------------------
-
-def mkdirIfNew(directory,perm=02775):
-    """Creates, with given permissions, the directory 'directory'
-(only if it does not already exist)
-This function originally used os.makedirs, but was rewritten
-as makedirs was not setting the permission perm.
-"""
-
-
-    perm = fixSBit(directory, perm)
-    if pathIsOwnedByMe(directory):
-        os.chmod(directory, perm)
-        return
-    else:
-        directory = os.path.normpath(directory)
-        head = str(directory)
-        start = head
-        if not isabs(head):
-            m = 'Cannot extend paths: start point must be absolute: %s' % head
-            logger.critical(m)
-            raise RTTCodingError(m)
-        tails = []
-        ncount = 0
-        while head != '/':
-            ncount += 1
-            if not  ncount % 100:
-                logger.warning('Making directories for path %s, sub dir count huge %d' % (start, ncount))
-            (head, tail) = os.path.split(head)
-            tails.append(tail)
-        tails.reverse()
-        for tail in tails:
-            head = os.path.join(head, tail)
-            if not os.path.exists(head):
-                try:
-                    os.mkdir(head,perm)
-                    changePerm(head,perm)
-                except Exception,e :
-                    msg  = 'Unable to make directory'
-                    msg += '%s\n%s\nTraceBack:\n%s' %(str(head),
-                                                      str(e),
-                                                      exc2string2())
-                    logger.error(msg)
-                    raise RTTCodingError(msg)
-            
-# ----------------------------------------------------
-
-def rmTreeErrorHandler(function, path, excinfo):
-    m = 'shutils.rmtree error. Traceback:\n%s\npath: %s\n' % (exc2string2(), path)
-    logger.error(m)
-
-    if os.path.isdir(path):
-        dirlist = os.listdir(path)
-        dirlist = [os.path.join(path, f) for f in dirlist]
-        for f in dirlist:
-            try:
-                os.remove(f)
-            except:
-                m = 'Could not remove %s' % f
-                logger.error(m)
-        
-#----------------------------------------------------
-
-def emptyDirMkDir(top,perm=02775):
-    perm = fixSBit(top, perm)
-
-    if not os.path.exists(top):
-        mkdirIfNew(top,perm)
-        return
-
-    # if the dire exists,  remove all subdirectories  and files            
-    [shutil.rmtree(os.path.join(top,d), False, rmTreeErrorHandler) for d in os.listdir(top) if os.path.isdir(os.path.join(top,d))]
-    # remove all files
-    [os.remove(os.path.join(top,f)) for f in os.listdir(top) if os.path.isfile(os.path.join(top,f))]
-    
-
-
-# ----------------------------------------------------
-    
-def rmdirMkdir(top,perm=02775):
-    perm = fixSBit(top, perm)
-
-    max = 3
-    ok  = False
-    while max:
-        if  not os.path.exists(top):
-            ok = True
-            break
-        max -= 1
-        shutil.rmtree(top, False, rmTreeErrorHandler)
-        
-    if not ok:
-        msg  = 'Unable to delete the directory tree below %s ' % str(top)
-        logger.error(msg)
-        raise RTTCodingError(msg)
-        
-    mkdirIfNew(top,perm)
-    logger.info('Made directory %s' % top)
-
-#takes the jobOptions+fileName and returns just the fileName
-def fileNameTrim(name,jobOptions):
-    """Takes the combined string jobOptions+fileName and returns just the fileName string"""
-    name = name.replace(jobOptions+"_","")
-    name = name.replace(jobOptions+".","")
-    return name
-
-#----------------------------------------------------
-
-def now():
-    """Returns a formatted string representing the present instant. Format example: 4May_1330"""
-    return time.strftime('%d%b_%H:%M', time.localtime())
-#----------------------------------------------------
-
-def nowSecs():
-    """Returns a formatted string representing the present instant. Format example: 4may_1330"""
-    return time.strftime('%d %b %H:%M:%S', time.localtime())
-    
-#----------------------------------------------------
-
-def isDateStampedToday(fileOrDir, midnight='00:00'):
-    """Returns true if the passed file or directory is date stamped today"""
-    from datetime import datetime
-
-    # first check out "now" date
-    todaysTime = time.localtime()
-    year,month,day,hour,min = todaysTime[0], todaysTime[1], todaysTime[2], todaysTime[3], todaysTime[4]
-    now_date = datetime(year,month,day,hour,min)
-
-    # check out the file or dir object now
-    stats = time.localtime(os.lstat(fileOrDir)[ST_CTIME])
-    fyear,fmonth,fday,fhour,fmin = stats[0], stats[1], stats[2], stats[3], stats[4]
-    file_or_dir_date = datetime(fyear, fmonth, fday, fhour, fmin)
-
-    diff_date = now_date - file_or_dir_date
-    if diff_date.days >= 1: return False
-
-    midn_hour = int(midnight.split(':')[0].strip())
-    midn_mins = int(midnight.split(':')[1].strip())
-    
-    def toSecs(h, m):
-        return (h*3600) + (m*60)
-
-    midnight_in_secs = toSecs(midn_hour, midn_mins)
-    now_in_secs      = toSecs(hour, min)
-    file_in_secs     = toSecs(fhour,fmin)
-
-    if midnight_in_secs > now_in_secs and midnight_in_secs > file_in_secs:
-        return now_in_secs > file_in_secs
-    if midnight_in_secs < now_in_secs and midnight_in_secs < file_in_secs:
-        return now_in_secs > file_in_secs
-    return file_in_secs > now_in_secs
-
-#----------------------------------------------------
-
-# taken from the python cookbook
-def unique(s):
-    """ return a list of the elements in s in arbitrary order,
-    but without duplicates"""
-
-    # get the special case of empty s out of the way rapidly
-    n = len(s)
-    if n==0: return []
-
-    # try using a dict - fast, usaually works
-    u={}
-    try:
-        for x in s: u[x]=1
-    except TypeError:
-        del u # move to next method
-    else:
-        return u.keys()
-
-    # can't hash elements: sort, and remove identical
-    # elements in a single pass
-
-    try:
-        t = list(s)
-        t.sort()
-    except TypeError:
-        del t # move to next method
-    else:
-        assert n>0
-        last = t[0]
-        lasti = i= 1
-        while i<n:
-            if t[i] != last:
-                t[lasti] = last= t[i]
-                lasti += 1
-            i += 1
-
-        return t[:lasti]
-
-    # only brute force remains
-
-    u = []
-    for x in s:
-        if x not in u:
-            u.append(x)
-    return u
-
-#----------------------------------------------------
-
-def sendMail(to,subject,message):
-    """A simple mailer"""
-    command = 'mail ' + str(to) + ' -s ' + '"' + str(subject) + '"'
-    os.popen(command,'w').write(message)
-
-#----------------------------------------------------
-
-def machineNameToIPaddress(machineName):
-    def legalIPformat(ipCandidate):
-        toks = ipCandidate.split('.')
-        if len(toks)!=4: return False
-        try:
-            [int(t) for t in toks]
-            return True
-        except:
-            return False
-        
-    po = Popen4('host %s' % machineName)
-    out = ''
-    while po.poll() == -1:
-        out += po.fromchild.read()
-    del po
-    if not out or out.find('not found')!=-1:
-        return 'http://localhost:8001'
-    ip = (out.split()[-1]).strip()
-    if not legalIPformat(ip): return 'http://localhost:8001'
-    return 'http://%s:8001' % ip
-
-# ----------------------------------------------------------------------
diff --git a/Tools/RunTimeTester/src/Tools2.py b/Tools/RunTimeTester/src/Tools2.py
deleted file mode 100755
index 9d8eeb87c02228a478b43b3a6fa77a9125520985..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Tools2.py
+++ /dev/null
@@ -1,433 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#
-# These functions have been split of from thr Tools module to avoid
-# circular module imports
-#
-
-import os, time, types
-from os.path       import isdir, join
-from stat          import *
-from popen2        import Popen4
-from RTTSException import RTTCodingError
-from RTTDateTime   import RTTDateTime
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-## relPath 
-# class to construct relative path between two absolute paths       
-class relPath:
-    """Author: K.Ciba"""
-
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @brief it is empty, as it should be
-    def __init__(self):
-        pass
-
-    ## split path to dir chunks
-    # @param self "Me, myself and Irene"
-    # @param path a path chunk
-    # @param rest a list with tails chunks
-    def pathSplit( self, path, rest=[] ):
-        ( head , tail ) = os.path.split( path )
-        if len(head) < 1: return [tail]+ rest
-        if len(tail) < 1: return [head]+ rest
-        return self.pathSplit( head, [tail] + rest )
-
-    ## finds common chunk in two paths
-    # @param self "Me, myself ad Irene"
-    # @param l1 1st path
-    # @param l2 2nd path
-    # @param common a list of common chunks
-    def commonPath( self,  l1, l2, common=[] ):
-        if len(l1) < 1: return (common, l1, l2)
-        if len(l2) < 1: return (common, l1, l2)
-        if l1[0] != l2[0]: return (common, l1, l2)
-        return self.commonPath( l1[1:], l2[1:], common+[l1[0]])
-
-    ## finds relative path between path1 and path2
-    # @param self "Me, myself and Irene"
-    # @param p1 1st path
-    # @param p2 2nd path
-    # @return string with relative path between p1 and p2
-    def __call__( self, p1, p2 ):
-        if ( p1 == p2 ): return "./"
-        ( common, l1, l2 ) = self.commonPath( self.pathSplit( p1 ), self.pathSplit( p2 ) )
-        p = []
-        if len(l1) > 0:
-            p = [ '../' * len(l1) ]
-        p = p + l2
-        return os.path.join( *p )
-
-
-def toRelPath(p1,p2):
-    r = relPath()
-    return r(p1,p2)
-
-def deSpamifyEmailAddress(address):
-    return address.replace('NOSPAM','')
-
-def isValidReleaseNumber(rel):
-    """Is the release string passed a valid numbered release i.e.
-    of the form X.Y.Z"""
-
-    splitRel = rel.split('.')
-    if not ( (len(splitRel) == 3) or (len(splitRel) == 4) ):
-        return False
-    else:
-        for thing in splitRel:
-            # if len(thing) != 1 or not thing in string.digits:
-            #    return False
-            try:
-                int(thing)
-            except Exception, inst:
-                return False # not necesarily an error - just not a valid release number
-
-    return True
-
-def releaseNumberComponents(relString):
-    """return the integer compoenents of a relase string"""
-
-    bad = (None, None, None)
-    if not isValidReleaseNumber(relString): return bad
-    
-    splitRel = relString.split('.')
-    if len(splitRel) != 3:
-        return bad
-    else:
-        rel=[]
-        for thing in splitRel:
-            # if len(thing) != 1 or not thing in string.digits:
-            #    return False
-            try:
-                rel.append(int(thing))
-            except Exception, inst:
-                return bad # not necesarily an error - just not a valid release number
-
-    return (rel[0], rel[1], rel[2])
-
-
-#----------------------------------------------------
-
-def popenCmd(cmd):
-    po = Popen4(cmd)
-
-    out = []
-        
-    while po.poll() == -1:
-        out.extend(po.fromchild.readlines())
-        
-    del po
-    return out
-
-#----------------------------------------------------
-
-def fixSortBranches(branches):
-    """Imposed 'sorting'."""
-    
-    theOrder = ['N.0.X','10.0.X','N.X.0','Numbered']
-    
-    theLen = len(branches)
-    done = False
-    while not done:
-        changedOrder = False
-        i = 0
-        while i < theLen - 1:
-            curr = branches[i]
-            next = branches[i+1]
-            if theOrder.index(curr) > theOrder.index(next):
-                branches[i],branches[i+1] = branches[i+1],branches[i]
-                changedOrder = True
-
-            i = i + 1 
-
-        done = (changedOrder==False)
-        
-    return branches
-    
-# ----------------------------------------------------------------------
-
-def extractReleaseFromNightlyLink(theLink):    
-    """Extract the release from the stamp file (which is a soft link)"""
-    theLink = str(theLink) # might be an RTTpath instance
-    if os.path.islink(theLink):
-        return os.readlink(theLink)
-    
-    return theLink
-# ----------------------------------------------------
-
-def getFileOrDirTimeStamp(fileOrDir):
-    """Returns a three-member list [hours,mins,secs] representing the time stamp of a given file or directory"""
-
-    fileOrDirTime = []
-    
-    if os.path.exists(fileOrDir):
-        return RTTDateTime(os.lstat(fileOrDir)[ST_ATIME])
-    else:
-        return RTTDateTime()
-# ----------------------------------------------------
-
-def getFileOrDirDateStamp(fileOrDir):
-    """Returns a three-member list [year,month,day] representing the date stamp of a given file or directory"""
-
-    fileOrDirDate = []
-    if os.path.exists(fileOrDir):
-        return RTTDateTime(os.lstat(fileOrDir)[ST_ATIME])
-    else:
-        return RTTDateTime()
-
-
-#----------------------------------------------------
-
-
-# def getAllResultDirs(resBase, legalBranches, legalBuilds, legalRunTypes):
-def getAllResultDirs(resBase, legals):
-    """Returns a list of all valid results directories below 'resBase'."""
-
-
-    
-    # get all valid results dirs below resBase
-    relDict = makeReleaseDictionary(resBase,legals)
-   
-    # sort the build keys reverse alphabetically
-    builds = relDict.keys()
-    builds.sort()
-    builds.reverse()    
-    
-    finalList = []
-
-    for build in builds:        
-        
-        branches = relDict[build].keys()
-        branches = fixSortBranches(branches)
-        
-        #        branches.sort() #alphabetically
-        
-        for branch in branches:
-            runTypes = relDict[build][branch]
-            
-            for runType in runTypes:
-
-                releases = relDict[build][branch][runType]
-
-                for release in releases:
-                    finalList.append({'build':build,'branch':branch,'runType':runType,'release':release}) 
-    
-    return finalList
-    
-# ----------------------------------------------------
-# ----------------------------------------------------
-        
-def jobStatusFromFile(job, resPath, statusFile=""):
-
-    numbEntriesPerLine = 7 # number of entries we normally have in a given line
-    
-    if statusFile == "":
-        statusFile = join(resPath, "status.txt")
-
-    if os.path.exists(statusFile):
-        ifile = open(statusFile,"r")
-
-        linesInFile = ifile.readlines()
-        ifile.close()
-        
-        linesInFile.reverse()
-
-        
-        matchesFound = []
-        for line in linesInFile:
-        
-            if line.find(job) != -1:
-                tokens = line.split()
-                # Previously each line in the status.txt log file had the form:
-                #      Job  Status  Date  Time
-                # now they each have the form:
-                #     Job  Status  Date  Time  KitOrReleaseRun  Platform  NICOSdate
-                # We need to take this into account here.                
-
-                numbEntriesFoundInLine = len(tokens)
-                while numbEntriesFoundInLine < numbEntriesPerLine:
-                    tokens.append('n/a')
-                    numbEntriesFoundInLine += 1
-
-                returnList = []
-                [returnList.append(thing) for thing in tokens[1:]]
-                returnTuple = tuple(returnList)
-
-                time.sleep(3)
-                return returnTuple
-
-    return ("n/a",)*(numbEntriesPerLine-1)
-
-#----------------------------------------------------
-
-def makeReleaseDictionary(resBase, legals):
-    """Searches for all valid results directories below 'resBase',
-    and builds a dictionary representing them, keyed on build whose
-    values themselves are dictionaries. These element dictionaries
-    are keyed on branch and have values which are lists of releases sorted
-    such that nightly releases are first, in alphabetical order, then numbered
-    releases in reverse numerical order. Returns the final dictionary.
-    """
-    
-    relDict = {}
-
-    lBuilds   = legals.vals('build')
-    lBranches = legals.vals('branch')
-    lRunTypes = legals.vals('runType')
-    
-    for buildDir in os.listdir(resBase):
-        if not buildDir in lBuilds or not isdir(join(resBase, buildDir)):
-            continue 
-        
-        resBaseBuild = join(resBase,buildDir)
-
-        relDict[buildDir] = {}
-            
-        for branchDir in os.listdir(resBaseBuild):            
-            if not branchDir in lBranches or not isdir(join(resBaseBuild,
-                                                            branchDir)):
-                continue
-            
-            resBaseBuildBranch = join(resBaseBuild,branchDir)
-
-            relDict[buildDir][branchDir] = {}
-
-            for runTypeDir in os.listdir(resBaseBuildBranch):
-                if not runTypeDir in lRunTypes: continue
-                if not isdir(join(resBaseBuildBranch, runTypeDir)): continue
-
-                releasesInDir = os.listdir(join(resBaseBuildBranch,runTypeDir))
-                relDict[buildDir][branchDir][runTypeDir] = sortReleases(releasesInDir, legals)
-
-    return relDict
-
-#----------------------------------------------------
-
-def sortReleases(rels, legals):
-    """Given a list of releases, returns a list of sorted releases, nightlies first in alphabetical order,
-    followed in reverse numerical order by numbered releases."""
-
-    sorted = []
-    numbered = [rel for rel in rels if isValidReleaseNumber(rel)]
-    nightly  = [rel for rel in rels if legals.isNightlyInRelease(rel)]
-    # sort the two lists
-    nightly.sort()
-    numbered = numericalSort(numbered)
-    
-    sorted.extend(nightly)
-    sorted.extend(numbered)
-
-    return sorted
-
-#----------------------------------------------------
-
-def numericalSort(rels):
-    """Takes a list of numbered releases and returns the list sorted in decreasing numerical order."""
-    if len(rels)== 0 or len(rels)==1: return rels
-
-    lists = {'1':[],'2':[]}
-    for thing in rels:
-        if len(thing.split('.')[0]) == 1:
-            lists['1'].append(thing)
-        else:
-            lists['2'].append(thing)
-        
-    lists['1'].sort()
-    lists['1'].reverse()
-    lists['2'].sort()
-    lists['2'].reverse()
-    finalList = lists['2']
-    finalList.extend(lists['1'])
-    return finalList
-
-def searchPaths(paths, file):
-    found = False
-    for p in paths:
-        tryPath = os.path.join(p, file)
-        if os.path.exists(tryPath): return p
-    return None
-
-
-#----------------------------------------------------
-
-def setProtectionFile(path, perm):
-    if not os.path.exists(path): return
-    os.chmod(path, fixSBit(path, perm))
-
-#----------------------------------------------------
-
-def fixSBit(directory, perm):
-    # kluge. The 2000 bit cnnot be set for afs
-    if not directory.startswith('/afs'):return perm
-    # if directory.startswith('/afs'):return perm
-
-    
-    operm = oct(perm)
-
-    
-    if len(operm)<5: return perm # not enough digits to require correction
-
-    
-    bit11set = ['2', '3', '6', '7']
-    if operm[-4:-3] in bit11set: perm = perm-02000
-    return perm
-
-#----------------------------------------------------
-
-def convertToSeconds(timeString):
-    """Takes a time string of format HH:MM:SS, and returns a
-    string representing an integer number of seconds."""
-    fields = timeString.split(':')
-    if len(timeString)!=8 or len(fields)!=3: return ''
-    try:
-        hours = int(fields[0])
-        mins  = int(fields[1])
-        secs  = int(fields[2])
-    except:
-        return ''
-
-    timeInSecs = (hours*3600) + (60*mins) + secs
-    return str(timeInSecs)
-
-    
-    # ----------------------------------------------------------------------
-
-def escapeTheCommand(cmd):
-    
-    toBeEscaped = ['(',')','"']
-    
-    for char in toBeEscaped:
-        pieces = cmd.split(char)
-        
-        cmd = pieces[0]
-        for piece in pieces[1:]:
-            cmd += '\\'+char+piece
-
-    return cmd
-    
-def isType(instance, name):
-    """name is a string. Determnes if class with name name is
-    in the class heirarchy of an instance"""
-    
-    if not (type(name) is types.StringType):
-        m = 'isType: argument passed is of type %s, should be of type StringType' % name.__class__.__name__
-        raise RTTCodingError(m)
-
-    name = name.strip()
-    
-    if instance.__class__.__name__ == name: return True
-
-    bases = [b.__name__ for b in instance.__class__.__bases__]
-    if name in bases: return True
-    return False
-            
-            
-if __name__ == '__main__':
-    thePath = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results/rel_4/dev/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/testDebug.log'
-    cmd = 'md5sum %s' % thePath
-    print popenCmd(cmd)
diff --git a/Tools/RunTimeTester/src/Tracer.py b/Tools/RunTimeTester/src/Tracer.py
deleted file mode 100644
index 8e29f2cfa17e38c95057308fd0f678589aa5e952..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Tracer.py
+++ /dev/null
@@ -1,22 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import types
-def tracing(f, name):
-    def traced_f(*a, **k):
-        # print '%s(%s,%s) -> '%(name, a,k),
-        print '\n-- Trace:%s --'%(name)
-        result = f(*a, **k)
-        print '-- Trace:%s result %s --'%(name, result)
-        return result
-    return traced_f
-
-class meta_tracer(type):
-    def __new__(self, classname, bases, classdict):
-        for f in classdict:
-            m = classdict[f]
-            if isinstance(m, types.FunctionType):
-                classdict[f]=tracing(m, 'Trace: %s.%s' % (classname, f))
-        return type.__new__(self, classname, bases, classdict)
-
-class Tracer:
-    __metaclass__ = meta_tracer
diff --git a/Tools/RunTimeTester/src/UnifiedConfigurationTestRun.py b/Tools/RunTimeTester/src/UnifiedConfigurationTestRun.py
deleted file mode 100644
index 59c5661717153d65dcfff673a1b6969ba85b0279..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/UnifiedConfigurationTestRun.py
+++ /dev/null
@@ -1,181 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-TestRun subclasses threading.Thread to allow concurrent running of instances.
-
-02/02/09 out of date documentation
-
-1. Submit the jobs. One JobMinder per job is created. JobMinders incoporate the
-   JobDescriptor, and then monitor the progress of the job.class 
-
-2. A poll command is periodically sent to the JobMinders. These check their job status.
-   This status is interogated, and html pages are updated with new status info.
-
-3. After all jobs have completed perform cleanup activities
-"""
-#----- Python modules -----
-
-import time, gc
-
-#----- RTT modules -----
-
-from Factory_Launcher    import launcherFactory
-from Tools               import sendMail
-from Tools2              import deSpamifyEmailAddress
-
-from MethodTimer         import MethodTimer
-
-from memstat             import memstat
-from exc2string2         import exc2string2
-from RTTSException       import RTTCodingError
-import threading
-from MethodTracer        import MetaTracer
-
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-# -------------------------------------------------------------------------
-# -------------------------------------------------------------------------
-
-from TestRun import TestRun
-
-class UnifiedConfigurationTestRun(TestRun):
-    
-    def __init__(self, minders, packageName, closeDownKey, closeDownKeyMaker,
-                 mode, devRun, mailer, logger, statusText=''):
-        # the summaryCallBack object passed in is the TesterXMLConverter.
-
-        statusText = 'initialising'
-        TestRun.__init__(self, minders, packageName, closeDownKey, closeDownKeyMaker,
-                         logger, statusText)
-
-        self.mode   = mode
-        self.devRun = devRun
-        self.mailer = mailer
-
-# -------------------------------------------------------------------------
-
-    def removeBadMinders(self):
-        msg  = 'package = '+str(self.name)
-        msg += ' No. of jobs before removing bad jobs '+str(len(self.jobMinders))
-        self.logger.debug(msg)
-        
-        # [m.makeJobGroupDirs() for m in self.jobMinders]
-        jobsToRemove = [m for m in self.jobMinders if m.removeMe]
-
-        lminders = len(self.jobMinders)
-        ltoremove = len(jobsToRemove)
-        [m.unlink() for m in jobsToRemove]
-        [self.jobMinders.remove(m) for m in jobsToRemove]
-        
-        assert (lminders-ltoremove-len(self.jobMinders)==0)
-        
-        msg  = 'package = '+str(self.name)
-        msg += ' No. of jobs after removing bad jobs '+str(len(self.jobMinders))
-        self.logger.info(msg)
-        
-# ------------------------------------------------------------------------- 
-
-    def run(self):
-        """split the jobs into various phases. Each phase runs to completion
-         before starting the next one."""
-        self.logger.info("TestRun run method starts")
-
-        self.statusText = 'running'
-
-        rttPilotJobs = [j for j in self.jobMinders if j.rttPilotJob]
-        
-        atnPilotJobs = [j for j in self.jobMinders if j.rttATNJob]
-                             
-        rttMainJobs  = [j for j in self.jobMinders if (j not in atnPilotJobs and j not in rttPilotJobs)]
-
-        phases = [('atnPilotJobPhase', atnPilotJobs), ('rttPilotJobPhase', rttPilotJobs), ('rttMainJobPhase',rttMainJobs)]
-        for phaseName, phaseJobs in phases:
-            self.phase = phaseName
-            m = "TestRun runMinder method starts for phase %s" % phaseName
-            self.logger.info(m)
-            self.runMinders(phaseJobs)
-            [self.jobMinders.remove(m) for m in phaseJobs] # allow garbage collection
-            [m.unlink() for m in phaseJobs]
-            self.xmlConverter.pushData()
-
-        self.mailer.sendMailToPackageDevelopers()
-
-        self.statusText = 'done'
-        self.logger.debug('status text set to done')
-        self.xmlConverter.finalise()
-        self.logger.debug('converter update method called.')
-
-# -------------------------------------------------------------------------
-
-    def runMinders(self, jobMinders):
-
-        # make a list of jobs which do not have self.done = True
-        msg = "TestRun job launching starts - active jobs: %d, jobs this phase %d" % (len(self.jobMinders), len(jobMinders))
-        self.logger.debug(msg)
-        s = 'Jobs:\n'
-        for j in jobMinders: s += '%s:%s\n' % (j.__class__.__name__, j.identifiedName)
-        self.logger.debug(s)
-
-        
-        launcher = launcherFactory(self.mode, jobMinders, self.logger)
-
-
-        sleepSeconds = 300
-        if self.devRun: sleepSeconds = 3 # for development purposes, shorten sleeptime
-        
-
-        msg = '%s entering poll loop: no of jobs: %d sleep time %d s' % (self.name,
-                                                                         len(jobMinders),
-                                                                         sleepSeconds
-                                                                         )
-        self.logger.debug(msg) # goes to screen
-
-            
-        while jobMinders:
-            self.pollLoopBody(jobMinders, launcher)
-
-            # self.printMethodTimes()
-
-            # don't take this out!
-            time.sleep(sleepSeconds)
-
-        self.logger.info('TestRun exited poll loop, leaving TestRun run() method.')
-            
-    # -------------------------------------------------------------------------
-
-    def pollLoopBody(self, jobMinders, launcher):
-
-            if self.operatorCloseDownTest():
-                self.logger.info('TestRun %s detected operator close down signal' % self.name)
-                [m.queueOperatorCloseDown() for m in jobMinders]
-                
-            launcher.launchNextUnlaunchedJob()
-
-            try:
-                [m.poll() for m in jobMinders]
-            except:
-                m = 'Error polling minder: %s ' % exc2string2()
-                self.logger.error(m)
-                self.logger.error(memstat())
-                self.logger.error('Garbage:\n%s' % gc.garbage)
-                self.logger.error('Collecting garbage...')
-                gc.collect()
-                self.logger.error('...Collected garbage')
-                raise RTTCodingError(m)
-                
-                
-            # the following print is done once when the minder goes to done
-            # the following is debug
-            # [m.printMethodTimes() for m in self.jobsActive]
-
-
-            toRemove  = [m for m in jobMinders if m.isDone()]
-            self.mailer.collectStats(toRemove) # info gathering for email sending purposes
-            [m.printMethodTimes() for m in toRemove]
-            [m.unlink() for m in toRemove]
-            [jobMinders.remove(m) for m in toRemove]
-            self.doneJobs       += len(toRemove)
-            self.successfulJobs += len([j for j in toRemove if j.isSuccess()])
-            
-            self.xmlConverter.pushData()
-
diff --git a/Tools/RunTimeTester/src/UseOfPaths.txt b/Tools/RunTimeTester/src/UseOfPaths.txt
deleted file mode 100755
index 84dc811cb1e88b79112705d9bd8ee420d6a2038f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/UseOfPaths.txt
+++ /dev/null
@@ -1,248 +0,0 @@
-Paths - created in Tester and passed to TestRun.__init__
-for all TR subclasses
-
-TestRun/__init__
-self.paths = paths !
-paths.package
-paths.confFile
-paths.resultsDirs['resBase/build/branch/runType/release']
-paths.mode
-paths.build
-paths.release
-self.paths.workDirs['workBase/']
-
-TestRun.makeLog
-self.paths.logDir
-
-TestRun.checkFreeSpace
-paths.site
-
-TestRun.checkForUserShutDownSignal
-paths.resultsDirs['resBase/']
-
-TestRun.getJobStatsForEmail
-m in self.jobMinders:
-m.jDescriptor.paths.paths.package
-m.jDescriptor.paths.paths.package
-
-TestRun.__str__
-self.paths
-
-
-UnifiedConfigurationTestRun.setKitsAndJobs
-SuperParser.parse(self.paths)
-
-UnifiedConfigurationTestRun.closeDown
-self.jobMinders[0].jDescriptor.paths.mailTo
-self.jobMinders[0].jDescriptor.paths.package
-self.jobMinders[0].jDescriptor.paths.confFile
-
-UnifiedConfigurationTestRun.makeRunDirectories
-self.paths.package
-
-UnifiedConfigurationTestRun.run
-self.paths.mode
-self.paths.platform
-self.paths.runTyp
-self.paths.kitID
-
-FailureTestRun.__init__
-paths.package
-self.paths = paths
-
-
-===============================
-
-SuperParser.parse
-paths.mailTo     = 
-paths.refRelease =
-JobMaker.(jobListElement, paths)
-
-==================================
-JobMaker.__init__
-self.paths       = paths
-
-JobMaker.makeWatcherDescs
-descriptorFactory('Watcher', self.paths)
-
-UnifiedConfigurationJobMaker.makeWorkerDescs
-JobsXMLReader.makeJobDescriptors(self.element, self.paths)
-self.paths.rttKitTest
-self.paths.release
-
-============================
-descriptorFactory
- clonedPaths = deepcopy(paths)  !!
-
-AthenaJobDescriptor
-paths.refRelease
-paths.installPath
-
-ReleaseJobDescriptor,
-self.paths            = paths
-
-
-Descriptor.__init__
-self.paths          = paths !
-paths.mode 
-
-Descriptor.setRunPath
-self.paths.workDirs['workBase/build/branch/runType/release']
-self.paths.workDirs[curWorkKey] = 
-self.paths.workDirs['workBase/build/branch/runType/release/jobGroup/jobName/SN']
-Descriptor.makeDBKey
-self.paths.build
-
-
-Descriptor.__str__
-self.paths.__str__
-
-
-WatcherJobDescriptor
-WorkerJobDescriptor
- Descriptor.__init__(self, paths)
-
-JobsXMLReader.removeNoNightlyJobs
-paths.release
-
-
-JobsXMLReader.makeJobDescriptors
- 
-handleAthenaJobs       (paths, athenaTags)
-handleRuleCheckerJobs  (paths, rulecheckerTags)
-handleQAmetricsJobs    (paths, qametricsTags)
-handleKitValidationJobs(paths, kitvalidationTags)
-handleMonitoringJobs   (paths, monitoringTags)
-handleRTTMonitoringJobs(paths, rttMonitoringTags)
-
-JobsXMLReader.handleAthenaJobs
-str(paths)
-handleAthenaJob(paths, job)
-
-JobsXMLReader.handleAthenaJob
-descriptorFactory('Athena', paths)
-
-JobsXMLReader.handleRuleCheckerJobs
-handleRuleCheckerJob(paths, job)
-
-JobsXMLReader.handleRuleCheckerJob(paths, job)
-descriptorFactory('Release', paths)
-handleRCEntryPoints(paths, job.getElementsByTagName("package"))
-
-JobsXMLReader. handleRCEntryPoints(paths, entryPointTags)
-rulecheckerAll(paths)
-rulecheckerSelected(paths, entryPoints)
-
-JobsXMLReader.(paths, entryPoints):
-getNewEntryPoints(paths, ep)
-
-JobsXMLReader.getNewEntryPoints(paths, entryPoint)
-CMTtools.writeTopCMTRequirementsFile(paths, epList)
-paths.workDirs['workBase/']
-UseToPackage(paths.releasePath)
-
-JobsXMLReader.rulecheckerAll(paths)
-CMTtools.writeTopCMTRequirementsFile(paths, ['AtlasRelease'])
-paths.workDirs['workBase/']
-UseToPackage(paths.releasePath) 
-
-
-JobsXMLReader.countSrcFiles(paths, package)
-paths.releasePath
-srcFileCounter(pathToCMTFile, paths)
-
-JobsXMLReader.handleKitValidationJob(paths, job)
-descriptorFactory('KitValidation', paths)
-
-JobsXMLReader.handleMonitoringJob(paths, job)
-descriptorFactory('Monitor', paths)
-
-
-================================
-Explicit
-
-Each TestRun gets a new Paths
-explicit use
-
-paths.package
-paths.confFile
-paths.resultsDirs['resBase/build/branch/runType/release']
-paths.mode
-paths.build
-paths.release
-self.paths.workDirs['workBase/']
-self.paths.logDir
-paths.site
-
-UnifiedConfigurationTestRun
-self.paths.platform
-self.paths.runType
-self.paths.kitID
-
-
-
-The paths object is expanded in SuperParser to enclude 
-paths.mailTo     = 
-paths.refRelease =
-
-
-is then passed on to the JobMaker classes which explitily ise
-
-self.paths.rttKitTest
-self.paths.release
- 
-and pass to descriptorFactory and JobsXMLReader.makeJobDescriptors
-
-descriptorFactory deep copies the Paths obj
-
-and creates descriptors. These explicitly use
-
-paths.refRelease
-paths.installPath
-paths.mode 
-self.paths.workDirs['workBase/build/branch/runType/release']
-self.paths.workDirs[curWorkKey] = 
-self.paths.workDirs['workBase/build/branch/runType/release/jobGroup/jobName/SN']self.paths.build
-paths.__str__
-
-
-JobsXMLReader explicitly uses
-paths.workDirs['workBase/']
-paths.releasePath
-
-------------------
-Explicit use with no overlaps
-
-TestRun
-
-paths.package
-paths.confFile
-paths.resultsDirs['resBase/build/branch/runType/release']
-paths.mode
-paths.build
-paths.release
-self.paths.workDirs['workBase/']
-self.paths.logDir
-paths.site
-self.paths.platform
-self.paths.runType
-self.paths.kitID
-
-Added in SuperParser
-paths.mailTo     = 
-paths.refRelease =
-
-In JobMaker
-self.paths.rttKitTest
-
-In Descriptors
-deepCopy, then uses  
-
-paths.refRelease
-paths.installPath
-paths.workDirs['workBase/']
-self.paths.workDirs['workBase/build/branch/runType/release']
-self.paths.workDirs[curWorkKey] = 
-self.paths.workDirs['workBase/build/branch/runType/release/jobGroup/jobName/SN']
-paths.__str__
-paths.releasePath
diff --git a/Tools/RunTimeTester/src/UserLogger.py b/Tools/RunTimeTester/src/UserLogger.py
deleted file mode 100755
index 26b418aba7497c2ac01a3d3bb717cc787f080afb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/UserLogger.py
+++ /dev/null
@@ -1,56 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-import logging
-
-from LoggerSetup import *
-
-class UserLogger:
-    # def __init__(self):
-    #    print 'instatiated UserLogger'
-        
-    def makeLogger(self, runPath, identifiedName):
-        logFileDir = runPath
-        
-        formatter = createFormatter()
-        
-        errorLogName   = createLogFile(logFileDir,
-                                       'postprocessing.error.log',
-                                       'postprocessing.error.log\n') 
-
-        warningLogName = createLogFile(logFileDir,
-                                       'postprocessing.warning.log',
-                                       'postprocessing.warning.log\n') 
-
-        infoLogName    = createLogFile(logFileDir,
-                                       'postprocessing.info.log',
-                                       'postprocessing.info.log\n') 
-
-        debugLogName   = createLogFile(logFileDir,
-                                       'postprocessing.debug.log',
-                                       'postprocessing.debug.log\n') 
-
-        handlers = []
-        handlers.append(createFileHandler(debugLogName,
-                                          logging.DEBUG,
-                                          formatter))
-        handlers.append(createFileHandler(errorLogName,
-                                          logging.ERROR,
-                                          formatter))
-        handlers.append(createFileHandler(warningLogName,
-                                          logging.WARNING,
-                                          formatter))
-        handlers.append(createFileHandler(infoLogName,
-                                          logging.INFO,
-                                          formatter))
-
-        myLogger = logging.getLogger('UserLogger.%s' % identifiedName)
-        [myLogger.removeHandler(h) for h in myLogger.handlers]
-        [myLogger.addHandler(h) for h in handlers] 
-
-        myLogger.setLevel(logging.DEBUG)
-        return myLogger
-
-if __name__ == '__main__':
-    ul = UserLogger()
-    ul.makeLogger()
diff --git a/Tools/RunTimeTester/src/UserStuffRetriever.py b/Tools/RunTimeTester/src/UserStuffRetriever.py
deleted file mode 100755
index 32e7cad3fb720c63704aabbad0487a0b51ba56bb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/UserStuffRetriever.py
+++ /dev/null
@@ -1,304 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from CVSPackage             import CVSPackage
-from ShellCommand           import ShellCommand
-from formatCollection       import formatCollection
-from RTTpath                import RTTpath
-from pathToProjectCMTDir    import pathToProjectCMTDir
-from pathToProjectCMTDir    import allProjectCMTDirs
-# import CMTtools
-
-import os, sys, glob, string
-
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-smallLogger = logging.getLogger('rtt.standalone')
-# -------------------------------------------------------------------------
-
-from MethodTracer        import MetaTracer
-
-# __metaclass__ = MetaTracer
-    
-"""
-UserStuffRetriever. Finds the configuration files for packages which define
-them in a CMT macro.
-
-A substring of the macro name is used to find macros with that name.
-By convention, the macro name must start with <package>_
-
-If the paths object passed in the constructor has no package name set, the
-entire release is queried via AtlasRelease for the macros.
-
-If the package _is_ given in the paths object, UserRetriever only asks that package.
-This assumes that the user has set some CMT stuff outside the RTT.
-
-"""
-
-class UserStuffRetrieverBase:
-    def __init__(self, argBag):
-        logger.debug('%s.__init__ starts' % self.__class__.__name__)
-        
-        # self.cmtHomeDir      = argBag.cmtHomeDir
-        # self.cmtProjDir      = argBag.cmtProjDir
-        self.hardCoded       = argBag.hardCoded
-        self.topProject      = argBag.topProject
-        self.substring       = argBag.confFileSuffix
-        self.projectCMTDirs  = argBag.projectCMTDirs
-        self.cmds            = argBag.cmds
-        self.init            = False
-        self.packageStuff    = {}
-        self.packages        = {}
-
-    # def homeRequirements(self):       return self.cmtHomeDir
-    # def projectRequirements(self):    return self.cmtProjDir
-        
-    def readMacros(self, project):
-
-        macros = self.macrosFromCMT(project)
-        # macros = self.macrosFromFile()
-
-        # print 'macros'
-        # print macros
-        
-
-        # the macros that interest us have one '=' sign
-        allMacros = {}
-        print 'no of macros', len(macros)
-        # print macros
-        for m in macros:
-
-            # if m.find('RunTimeTester')>-1: print m
-
-            split = m.split('=')
-            if len(split) != 2: continue
-
-            # if split[0].find('RunTimeTester')>-1: print split
-            
-            allMacros[split[0].strip()] = split[1].strip()
-
-        # print
-        # print 'self.allMacros'
-        # print formatCollection(self.allMacros)
-        # print 'self.allMacros end'
-        # print
-
-        
-        return allMacros
-    
-    def packageInfo(self, package):
-
-        self.userStuff()
-
-        return self.packages.get(package,
-                                CVSPackage(package, '', '', '', '', '', '',
-                                           '', '',
-                                           'Insufficient info from  CMT'))
-    
-    def userStuff(self):
-        " sets up a dictionary packageName: CVSPackage returns list packages"
-
-        if not self.init:
-            self.makePackages()
-            self.init = True
-
-        return self.packages.values()
-
-    def getPackageInfo(self, allMacros):
-
-        terms = ['_project_release', '_offset', '_cmtPath', '_project',
-                 '_TestConfiguration']
-
-        # build a dictionary with
-        # packageName : dict
-        # dict is a dictionary of term: macroValue
-        for macroName in allMacros.keys():
-            if macroName.find(self.substring)>=0:
-                packageName = macroName[:macroName.rfind(self.substring)]
-                self.packageStuff[packageName] = {}
-
-        # save macros which seem to be related to the package.
-        for macroName in allMacros.keys():
-            for packageName in self.packageStuff.keys():
-                if macroName.startswith(packageName):
-                    key = macroName[len(packageName):]
-                    val = allMacros[macroName][1:-1]
-                    self.packageStuff[packageName][key]=val
-
-
-        # handle the case where the package part of the macro is uppercase
-        terms = ['VERSION']
-        for macroName in allMacros.keys():
-            for t in terms:
-                if macroName.find(t)>0:
-                    packageName = macroName[:macroName.rfind(t)]
-                    for p in self.packageStuff.keys():
-                        # add to dictionary only if the packagename is already
-                        # present
-                        if p.upper() == packageName:
-                            #strip the surrounding quotes
-                            val = allMacros[macroName][1:-1]
-                            self.packageStuff[p][t] = val
-                                                               
-
-        # make dictionary of CVSpackages
-        m  = 'About to examine packages for sufficient information from CMT.\n'
-        m += 'Currently have %d packages\n' % len(self.packageStuff.keys())
-        logger.debug(m)
-        packages = {}
-        for package in self.packageStuff.keys():
-            dict = self.packageStuff[package]
-            print 'packageStuff dict',dict
-            #needs = ['_cmtpath', '_TestConfiguration', '_offset',
-            #         'VERSION', '_root']
-            needs = ['_cmtpath', '_TestConfiguration',
-                     'VERSION', '_root']
-            notPresent = [n for n in needs if n not in dict.keys()]
-            if notPresent:
-                m  = 'Could not find package information '
-                m += '%s for package %s: ' % (str(notPresent), package)
-                m += 'not handling this package'
-                m += 'package dictionary:\n%s\nend' % formatCollection(dict)
-                
-                logger.warning(m)
-                print m
-                continue
-
-            configFilePath = RTTpath(
-                dict['_root'],
-                'cmt',
-                dict['_TestConfiguration']
-                )
-
-            pathToConfigFile = str(configFilePath)
-            errMsg = '' # will be passed onto the CVSPackage
-            if not configFilePath.exists():
-                errMsg = 'Non-existent config file %s' % pathToConfigFile
-                logger.error(errMsg)
-             
-
-            # packageOffset     = dict['_offset']
-            packageOffset     = 'notUsed'
-            packageVersion    = dict['VERSION']
-            packageCmtPath    = dict['_cmtpath']
-            packageRoot       = dict['_root']
-
-            packageSharePath = self.makePackageSharePath(dict) # allow subclasses to alter for eg local running 
-            # packageSharePath  = RTTpath(packageCmtPath,
-            #                            self.hardCoded.installAreaDirName(),
-            #                            self.hardCoded.shareAreaDirName()
-            #                            )
-            if not packageSharePath.exists():
-                msg = 'Non-existant share path %s\n ' % str(packageSharePath)
-                msg += 'for package %s' % package
-                logger.error(msg)
-                continue
-                
-            print '**************************************'
-            print 'package name............', package
-            print 'packageOffset ',   packageOffset
-            print 'packageCmtPath ',  packageCmtPath
-            print 'pathToConfigFile', pathToConfigFile
-            
-            packages[package] = CVSPackage(package,
-                                           pathToConfigFile,
-                                           packageOffset,
-                                           packageVersion,
-                                           packageCmtPath,
-                                           packageRoot,
-                                           str(packageSharePath),
-                                           # self.cmtHomeDir,
-                                           # self.cmtProjDir,
-                                           errMsg)
-            
-        plist = packages.keys()
-        plist.sort()
-        m += '%s found %d packages\n%s\n' % (self.__class__.__name__,
-                                             len(plist),
-                                             str(plist))
-
-        
-        logger.info(m)
-
-        return packages
-
-    def makePackageSharePath(self, dict):
-        return RTTpath(dict['_cmtpath'],
-                       self.hardCoded.installAreaDirName(),
-                       self.hardCoded.shareAreaDirName())
-
-    def macrosFromCMT(self, project):
-        raise NotImplementedError
-    
-    def makePackages(self):
-        raise NotImplementedError
-
-#===================================================================
-
-class UserStuffRetrieverLocal(UserStuffRetrieverBase):
-    def __init__(self, paths):
-        UserStuffRetrieverBase.__init__(self, paths)
-
-    def macrosFromCMT(self, project):
-        self.cmds.extend(['cmt broadcast cmt -quiet show macros'])
-        
-        logger.debug('CMT query cmd: %s' % str(self.cmds))
-        print 'CMT query cmd: %s' % str(self.cmds)
-        sc = ShellCommand(self.cmds, logger)
-        macros = sc.getReply()
-        
-        return macros
-
-    def makePackages(self):
-        macros = self.readMacros('all')
-        self.packages = self.getPackageInfo(macros) # list of CVSPackage objects
-
-    def makePackageSharePath(self, dict):
-        return RTTpath(dict['_root'], self.hardCoded.shareAreaDirName())
-#===================================================================        
-
-class UserStuffRetrieverNonLocal(UserStuffRetrieverBase):
-    def __init__(self, paths):
-        UserStuffRetrieverBase.__init__(self, paths)
-
-    def cmdsToSourceProjects(self, project):
-        if project == 'all':
-            cmtdirs = allProjectCMTDirs(self.projectCMTDirs)
-        else:
-            cmtdirs = [pathToProjectCMTDir(project, self.projectCMTDirs)]
-            
-        
-        cwd = os.getcwd()
-
-        cmd = []
-        for dir in cmtdirs:
-            cmd.append('cd %s' % dir)
-            cmd.append('cmt -quiet show macros')
-        cmd.append('cd %s' % cwd)
-
-        logger.info('Command fragment to source all projects:\n%s' % str(cmd))
-        return cmd
-        
-    def macrosFromCMT(self, project):
-        self.cmds.extend(self.cmdsToSourceProjects(project))
-
-        logger.debug('macrosFromCMT - issuing command %s', str(self.cmds))
-        sc = ShellCommand(self.cmds, logger)
-        return sc.getReply()
-
-    def makePackages(self):
-        macros = self.readMacros('all')
-        self.packages = self.getPackageInfo(macros) # list of CVSPackage objects
-
-        # overwrite the package if there is a patch package in the top package
-        macros = self.readMacros(self.topProject)
-        self.packages.update(self.getPackageInfo(macros))
-        
-#===================================================================
-
-
-
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/UserTestDescriptor.py b/Tools/RunTimeTester/src/UserTestDescriptor.py
deleted file mode 100755
index 63143e1c8d754ac4ffa1339b101303d3d7dbfd38..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/UserTestDescriptor.py
+++ /dev/null
@@ -1,24 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from formatCollection import formatCollection
-
-import logging
-logger = logging.getLogger('rtt')
-smallLogger = logging.getLogger('rtt.standalone')
-# -------------------------------------------------------------------------
-
-class UserTestDescriptor:
-    def __init__(self, runInAtlasEnv, pos, moduleName, testName, argDict):
-        self.moduleName    = moduleName
-        self.testName      = testName
-        self.argDict       = argDict
-        self.position      = str(pos)
-        self.runInAtlasEnv = str(runInAtlasEnv)
-
-    def addParameter(self, item):
-        self.argDict[str(item[0])] = item[1]
-
-    def __str__(self):
-        str  = 'TestDescriptor: module name %s test name %s \n' % (self.moduleName, self.testName)
-        str += 'argDict: \n%s' % formatCollection(self.argDict)
-        return str
diff --git a/Tools/RunTimeTester/src/Utilities/FileCounter.py b/Tools/RunTimeTester/src/Utilities/FileCounter.py
deleted file mode 100755
index 1636b4c09ed0d04fd944e5fc20fb30bd51822d21..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Utilities/FileCounter.py
+++ /dev/null
@@ -1,77 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from visitor import FileVisitor
-import os
-
-
-class FileCounter(FileVisitor):
-
-    def __init__(self, fnFrag, grandParFrag, data=None, listonly=0):
-
-        FileVisitor.__init__(self, data, listonly)
-        
-        self.nFiles       = {}
-        self.fnFrag       = fnFrag
-        self.grandParFrag = grandParFrag
-        print 'Looking for filenames containing ', self.fnFrag
-        print '... with grandparent directory containing', self.grandParFrag
-        
-    def visitor(self, data, dirPath, filesInDir):
-
-        parentDir, temp = os.path.split(dirPath)
-        parentDir, temp = os.path.split(parentDir)
-        temp, parentDir = os.path.split(parentDir)
-        # print parentDir
-        if parentDir.find(self.grandParFrag) != -1:
-            # print '........',parentDir
-            # print '...........',dirPath
-
-            fileCount = 0
-            for file in filesInDir:
-                if file.find(self.fnFrag) != -1: fileCount += 1
-
-            # print ' has psFiles: ',fileCount
-            self.nFiles[dirPath]=fileCount
-
-def printit(fileCounter):
-
-    items   = fileCounter.nFiles.items()
-
-    nzero = 0
-    for item in items:
-
-        if int(item[1]) == 0:
-            print item
-            nzero+=1
-
-    print
-    
-
-    nsome=0
-    for item in items:
-
-        if int(item[1]) != 0:
-            print item
-            nsome+=1
-
-    print 'Looked for files containing ', fileCounter.fnFrag, 'in the filename, if these had'
-    print fileCounter.grandParFrag,' in the grandparent directory'
-    print 'total:         ',nsome+nzero
-    print 'directories with   no such files:' ,nzero
-    print 'directories with some such files:' ,nsome
-    
-
-
-def afCountFiles(fnFrag):
-    fc = FileCounter(fnFrag, 'AthenaAtlfast')
-    fc.run('.')
-    printit(fc)
-
-def rcCountFiles(fnFrag):
-    fc = FileCounter(fnFrag, 'RuleChecker')
-    fc.run('.')
-    printit(fc)
-
-if __name__ == '__main__':
-
-    afCountFiles('.ps')
diff --git a/Tools/RunTimeTester/src/Utilities/dirDiffer.py b/Tools/RunTimeTester/src/Utilities/dirDiffer.py
deleted file mode 100755
index e1bbfcec7c5623f001c4442aea0cd12271fd720e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Utilities/dirDiffer.py
+++ /dev/null
@@ -1,86 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import filecmp
-import sys
-import os.path
-
-
-totalLeftOnly  = []
-totalRightOnly = []
-totalDiffer    = []
-
-def cmpDir(left, right):
-
-    
-    dirlist = os.listdir(right)
-    dirlist.extend(os.listdir(left))
-
-    
-    ignore = [file for file in dirlist if file.find('.pyc') != -1]
-    ignore.extend([file for file in dirlist if file.find('~') != -1])
-    ignore.append('CVS')
-    ignore.append('Config')
-
-    ignoreDict = {}
-    for file in ignore:
-        ignoreDict.setdefault(file, 1)
-
-    ignore = ignoreDict.keys()
-
-    dc = filecmp.dircmp(left, right, ignore)
-    
-    print "files in  ",left,' and not in ', right,":"
-    print dc.left_only
-    totalLeftOnly.extend(dc.left_only)
-    
-    print "files in  ",right,' and not in ', left,":"
-    print dc.right_only
-    totalRightOnly.extend(dc.right_only)
-    
-    print "files in  ",right,' and not in ', left,":"    
-    print "files in  both, but which differ:"
-    print dc.diff_files
-    totalDiffer.extend(dc.diff_files)
-    
-    print "common dirs:"
-    print dc.common_dirs
-
-    for dir in dc.common_dirs:
-        newRight= os.path.join(right,dir)
-        newLeft = os.path.join(left,dir)
-
-        cmpDir(newRight, newLeft)
-
-
-if len(sys.argv) != 3:
-    print "Give 2 directories as arguments"
-    sys.exit(0)
-
-right = sys.argv[1]
-if not (os.path.exists(right) or os.path.isdir(right)):
-    print right,' is not a directory'
-    sys.exit(0)
-    
-left = sys.argv[2]
-if not (os.path.exists(left) or os.path.isdir(left)):
-    print right,' is not a directory'
-    sys.exit(0)
-
-cmpDir(left, right)
-
-
-
-print
-print "Total left only:"
-print totalLeftOnly
-print
-
-print
-print "Total right only:"
-print totalRightOnly
-print
-
-print
-print "Total common but differ:"
-print totalDiffer
-print
diff --git a/Tools/RunTimeTester/src/Utilities/rttCheck.py b/Tools/RunTimeTester/src/Utilities/rttCheck.py
deleted file mode 100755
index 96a6228d16081fe52ded688ae109cf3bda73f426..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Utilities/rttCheck.py
+++ /dev/null
@@ -1,7 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from FileCounter import *
-
-afCountFiles('.ps')
-afCountFiles('core')
-rcCountFiles('viol')
diff --git a/Tools/RunTimeTester/src/Utilities/visitor.py b/Tools/RunTimeTester/src/Utilities/visitor.py
deleted file mode 100755
index 427eaafaedbe3566140ddde23efb18596b7a5012..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/Utilities/visitor.py
+++ /dev/null
@@ -1,89 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-Taken from Lutz, programming python 2nd edition p 211.
-
-The classes in the module wrap os.path.walk
-
-The classes in this module are to be overridden to obtain desired behaviour
-"""
-
-import os, sys, string
-listonly = 0
-
-class FileVisitor:
-    """
-    visits all non-directory files below startDir;
-    overide visitfile to provide a file handler;
-    """
-
-    def __init__(self,data=None, listonly=0):
-
-        self.context  = data
-        self.fcount   = 0
-        self.dcount   = 0
-
-        self.listonly = listonly
-
-    def run(self, startDir=os.curdir):
-        os.path.walk(startDir, self.visitor, None)
-
-    def visitor(self, data, dirName, filesInDir):
-        
-        self.visitDir(dirName)
-        for fname in filesInDir:
-            fpath = os.path.join(dirName, fname)
-            if not os.path.isdir(fpath):
-                self.visitfile(fpath)
-                
-    def visitDir(self, dirPath):
-        self.dcount += 1
-        print dirPath,'...'
-
-    def visitfile(self, filepath):
-        self.fcount += 1
-        print self.fcount, '=>', filepath
-
-class SearchVisitor (FileVisitor):
-
-    "search files at and below startDir for a string"
-
-    skipexts = ['.gif', '.exe', '.pyc', '.o', '.a']
-
-    def __init__(self, key, listonly=0):
-
-        FileVisitor.__init__(self)
-        self.scount = 0
-
-    def visitfile(self, fname):
-        FileVisitor.visitfile(self, fname)
-        if not self.listonly:
-            if os.path.splitext(fname)[1] in self.skipexts:
-                print 'skipping', fname
-            else:
-                text = open(fname).read()
-                if string.find(text, self.context) != -1:
-                    self.visitmatch(fname, text)
-                    self.scount += 1
-
-    def visitmatch(self, fname, text):
-        raw_input('%s has %s' % (fname, self.context))
-
-# self-test logic
-
-dolist   = 1
-dosearch = 2
-donext   = 4
-
-def selftest(testmask):
-    if testmask & dolist:
-        visitor=FileVisitor()
-        visitor.run('.')
-        print 'Visited %d files and %d directories' % (visitor.fcount,
-                                                       visitor.dcount)
-
-if __name__ == '__main__':
-
-    selftest(int(sys.argv[1]))
-            
-        
diff --git a/Tools/RunTimeTester/src/VersionedFileSaver.py b/Tools/RunTimeTester/src/VersionedFileSaver.py
deleted file mode 100755
index 85ee27e0feb75e3c64e7354e6f6d649635d6d3a8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/VersionedFileSaver.py
+++ /dev/null
@@ -1,118 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTSException import RTTCodingError
-from types         import IntType
-
-import os, shutil, logging
-
-logger=logging.getLogger('rtt')
-
-def isInt(a):
-    try:
-        i = int(a)
-        return True
-    except:
-        return False
-
-def stripVersion(fn, sep):
-    (dir, file) = os.path.split(fn)
-    file = file.split(sep)[0]
-    return file
-    
-def ctimeCmp(a, b):
-    return cmp(os.path.getctime(a), os.path.getctime(b))
-        
-class VersionedFileSaver:
-    def __init__(self, maxVersion, destDir=None):
-        if not type(maxVersion) == IntType:
-            msg = 'VersionedFileSaver contructor arg not an int: %s ' % str(maxVersion)
-            raise RTTCodingError(msg)
-        self.maxVersion = maxVersion;
-        if destDir:
-            self.destDir = destDir
-        else:
-            self.destDir = os.getcwd()
-
-    def getVersionedFiles(self, fn, sep):
-
-        vFiles = []
-        if not os.path.exists(fn): return vFiles
-
-        if not os.path.isfile(fn): return vFiles
-
-        # remove version if any from file name
-        noVerFile = stripVersion(fn, sep)
-
-        # look for versioned files: these have the form xxx;n
-        # where n is in integer
-        (dir, vfile)= os.path.split(noVerFile)
-        vfile += sep
-
-        vFiles = os.listdir(self.destDir)
-        vFiles = [f for f in vFiles if f.find(vfile) >=0]
-        vFiles = [f for f in vFiles if f.count(sep) == 1]
-
-
-        vFiles = [f for f in vFiles if isInt(f.split(sep)[1])]
-        vFiles = [os.path.join(self.destDir, f) for f in vFiles]
-
-        return vFiles
-
-    
-
-    def saveFile(self, fn):
-        sep = '-'
-        self.vFiles = self.getVersionedFiles(fn, sep)
-        nextVersion = -1
-
-        # find the last (in time) created file
-        self.vFiles.sort(ctimeCmp)
-
-        if self.vFiles:
-            lastFile = os.path.basename(self.vFiles[-1])
-            nextVersion = (int(lastFile.split(sep)[1])+1) % self.maxVersion
-        else:
-            nextVersion = 1
-
-
-        noVerFile = stripVersion(fn, sep)
-        toFile = noVerFile+sep+str(nextVersion)
-        if self.destDir:
-            toFile =  os.path.join(self.destDir, os.path.basename(toFile))
-
-
-        try:
-            shutil.copyfile(fn, toFile)
-        except (IOError, os.error), why:
-            logger.error("Cannot copy  %s to %s: %s" % (fn, toFile, str(why)))
-        
-if __name__ == '__main__':
-
-
-    fn = './temp'
-    file = open(fn, 'w')
-
-    file.write('junk\n')
-    file.close()
-
-    try:
-        vfs = VersionedFileSaver(4)
-    except:
-        print 'caught bad VSF construction'
-
-
-    destDir = os.environ['HOME']
-    # vfs = VersionedFileSaver(8,destDir)
-    vfs = VersionedFileSaver(4)
-    print vfs.destDir
-    nvers = 1
-    while nvers > 0:
-        nvers -= 1
-        vfs.saveFile(fn)
-            
-    for f in vfs.vFiles:
-        print f,os.path.getctime(f)
-        
-
-        
-        
diff --git a/Tools/RunTimeTester/src/WWWebPageWriter.py.obsolete b/Tools/RunTimeTester/src/WWWebPageWriter.py.obsolete
deleted file mode 100755
index cd8ae6370c18ccbd001e997e0f1bd386cd467536..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WWWebPageWriter.py.obsolete
+++ /dev/null
@@ -1,387 +0,0 @@
-# This class acts as the entrance to the system that writes out the xml files
-# used to generate web pages.
-# It manages incomoming summaries, establishes if the web summary is more recent than
-# others with the same parameter set.
-#
-# It then wraps most recent the summaries into a DOM document and passes them to PageWriter.
-# the document is destroyed after the web pages have been written.
-#
-
-import os, types, string
-import time
-import page0DefaultFiles
-
-from PageWriter import Page1Writer
-from xml.dom.minidom            import parseString, parse
-
-from Evaluate  import Evaluate
-from listFiles import listFiles
-from XMLTools  import relPath
-
-from exc2string2 import exc2string2
-# print level: current settings are at 1, 10, 100. 0 = silent.
-debug = 10
-
-def setPageLocations(dom):
-
-    def makeShorterPath(path, backup):
-
-        while backup:
-            backup -= 1
-            path, tail = os.path.split(path)
-
-        return path
-
-
-    # hardcode the location of page 2.
-    # original idea was to have the source of all path info
-    # from minder results paths. Does not work as some packages (vetoed eg)
-    # have no minders.
-
-    absPage2Path = os.path.join(Evaluate('overview/resReleaseDir/text()', dom.documentElement), 'page2.xml')
-    
-    # absPage1Path = os.path.join(makeShorterPath(absPage2Path, 3), 'page1.xml')
-    # absPage0Path = os.path.join(makeShorterPath(absPage1Path,4), 'page0.xml')
-
-    absPage1Path = os.path.join(makeShorterPath(absPage2Path, 6), 'page1.xml')
-
-    # print 'absPage1Path',absPage1Path
-    # print absPage0Path
-    # relPage0Page1Path = relPath(absPage0Path, absPage1Path)
-    relPage1Page2Path = relPath(absPage1Path, absPage2Path)
-
-
-    newPaths = [
-        # ('absPage0Path',      absPage0Path),
-        ('absPage1Path',      absPage1Path),
-        ('absPage2Path',      absPage2Path),
-        # ('relPage0Page1Path', relPage0Page1Path),
-        ('relPage1Page2Path', relPage1Page2Path)
-        ]
-
-    addedPaths = []
-    addedPaths.extend(newPaths)
-
-    # for p in newPaths: print '%20s: %s' % p
-    
-    on = Evaluate('overview', dom.documentElement)[0]
-
-    def createPathNode(name, path, dom):
-        el = dom.createElement(name)
-        el.appendChild(dom.createTextNode(path))
-        return el
-    
-    [on.appendChild(createPathNode(name, path, dom)) for name, path in newPaths]
-
-    
-    packageNodes = dom.getElementsByTagName('package')
-    for pn in packageNodes:
-        minderNodes = pn.getElementsByTagName('minder')
-        
-        # sometimes package has no minder eg when package in vetoed
-        # page 3 contains the list of pointers to all the page4s (jobs) for
-        # a package.
-        if not minderNodes: continue
-
-        def findPackageLocation(minderNode):
-            path = Evaluate('resultsPath/text()', minderNode)
-            path = makeShorterPath(path, 3)
-            return path
-        
-        packageLocations = [findPackageLocation(mn) for mn in minderNodes]
-        # print pn.toxml()
-        # assert len(packageLocations)>0
-
-            
-        absPage3Path = packageLocations[0]
-        for p in packageLocations: assert absPage3Path == p
-
-        absPage3Path = os.path.join(absPage3Path, 'page3.xml')
-        relPage2Page3Path = relPath(absPage2Path, absPage3Path)
-        newPaths = [
-            ('absPage3Path',      absPage3Path),
-            ('relPage2Page3Path', relPage2Page3Path)
-            ]
-        [pn.appendChild(createPathNode(name, path, dom)) for name, path in newPaths]
-        addedPaths.extend(newPaths)
-
-        for mn in minderNodes:
-            absPage4Path = Evaluate('resultsPath/text()', mn)
-            absPage4Path = os.path.join(absPage4Path, 'page4.xml')
-            relPage3Page4Path = relPath(absPage3Path, absPage4Path)
-            mn.appendChild(createPathNode('relPage3Page4Location', relPage3Page4Path, dom))
-
-
-    return addedPaths
-         
-    
-class SummaryLocationKey:
-    def __init__(self, path):
-        self.keyType = 'relativePath'
-        self.key     = path
-
-def setupSummaryDict(topDir, summaryListFileName):
-    """
-    set up a dictionary of location: document pairs
-    location is  a general concept, here represented by a relative path
-
-    the locations of the summary files are stored in the file summaryFilePaths.txt
-    for rapid access. If this file does not exist, the file system is searched
-    for the summaryFiles.
-    """
-    fn = 'summaryFilePaths.txt'
-
-    if os.path.exists(fn):
-        summaryPaths=open(fn, 'r').readlines()
-        summaryPaths = [s.strip() for s in summaryPaths]
-        # print 'read %d locations from %s' % (len(summaryPaths), fn)
-    else:
-        summaryPaths =  listFiles(topDir, 'RTTSummary.xml')
-        of = open(summaryListFileName, 'w')
-        [of.write('%s\n' % l) for l in summaryPaths]
-        of.close()
-
-    summaryPaths = [s for s in summaryPaths if os.path.exists(s)]
-    # print 'len summaryPaths', len(summaryPaths)
-    return summaryPaths
-
-import mutex
-
-accessLock = mutex.mutex()
-
-class WebPageWriterBase:
-    def __init__(self, resBaseDir, initTables, writePages):
-
-        #logger.debug('WebPageWriter __init__ start')
-
-        self.totPageUpdates = 0
-        self.summaryListFileName = os.path.join(os.getcwd(), 'summaryFilePaths.txt')
-
-        self.docs = [] # path names to summary files
-        self.doms = {} # identifier to dom mapping
-
-        # On request, look for all summary files below resBaseDir
-        if initTables: self.docs = setupSummaryDict(resBaseDir, self.summaryListFileName)
-        
-        self.resBaseDir = resBaseDir
-        self.writer = self.makeWriter()
-        # ensure that some enssential web files are present
-        
-        filesToWrite = {
-            'messageOfTheDay.js':        page0DefaultFiles.messageOfTheDay,
-            'RTTpage1_LinksAtTopOfPage.js': page0DefaultFiles.RTTpage1_LinksAtTopOfPage
-            }
-
-        def writeOut((fn, text), resBase):
-            fullPath = os.path.join(resBase, fn)
-            if os.path.exists(fullPath): return
-            of = open(fullPath,'w')
-            of.write(text)
-            of.close()
-
-        [writeOut(item, resBaseDir) for item in filesToWrite.items()]
-
-        self.updates = []
-
-        ndocs = len(self.docs)
-        ndoc  = 0
-        for s in self.docs:
-            ndoc += 1
-            print 'processing  %s' % (s)
-            self.internalUpdateSingleSummary(s, writePages)
-            print '%d/%d pages complete: last = %s' % (ndoc, ndocs, s)
-
-    def internalUpdateSingleSummary(self, pathToSummaryFile, writePage):
-
-        
-        try:
-            xmlSummary = parse(pathToSummaryFile)
-            # print '\n %s::internalUpdateSingleSummary - parsed string from file\n  %s' % (self.__class__.__name__,
-            #                                                                              pathToSummaryFile)
-        except:
-            print exc2string2()
-            # print open(pathToSummaryFile, 'r').read()
-            print 'could not parse string from file %s' % pathToSummaryFile
-            return 1
-        
-        identifier = Evaluate('overview/startTime1/text()', xmlSummary.documentElement).strip()
-
-        argdict = {'packagesToUpdate':   [],
-                   'mindersToUpdate':    [],
-                   'cmttargetsToUpdate': [],
-                   'writePages':         writePage,
-                   'identifier':         identifier,
-                   'xmlSummary':         xmlSummary,
-                   'xmlSummaryLocation': pathToSummaryFile,
-                   'finalise':           True
-                   }
-
-        self.doms[identifier] = xmlSummary
-        self.update(argdict)
-
-        # self.router.deregisterSummary(identifier)
-        return 0
-    
-    def updateSingleSummary(self, pathToSummaryFile):
-        'Method to force reading from summary from given path and to write webPages'
-        rc = self.internalUpdateSingleSummary(pathToSummaryFile, True)
-        if rc !=0: return rc
-        
-        try:
-            of = open(self.summaryListFileName, 'r')
-        except:
-            return 2
-
-        lines = of.readlines()
-        of.close()
-        found = False
-        for l in lines:
-            if l.startswith(pathToSummaryFile):
-                of.close()
-                return 0
-
-        lines.append(pathToSummaryFile+'\n')
-        
-
-        text =string.join(lines)
-
-        try:
-            of = open(self.summaryListFileName, 'w')
-        except:
-            return 3
-
-        of.write(text)        
-        of.close()
-        return 0
-    
-    def update(self, dict):
-        if debug>=50: print time.ctime(), 'WebPageWriter start update'
-        startTime = time.ctime()
-        timePair = (startTime,0)
-        self.updates.append(timePair)
-        accessLock.lock(self.lockedUpdate, dict)
-
-        timePair = (startTime,time.ctime())
-        self.updates[-1] = timePair
-        if debug >=50: print time.ctime(), 'WebPageWriter ends update'
-        
-        return 0
-    
-    def lockedUpdate(self, dict):
-        if debug>=100: print 'start lockedUpdate'
-
-        document = dict['xmlSummary']
-        if isinstance(document, types.StringTypes): # strings passed from Tester, DOMS from __init__
-            try:
-                document           = parseString(dict['xmlSummary'])
-            except:
-                if debug >= 10: print dict['xmlSummary']
-
-            
-        xmlSummaryLocation = relPath(self.resBaseDir, dict['xmlSummaryLocation'])
-        packagesToUpdate   = dict['packagesToUpdate']
-        mindersToUpdate    = dict['mindersToUpdate']
-        cmttargetsToUpdate = dict['cmttargetsToUpdate']
-        writePages         = dict['writePages']
-        summaryIdentifier  = dict['identifier']
-        finalise           = dict.get('finalise', False)
-
-
-        # print 'Calling setPageLocations start'
-        try:
-            npaths = setPageLocations(document)
-        except:
-            print 'Error setting path locations'
-        # else:
-        #    for np in npaths: print np
-
-        # print document.toxml()
-
-        # print 'Calling setPageLocations end'
-                
-        self.doms[summaryIdentifier] = document
-
-        de = document.documentElement
-        overviewNode = Evaluate('overview', de)[0]
-        # Some pages containing the overview node have the lifetime of the server.
-        # Once a document is processed, it is deleted to avoid running out of memory.
-        # The overview node needs to be cloned to avoid its destruction while in a live page
-        clonedOverview =overviewNode.cloneNode(True)
-        nupdates = self.writer.update(cmttargetsToUpdate, packagesToUpdate, mindersToUpdate, writePages,
-            xmlSummaryLocation, clonedOverview, de)
-
-        self.totPageUpdates += nupdates
-        if debug >= 10: print 'WebPageWriter: pages updated  this doc: %d total %d' % (nupdates, self.totPageUpdates)
-        if finalise:
-            if debug >= 20: print 'Deregistration starts'
-            self.writer.deregisterChildren()
-            document.unlink()
-            del self.doms[summaryIdentifier]
-        
-        accessLock.unlock()
-        return 0
-
-    def getDocs(self):
-        if self.docs: return self.docs
-        return 0
-
-    def countDoms(self):
-        print len(self.doms)
-        return 0
-
-    def getOutputDir(self): return self.resBaseDir
-
-    def areYouAlive(self): return 'yes'
-    
-    def dumpTable(self):
-        # return '\n---- WebPageWriter Tables ---\n\n%s\n\n%s\n\n%s\n\n%s\n\n' % (
-        #    str(self.page0Table),
-        #    str(self.page1Table),
-        #    str(self.page2Table),
-        #    str(self.page3Table)
-        #    )
-        counts = {}
-        s = str(self.updates) + '\n\n' + self.writer.dump(0, counts)
-        s += '\n\ncounts by page type \n\n'
-
-        ks = counts.keys()
-        ks.sort()
-        for k in ks:
-            s += '%10s %d\n' % (k, counts[k])
-        print s
-        return 0
-
-class WebPageWriter(WebPageWriterBase):
-    def __init__(self, resBaseDir, initTables=True, writePages=False):
-
-
-        if debug>=100: print 'WebPageWriter: start construction: initTable:', initTables,'writePages:',writePages
-        ### self.writerClass = Page0Writer
-        
-        WebPageWriterBase.__init__(self, resBaseDir, initTables, writePages)
-
-        if debug>=100: print 'WebPageWriter: construction complete'
-
-    def makeWriter(self):
-        # return Page0Writer(self.resBaseDir, 'Top', None, None)
-        return Page1Writer(self.resBaseDir, 'Top')
-
-if __name__ == '__main__':
-    from xml.dom.minidom import parse
-    import sys
-    
-    resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results'
-
-    # dict = {
-    #    'keyDocList':        docList,
-    #    'cmtTargetToUpdate': [],
-    #    # 'packageToUpdate':   ['BPhysValidation'],
-    #    'packageToUpdate':   [],
-    #    'mindersToUpdate':   []
-    #    }
-
-    writer = WebPageWriter(resBaseDir,True)
-
-    
-    print writer.dumpTable()
-    
diff --git a/Tools/RunTimeTester/src/WWebPageWriter.py.obsolete b/Tools/RunTimeTester/src/WWebPageWriter.py.obsolete
deleted file mode 100755
index 89e165a7c010c91bb3372641859fed1638d14c2e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WWebPageWriter.py.obsolete
+++ /dev/null
@@ -1,339 +0,0 @@
-# This class acts as the entrance to the system that writes out the xml files
-# used to generate web pages.
-# It manages incomoming summaries, establishes if the web summary is more recent than
-# others with the same parameter set.
-#
-# It then wraps most recent the summaries into a DOM document and passes them to PageWriter.
-# the document is destroyed after the web pages have been written.
-#
-
-import os.path
-
-import page0DefaultFiles
-
-from PageWriter import Page0Writer
-from PageWriter import summaryPacker
-from xml.dom.minidom            import parseString
-
-from Evaluate import Evaluate
-
-
-def selectKey(key, selection):
-    """
-    selection and key are lists if (name, value) tuples.
-    Return true if every such tuple in selection is in key
-    """
-
-    key = eval(key)
-
-    print key
-    print selection
-    for s in selection:
-        if s not in key: return False
-
-    return True
-
-
-
-def stripTimeFromKey(key):
-    key = eval(key)
-    dict = {}
-    for k, v in key: dict[k] = v
-    # time = dict['time']
-    del dict['time']
-    nkey = dict.values()
-    nkey.sort()
-    nkey = nkey.__repr__()
-    # return (time, nkey)
-    return nkey
-
-def getDocumentTime(doc):
-    """
-    fomd the time stamp of the document. older docuemtns do not
-    have a latestUpdateTime in the overview - in which case use the
-    start time.
-    """
-    
-    try:
-        docTime = float(Evaluate('overview/latestUpdateTime2/text()', doc.documentElement))
-    except:
-        docTime = float(Evaluate('overview/startTime2/text()', doc.documentElement))
-
-    return docTime
-
-import mutex
-
-accessLock = mutex.mutex()
-
-class PageTable:
-    def __init__(self):
-        # data structure: {xpath: {location: page} } to prevent cumulation of old writers
-        self.dict = {}
-    def register(self, parentWriter, writer):
-
-        key = parentWriter.name
-        # print 'registering writer %s  at %s  for parent %s' % (writer.name, writer.location, key)
-        dict = self.dict.setdefault(key, {})
-        
-        # stop memory leaks
-        try:
-            oldWriter = dict[writer.location]
-        except:
-            # no entry yet for this location
-            pass
-        else:
-            # oldWriter.unlink()
-            pass 
-        #fill in table and return
-        dict[writer.location] = writer
-        # print 'registered writer %s at location %s' % (writer.name, key)
-        return
-
-    def __getitem__(self, key): return self.dict[key].values()
-
-    def __str__(self):
-        s = '\n\n'
-        for k in self.dict.keys():
-            s += '\n%s\n' %k 
-            s += '   %s\n' % str([w.name for w in self.dict[k].values()])
-        return str(s)
-    
-class PageTableRouter:
-    def __init__(self, dict):
-        self.dict = dict
-    def register(self, parentWriterName, writer):
-
-        # place the writer in the doctionary. One writer per file location.
-        dict = self.dict.setdefault(parentWriterName, {})
-        dict[writer.location] = writer
-
-    def registeredWriters(self, writerName):
-        'returns ALL the child writers for the writer passed in.'
-        return (self.dict[writerName]).values()
-
-class WebPageWriterBase:
-    def __init__(self, resBaseDir):
-
-        #logger.debug('WebPageWriter __init__ start')
-        self.docs = {}
-        self.resBaseDir = resBaseDir
-        # set up tables to store pages.
-        # the update algorithm is to update only thos pages that are affected.
-        # for a  minder that is updated: the page three for the package is updated,
-        # the page 2 for the cmtTarget is updated, the page1 is updated. Page 0
-        # is updated if the minder corresponds to a new cmtTarget.
-        #
-        # each table holds the pages for a single class of page writer. it is a dictionary
-        # of dictionaries. The key into the frst dictionary is the xpath to get to the
-        # nodes for the next page writer, the inner dictionary key is the page location, and
-        # the value is the page.
-        #self.page0Table = PageTable()
-        # self.page1Table = PageTable()
-        # self.page2Table = PageTable()
-        #self.page3Table = PageTable()
-        self.pageTable = {}
-        # ensure that some enssential web files are present
-        
-        filesToWrite = {
-            'currentStatusN0X.js':       page0DefaultFiles.currentStatusN0X,
-            'currentStatusNX0.js':       page0DefaultFiles.currentStatusNX0,
-            'messageOfTheDay.js':        page0DefaultFiles.messageOfTheDay,
-            'page0.html':                page0DefaultFiles.page0,
-            'page0.noTopFrame.html':     page0DefaultFiles.page0_noTopFrame,
-            'status.html':               page0DefaultFiles.status,
-            'page0.xsl':                 page0DefaultFiles.page0XSL,
-            'page1.xsl':                 page0DefaultFiles.page1XSL,
-            'page2.xsl':                 page0DefaultFiles.page2XSL,
-            'page3.xsl':                 page0DefaultFiles.page3XSL,
-            'page4.xsl':                 page0DefaultFiles.page4XSL,
-            'RTTpage1_LinksAtTopOfPage.js': page0DefaultFiles.RTTpage1_LinksAtTopOfPage
-            }
-
-        def writeOut((fn, text), resBase):
-            fullPath = os.path.join(resBase, fn)
-            if os.path.exists(fullPath): return
-            of = open(fullPath,'w')
-            of.write(text)
-            of.close()
-
-        [writeOut(item, resBaseDir) for item in filesToWrite.items()]
-
-
-    def update(self, dict):
-        print 'start update'
-        accessLock.lock(self.lockedUpdate, dict)
-        return 0
-    
-    def lockedUpdate(self, dict):
-        print 'start lockedUpdate'
-        
-        keyDocList         = dict['keyDocList']
-        cmtTargetToUpdate  = dict['cmtTargetToUpdate']
-        packageToUpdate    = dict['packageToUpdate']
-        mindersToUpdate    = dict['mindersToUpdate']
-
-        # print '\n------------- lockedUpdate starts for ----------------\n'
-        # print 'no of docs', len(keyDocList)
-        # print 'cmtTarget',  cmtTargetToUpdate
-        # print 'package',    packageToUpdate
-        # print 'minders',    mindersToUpdate
-        # print '\n------------- lockedUpdate starts for ----------------\n'
-
-
-        class ParseIt:
-            def __init__(self):
-                self.ncalls = 0
-
-            def run(self, k, xml):
-                print self.ncalls, k
-                self.ncalls += 1
-                return parseString(xml)
-        
-        # parseIt = ParseIt()
-        # keyDocList = [( k, parseIt.run(k, xml)) for k, xml  in keyDocList]
-        keyDocList = [( k, parseString(xml)) for k, xml  in keyDocList]
-        
-
-        newDocs = {}
-        for key, ndoc in keyDocList:
-            # ntime is latest update time if available, start time otherwise.
-            ntime = getDocumentTime(ndoc)
-            nkey  = stripTimeFromKey(key)
-
-            if nkey not in self.docs:
-                self.docs[nkey] = (ntime, ndoc)
-                newDocs[nkey] = ndoc
-                # print 'new doc, key = ', nkey 
-            else:
-                otime, odoc  = self.docs[nkey]
-                # if ntime not latestUpdateTime, the web pages will not update.
-                # this should not cause a problem because all documents that
-                # are unfinished, ir need updating have a latestUpdate time
-                # using ntime>otime with ntime = startTime stops updates
-                #       ntime>=otime with ntime = startTime causes unnecessary reinitialisation
-                # when a new RTT instance sends a database load of data.
-                #
-                if ntime>otime:
-                    self.docs[nkey] = (ntime, ndoc)
-                    newDocs[nkey] = ndoc
-
-        newKeys = newDocs.keys()
-        newDocs = newDocs.values()
-        # print 'found %d new docs' % len(newDocs)
-        # for k in newKeys: print k
-        # return 0
-        packedSummaries = summaryPacker(newDocs) 
-        writer = self.writerClass(self.router, None, self.resBaseDir, 'Top', packedSummaries.documentElement )
-        writer.update(cmtTargetToUpdate, packageToUpdate, mindersToUpdate)
-        # print 'returning from WebPageWriter.update'
-
-
-        # print '\n------------- lockedUpdate ends for ------------------\n'
-        # print 'lockedUpdate ends for: '
-        # print 'no of docs', len(keyDocList)
-        # print 'cmtTarget',  cmtTargetToUpdate
-        # print 'package',    packageToUpdate
-        # print 'minders',    mindersToUpdate
-        # print '\n------------- lockedUpdate ends for ------------------\n'
-        
-        accessLock.unlock()
-        return 0
-
-    def getDocs(self):
-        if self.docs: return self.docs
-        return 0
-
-    def getOutputDir(self): return self.resBaseDir
-
-    def dumpTable(self): return self.__str__()
-
-    def areYouAlive(self): return 'yes'
-    
-    def __str__(self):
-        # return '\n---- WebPageWriter Tables ---\n\n%s\n\n%s\n\n%s\n\n%s\n\n' % (
-        #    str(self.page0Table),
-        #    str(self.page1Table),
-        #    str(self.page2Table),
-        #    str(self.page3Table)
-        #    )
-        return '\n---- WebPageWriter Tables ---\n\n%s\n\n' % str(self.pageTable)
-
-class WebPageWriter(WebPageWriterBase):
-    def __init__(self, resBaseDir):
-        WebPageWriterBase.__init__(self, resBaseDir)
-        self.writerClass = Page0Writer
-        # a class to route the table to the appropiate table.
-        # self.router     = PageTableRouter({
-        #     'Page0Writer': self.page0Table,
-        #    'Page1Writer': self.page1Table,
-        #    'Page2Writer': self.page2Table,
-        #    'Page3Writer': self.page3Table
-        #   })
-
-        self.router = PageTableRouter(self.pageTable)
-        
-class CERNWebPageWriter(WebPageWriter):
-    """
-    selects RTT summaries made at CERN.
-    """
-    def lockedUpdate(self, dict):
-        keyDocList         = dict['keyDocList']
-
-        selection = [('site', 'CERNslc4')]
-
-
-        docList = [(k,d) for k,d in keyDocList if selectKey(k, selection)]
-        if not docList: return 0
-        
-        dict['keyDocList'] = docList
-
-        return WebPageWriter.lockedUpdate(self, dict)
-        
-if __name__ == '__main__':
-    from xml.dom.minidom import parse
-    import sys
-    
-    fn = 'testShort.xml'
-
-    xml = open(fn, 'r').read()
-    #dom = parse(fn)
-    #de = dom.documentElement
-    #on = Evaluate('overview', de)[0]
-
-    mindersToUpdate = [
-        'Atlfast_POOL_False_high_kt_tracksPilot29',
-        'Atlfast_POOL_False_high_shared_tracksPilot81'
-        ]
-    packageToUpdate = 'TestAtlfast'
-
-    key = [
-        ('originalBranch',  'dev'),
-        ('release',         'rel_5'),
-        ('releaseType',     'dev'),
-        ('runType',         'build'),
-        ('targetCMTCONFIG', 'i686-slc3-gcc323-opt'),
-        ('time',             1172855233.26)
-        ]
-
-    key = key.__repr__()
-    docList = [(key, xml)]
-
-    resBaseDir = '/afs/cern.ch/atlas/RTT/data/Peter/Results'
-
-    dict = {
-        'keyDocList':        docList,
-        'cmtTargetToUpdate': [],
-        # 'packageToUpdate':   ['BPhysValidation'],
-        'packageToUpdate':   [],
-        'mindersToUpdate':   []
-        }
-
-    writer = WebPageWriterTest(resBaseDir)
-    # writer.update(docList, packageToUpDate, mindersToUpdate)
-    writer.update(dict)
-    print '----pass 2---'
-    writer.update(dict)
-
-    
-    print writer
-    
diff --git a/Tools/RunTimeTester/src/WWebPageWriterTest.py.obsolete b/Tools/RunTimeTester/src/WWebPageWriterTest.py.obsolete
deleted file mode 100755
index b0d0990e2529996f630132dfcd2525aa49643ab6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WWebPageWriterTest.py.obsolete
+++ /dev/null
@@ -1,15 +0,0 @@
-from WWebPageWriter import WebPageWriterBase, PageTableRouter
-from PageWriterTest import Page0WriterTest
-
-class WebPageWriterTest(WebPageWriterBase):
-    def __init__(self, resBaseDir):
-        WebPageWriterBase.__init__(self, resBaseDir)
-        self.writerClass = Page0WriterTest
-        # a class to route the table to the appropiate table.
-        self.router     = PageTableRouter({
-            'Page0WriterTest': self.page0Table,
-            'Page1WriterTest': self.page1Table,
-            'Page2WriterTest': self.page2Table,
-            'Page3WriterTest': self.page3Table
-            })
-            
diff --git a/Tools/RunTimeTester/src/WatcherJobDescriptor.py b/Tools/RunTimeTester/src/WatcherJobDescriptor.py
deleted file mode 100755
index 4ad1e14dd0338b1a8a68436edbf8f42014b5a394..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WatcherJobDescriptor.py
+++ /dev/null
@@ -1,43 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Class descriptor for Watcher jobs. Watcher jobs 
-iterate over some or all of the ATLAS offline release packages.
-They are not Athena jobs.
-
-Watcher descriptor has been found to need a lot of
-athena job characteristics: it should be split off
-into a derived class AthenaWatcherJobDescriptor....
-"""
-
-from Descriptor import Descriptor
-
-class WatcherJobDescriptor(Descriptor):
-
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-
-        Descriptor.__init__(self, paths, jobSerialNumber, logger)
-        if self.mode != 'Dummy': self.mode = 'Watcher'
-
-        # a list of worker job minders
-        self.confederation    = []
-        self.package          = paths.package
-
-        self.__dict__.update(argDict)
-
-        self.name = self.jobGroup
-
-        self.fixName()        
-        self.setPathExtensions()
-        self.setRunPath(self.pathExtensions)
-
-    def setPathExtensions(self):
-
-        # NOTE: we assume here that all watchers are watching athena jobs
-        #       i.e ones with a package. Probably wrong long term, but we leave
-        #       it as-is for now in order to go faster.
-        self.pathExtensions = [
-            ('package',    self.package),
-            ('jobGroup',   self.jobGroup),
-            ('jobName',    self.name),
-            ('SN',         str(self.jobSerialNumber))
-            ]
diff --git a/Tools/RunTimeTester/src/WatcherJobMinder.py b/Tools/RunTimeTester/src/WatcherJobMinder.py
deleted file mode 100755
index ffd8070b56224bea619a25ba78e564cd817cc781..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WatcherJobMinder.py
+++ /dev/null
@@ -1,108 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-JobMinder that watches other jobs
-"""
-from Minder                         import Minder
-from formatCollection               import formatCollection
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class WatcherJobMinder(Minder):
-    def __init__(self, argBag, jDescriptor):
-
-        self.jobsToWatch = jDescriptor.confederation
-        
-
-        Minder.__init__(self,argBag, jDescriptor, logger)
-
-    # =======================================================
-    def removeJob(self, job):
-        self.logger.debug('Removed job: '+job.identifiedName)
-        self.jobsToWatch.remove(job)
-
-    # =======================================================
-
-    def __str__(self):
-        s  = Minder.__str__(self)
-        names = [job.identifiedName for job in self.jobsToWatch]
-        s += 'job to watch.:      %s' % formatCollection(names)
-        return s
-
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|WatcherJobMinder  dump                     |')
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug(self.__str__()+'\n')
-        self.logger.debug(Minder.__str__(self)+'\n')
-
-    def jobIsRunning(self):
-        queryOK = True
-        isRunning = False
-        for j in self.jobsToWatch:
-            isRunning = True
-            if not j.isDone(): return (queryOK,isRunning)
-        return (queryOK,isRunning)
-   # -----------------------------------------------
-    
-    def fromQueued(self):
-        return 'running'
-            
-    # -----------------------------------------------
-    
-    def fromRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-        
-        if isRunning:  return 'running'# no state transition
-
-        self.fromRunningActions()
-
-        if self.doPostProcessing():
-            return 'postProcessingQueued'
-
-        return 'done'
-
-
-    # -----------------------------------------------    
-
-    def fromRunningActions(self):
-        'in a method to allow derived class extensions'
-
-        # self.minder.runMoniActions()
-        self.runChecks()
-
-        # chainFileCopier present only if running in a chain 
-        if self.chainFileCopier:
-            self.minder.chainFileCopier.copyFiles()
-
-        # if keep file is also a chain file, this code
-        # will silently fail. Intended that the chain
-        # file copier is interogated to find new location
-        self.copyKeepFiles()
-
-    # -----------------------------------------------
-    
-    def fromPostProcessingQueued(self):        
-        self.fire('postProcessorScript')
-        return 'postProcessingRunning'
-            
-    # -----------------------------------------------
-    
-    def fromPostProcessingRunning(self):
-        queryOK, isRunning =  self.jobIsRunning()
-        if isRunning:  return 'postProcessingRunning'# no state transition
-
-        self.minder.readTestResults()
-        self.minder.copyKeepFiles() # grab any new keep files
-
-        # make sure ALL files in run and results directory are
-        # accessible by all members of the UCL Atlas group
-        self.calcPostProcessingResult()
-
-        return 'done'
-        
-            
diff --git a/Tools/RunTimeTester/src/WatcherMinderStateEngine.py b/Tools/RunTimeTester/src/WatcherMinderStateEngine.py
deleted file mode 100644
index e41a99564eb990ee80a3442588a85c836dafdaaa..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WatcherMinderStateEngine.py
+++ /dev/null
@@ -1,29 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-from exc2string2            import exc2string2
-from MinderStateEngineBase  import MinderStateEngineBase
-
-import time
-
-class WatcherMinderStateEngine(MinderStateEngineBase):
-    """
-    Empty state machine. This will be stepped by calls to poll(),
-    and will end up in state = success. Different behaviour is obtained
-    by subclassing the xxxTest and xxxAction methods.
-    """
-
-    
-    def __init__(self, minder, state='queued'):
-        MinderStateEngineBase.__init__(self, state, minder.identifiedName)
-
-        self.dispatcher = {
-            'queued'               :  minder.fromQueued,
-            'running'              :  minder.fromRunning,
-            'operatorCloseDown'    :  minder.fromOperatorCloseDown,
-            'postProcessingQueued' :  minder.fromPostProcessingQueued,
-            'postProcessingRunning':  minder.fromPostProcessingRunning,
-            'done'                 :  minder.fromDone
-            }
-
- 
diff --git a/Tools/RunTimeTester/src/WebPageCollector.py.obsolete b/Tools/RunTimeTester/src/WebPageCollector.py.obsolete
deleted file mode 100755
index 09d7e2f3e4307017cf854e25bc55bdef2d00d567..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WebPageCollector.py.obsolete
+++ /dev/null
@@ -1,105 +0,0 @@
-import xmlrpclib, time, sys, getopt
-from exc2string2 import exc2string2
-
-
-python = '/afs/cern.ch/sw/lcg/external/Python/2.4.2/slc4_ia32_gcc34/bin/python'
-
-from RTTSummaryFilePathsWriter import RTTSummaryFilePathsWriter
-
-import SimpleXMLRPCServer
-
-def usage():
-    print
-    print 'WebPageCollecter -- starts server web page collector server for the RTT.'
-    print
-    print 'usage: python WebPageCollector [options]'
-    print
-    print 'Options'
-    print '   -n do not init tables from existing summary files.'
-    print '   -w write pages from existing summaries at initalisation'
-    print '   -h print this message and exit'
-    return
-
-def runServer(resBaseDir, ip='localhost', port='8001'):
-    
-    print 'RunServer.run()'
-    socket = 'http://'+ip+':'+port
-    print 'connecting to socket %s '%socket
-    
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'nwh', ['noprompt'])
-        print opts
-    except getopt.GetoptError:
-        usage()
-        sys.exit(2)
-
-    try:
-        server = xmlrpclib.ServerProxy(socket, xmlrpclib.Transport(), 'UTF-8', True)
-    except Exception, e:
-        print 'Could not get server proxy %s: '%str(e)
-    else:
-        try:
-            a =  server.areYouAlive()
-            print '%s: server is running' % time.ctime()
-            return
-        except:
-            print '%s: server not responding' % time.ctime()
-
-    print 'Preparing to instantiate server'
-                                
-    mmaxtry    = 3
-    maxtry    = mmaxtry
-    sleepTime = 20
-    running   = False
-    
-    while maxtry:
-        
-        try:
-            server = SimpleXMLRPCServer.SimpleXMLRPCServer(("", int(port)))
-            running = True
-            break
-        except:
-            print exc2string2()
-            print 'Could not instantiate server, will sleep %d secs, remaining tries: %d' % (sleepTime,
-                                                                                             maxtry)
-            time.sleep(sleepTime)
-        maxtry -= 1
-
-    if not running:
-        print 'Abandoning attempt to instantiate server after %d attempts' % mmaxtry
-        return
-
-    print 'Server is instantiated'
-    print 'Preparing to instantiate WebPageWriter'
-        
-    summaryFilePathsWriter = RTTSummaryFilePathsWriter(resBaseDir)
-
-
-    print 'RTTSummaryFilePathsWriter is instantiated'
-    print 'Preparing to register this instance with the server.'
-    try:
-        server.register_instance(summaryFilePathsWriter)
-    except:
-        print 'registration failed\n   %s' % exc2string2()
-        return
-
-    print 'RTTSummaryFilePathsWriter instance registered'
-    print 'Preparing to serve forever'
-
-    server.serve_forever()
-        
-if __name__ == '__main__':
-    ip      = 'localhost'
-    port    = '8001'
-    resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results'
-    # resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results'
-    # resBaseDir = '/afs/cern.ch/atlas/project/RTT/Results'
-    #resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/peter/Results'
-    # resBaseDir = '/afs/cern.ch/atlas/project/RTT/Results'
-    if '--noprompt' not in sys.argv:
-        s = raw_input ('resBaseDir = %s CR = No all else yes ' % resBaseDir)
-        if s == '':
-            sys.exit(0)
-            
-    runServer(resBaseDir, ip, port)
-                                                                                 
diff --git a/Tools/RunTimeTester/src/WebPageWriter.py.obsolete b/Tools/RunTimeTester/src/WebPageWriter.py.obsolete
deleted file mode 100755
index 1cfa5dc45d274a1100298500cb02a1427a16a4ad..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WebPageWriter.py.obsolete
+++ /dev/null
@@ -1,31 +0,0 @@
-from Page0XMLWriter import Page0XMLWriter
-from Page1XMLWriter import Page1XMLWriter
-from Page2XMLWriter import Page2XMLWriter
-from Page3XMLWriter import Page3XMLWriter
-from Page4XMLWriter import Page4XMLWriter
-
-class WebPageWriter:
-	def __init__(self, dbPath, key, tester, logger):
-
-		logger.debug('WebPageWriter __init__ start')
-		self.pageWriters = [
-			Page0XMLWriter(tester, logger),
-			Page1XMLWriter(dbPath, key, tester, logger),
-			Page2XMLWriter(tester, logger),
-			Page3XMLWriter(tester, logger),
-			Page4XMLWriter(tester, logger)
-			]
-		
-		# order for updating is important
-		self.pageWriters.reverse()
-		self.logger = logger
-		logger.debug('WebPageWriter __init__ finish')
-	def update(self, package, updatedMinders):
-		self.logger.info('Update method starts for WebPageWriter')
-		[p.update(package, updatedMinders) for p in self.pageWriters]
-		self.logger.info('Update method ends for WebPageWriter')
-
-	def finalise(self):
-		self.logger.info('Finalise method starts for WebPageWriter')
-		[p.finalise() for p in self.pageWriters]
-		self.logger.info('Finalise method ends for WebPageWriter')
diff --git a/Tools/RunTimeTester/src/WhoHasConfigFiles.py.obsolete b/Tools/RunTimeTester/src/WhoHasConfigFiles.py.obsolete
deleted file mode 100755
index 1975c0aac064179f162649208df3d5cea06e8091..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WhoHasConfigFiles.py.obsolete
+++ /dev/null
@@ -1,74 +0,0 @@
-#! /usr/local/bin/python
-
-
-from RTTConfigParser             import rttConfiguration
-from Paths                       import NoPackagePaths
-from RunTimer                    import RunTimer
-from formatCollection            import formatCollection
-from findRTThome                 import findRTThome
-from UserStuffRetriever          import UserStuffRetrieverBase
-from xml.parsers.xmlproc         import xmlproc, xmlval
-
-import getopt, sys, logging, os.path
-
-def handleSubstring(subString, paths, confDict):
-    usr = UserStuffRetriever(paths, subString, confDict)
-    fnames = usr.userStuff()
-    
-    logger.info('fnames')
-    logger.info(fnames)
-    
-    status = {}
-    for fname in fnames:
-        finfo = {}
-        if os.path.exists(fname):
-            finfo['exists']='Yes'
-        else:
-            finfo['exists']='No'
-            finfo['valid'] ='NA'
-            continue
-        
-        try:
-            validator = xmlval.XMLValidator()
-            validator.parse_resource(fname)
-            finfo['valid'] = 'Yes'
-        except Exception, e:
-            finfo['valid'] = 'No'
-                
-        status[fname] = finfo
-
-        logger.info('status')
-        logger.info(status)
-        print
-        print 'file name substring =',subString
-        if status:
-            print 'file','exists','valid'
-            for fn in status:
-                finfo = status[fn]
-                print finfo['exists'],finfo['valid']
-            else:
-                print 'No configuration files found from the release'
-
-if __name__ == '__main__':
-
-    from setupLoggerForDebug import setupLoggerForDebug
-    setupLoggerForDebug('INFO')
-    logger = logging.getLogger()
-    
-    (basePath, confFile, srcDir) = findRTThome()
-    print 'config file: ', confFile
-    print 'source dir:  ', srcDir
-      
-    confDict = rttConfiguration(confFile)
-    paths      = NoPackagePaths(confDict, basePath, RunTimer(60*60*20))
-    # paths.dump()
-    
-    # subString = '_RTT_things'
-
-    subStrings  = ['_TestConfiguration', '_RTT', '_ReleaseTest']
-    
-    print 'config file used    =',confFile
-    print 'release             =',paths.release
-    print 'branch              =',paths.branch
-    for string in subStrings:
-        handleSubstring(string, paths, confDict)
diff --git a/Tools/RunTimeTester/src/WorkerJobDescriptor.py b/Tools/RunTimeTester/src/WorkerJobDescriptor.py
deleted file mode 100755
index d5594e6374e30484bab326a70286831136f4aeb9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WorkerJobDescriptor.py
+++ /dev/null
@@ -1,59 +0,0 @@
-#! /usr/local/bin/python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-"""
-Class that stocks information about Worker jobs.
-Responsible for setting up the run directory for a job. Eventually becomes a data member
-of JobMinder.
-"""
-
-from Paths               import Paths
-from Descriptor          import Descriptor
-
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class WorkerJobDescriptor(Descriptor):    
-    
-    def __init__(self, paths, logger, jobSerialNumber, argDict={}):
-
-        self.package          = '' # set by Jobs
-
-        # stuff to do with data sets
-        self.datasets      = {}
-
-        # init the base class
-        Descriptor.__init__(self, paths, logger, jobSerialNumber, argDict)
-
-
-    # --------------------------------------------------------------
-
-    def hasData(self):
-        return len(self.datasets.keys()) > 0 
-
-    # --------------------------------------------------------------
-    
-    def __str__(self):
-        "Converts self to a string"
-
-        s  = Descriptor.__str__(self)
-        s += 'package:             %s\n' % str(self.package)
-        s += 'Descriptor hasData:  %s\n' % str(self.hasData())
-        for datasetType in self.datasets.keys():
-            for dataset in self.datasets[datasetType]:
-                s += 'Dataset type: ' + str(datasetType) + ', dataset physical name: ' + str(dataset.name) + '\n'
-                
-        return s
-
-    # --------------------------------------------------------------
-    
-    def dump(self):
-        """Dumps out descriptor"""
-
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info('|Worker job descriptor  dump                |')
-        self.logger.info('|-------------------------------------------|')
-        self.logger.info(self.__str__())
-        self.logger.info('')
diff --git a/Tools/RunTimeTester/src/WorkerJobMinder.py b/Tools/RunTimeTester/src/WorkerJobMinder.py
deleted file mode 100755
index 5bc04507233c2f93491545f9972d758287ce0159..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WorkerJobMinder.py
+++ /dev/null
@@ -1,152 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""
-This file contains:
-an intermediate base class         -   WorkerJobMinder
-
-- Has a the command string.
-- has a 'fire' method to submit the jobs
-- has a popen method to monitor the subshell used to run the job.
-
-The base class also specifies common tasks that must be performed across job mode
-boundaries:
-
-- database manipulations for jobs which store reference files
-- various protection manipulations to allow other mebers of the same UNIX group
-  to manipulate files on future runs.
-- copying files to be stored at the end of the job
-- actions to be taken of there is a forced shutdown
-  (triggered by receiving a signal 15)
-PS
-"""
-
-
-from ShellCommand          import ShellCommand
-from formatCollection      import formatCollection
-from Minder                import Minder
-from Defs                  import RTTdefs
-
-import os, shutil, sys, string, copy
-# ------------------------------------------------------------------------
-
-class WorkerJobMinder(Minder):    
-
-    def __init__(self, argBag, jDescriptor):
-        """
-        resDir     - results directory: recieves the logfile
-        status
-        """
-
-        # Minder __init__ needs a JobGroupDirectoryMaker to have been
-        # instantiated
-        Minder.__init__(self, argBag, jDescriptor)
-
-        self.jobWeight      = 1 # These jobs are counted in the Launcher stats.
-        self.subLog         = [] # to hold reply from the submission command
-        self.logChecker     = copy.copy(argBag.jobGroupKit.logChecker)
-        self.logResult      = 'unavailable' # should be cached in logChecker
-        
-        # extend the Minder checks to perform. Kit was obtained by base class
-        self.checks.extend(argBag.jobGroupKit.getItem('checkers'))
-        self.logChecker.logger = self.logger
-        for c in self.checks: c.logger = self.logger
-        # Add the log file to the keep files
-        # directory depends on the submit command.
-        # The name is stored in the job Descriptor at submission time        
-        destLogName = self.fullResultsFileName(self.log)
-        srcFile     = os.path.join(self.runPath,self.log)
-
-        # add the log file as a keep file
-        infoString = "Job Log file"
-        displayColor = "#cc3333"
-
-        self.keepFiles[srcFile] = {'keepFileString' : destLogName,
-                                   'infoString'     : infoString,
-                                   'displayColor'   : displayColor}
-
-       
-    # ------------------------------------------------------------------------
-
-    def collectScripts(self, scriptWriter):
-        Minder.collectScripts(self, scriptWriter)
-
-        script               = scriptWriter.makeScript()
-
-        dict = {'runScript': {'script':       script,
-                              'scriptName':  'runScript.sh',
-                              'cmd':          self.submitCommand # set in subclasses
-                              }
-                }
-        
-        self.scripts.update(dict)
-       
-    # ------------------------------------------------------------------------
-
-    def checkLog(self):
-        self.logResult        = RTTdefs.status(self.logChecker.check(self))
-        self.processingResult = self.logResult # determines is post processing should be done.
-        # log checker determines how many of the chained jobs succeeded.
-        self.chainSuccesses   = self.logChecker.nSuccesses
-        self.chainLength      = self.logChecker.chainLength
-    # ------------------------------------------------------------------------
-    
-    def fire(self, what):
-        """
-        Submits the job by spawning a subprocess.
-        The Popen objects captures std error, and std out
-        'what' is a a key in to tthe self.scripts dictionary
-        """
-
-        cmd = self.scripts[what]['cmd']
-
-        m =  'fire commandL\n: %s' % cmd
-        self.logger.debug(m)
-
-        sc = ShellCommand([cmd], self.logger)
-        self.subLog = sc.getReply()
-
-
-    # ------------------------------------------------------------------------
-
-    def dataForMonitoring(self):
-        dict = Minder.dataForMonitoring(self)
-        dict['jobSuccess'] = self.isSuccess()
-        dict['jobFailure'] = self.isFailure()
-        return dict
-
-    # ------------------------------------------------------------------------
-
-    def isSuccess(self): return self.logResult=='success'
-
-    # ------------------------------------------------------------------------
-
-    def isFailure(self): return self.logResult=='error'
-
-    # ------------------------------------------------------------------------
-
-    def status(self):
-
-
-        if self.logResult == 'unavailable': return self.stateEngine.state.state
-        return self.logResult
-
-    # ------------------------------------------------------------------------
-
-    def __str__(self):
-
-        s = Minder.__str__(self)
-        s += '----------- Worker Minder ---------------\n'
-        s += ' checks:              %s' % formatCollection(self.checks)
-        return s
-    
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|        Worker Job Minder  dump            |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-        
-# ==================================================================
diff --git a/Tools/RunTimeTester/src/WorkerJobMinder.py.obsolete b/Tools/RunTimeTester/src/WorkerJobMinder.py.obsolete
deleted file mode 100755
index dbdd311e50d66bff7f2ef7ed692771f7c247f58f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WorkerJobMinder.py.obsolete
+++ /dev/null
@@ -1,1162 +0,0 @@
-"""
-This file contains:
-an intermediate base class         -   WorkerJobMinder
-
-- Has a the command string.
-- has a 'fire' method to submit the jobs
-- has a popen method to monitor the subshell used to run the job.
-
-The base class also specifies common tasks that must be performed across job mode
-boundaries:
-
-- database manipulations for jobs which store reference files
-- various protection manipulations to allow other mebers of the same UNIX group
-  to manipulate files on future runs.
-- copying files to be stored at the end of the job
-- actions to be taken of there is a forced shutdown
-  (triggered by receiving a signal 15)
-PS
-"""
-
-import os, shutil, string,time,string,re, sys
-from popen2 import Popen4
-
-from JobGroupKits            import JobGroupKits
-from os.path                 import join
-from PathCompleter           import PathCompleter
-from Reporter                import Reporter
-from getInheritanceHeirarchy import isAthenaJob
-from RunTimer                import *
-from formatCollection        import formatCollection
-from Minder                  import Minder
-from Tools                   import isValidReleaseNumber, dirSnapShot,dirSnapShot2,dirSnapShot3,dirSnapShot4
-from Tools2                  import convertToSeconds
-from exc2string              import exc2string
-from RTTpath                 import RTTpath
-from ShellCommand            import ShellCommand
-
-from Factory_JobGroupDirectoryMaker import JobGroupDirectoryMakerFactory
-from Factory_Commander              import CommanderFactory
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-
-
-def match(patterns, lines):
-    line = None
-    for p in patterns:
-        cp = re.compile(p)
-        line = None
-
-        for l in lines:
-            if re.search(cp,l):
-                line = l
-                break
-                
-        if line: break
-
-    return line
-
-def allMatches(patterns, lines):
-    matchedLines = []
-    for p in patterns:
-        cp = re.compile(p)
-        matchedLines.extend([l for l in lines if re.search(cp,l)])
-    return matchedLines
-
-def getFromEnvLog(runPath,envVar,pathSep=':'):
-    envLogName = os.path.join(runPath, 'env.log')
-    if os.path.exists(envLogName):
-        content = open(envLogName, 'r').readlines()
-    else:
-        return []
-
-    envVar = envVar.strip() + '='
-    libstr = ''
-    for line in content:
-        if line.startswith(envVar):
-            libstr = line[len(envVar):]
-            break
-    if not libstr: return []
-        
-    return libstr.split(pathSep)
-
-class RetryTester:
-    #def __init__(self, log, elog, logger, sharePath, runPath):
-    def __init__(self, log, logger, sharePath, runPath):
-        self.logName   = log # full abs path to log file (which doesn't exist yet)
-        # self.elogName  = elog
-        self.log       = []
-        # self.elog      = []
-        self.logger    = logger
-        self.sharePath = sharePath
-        self.runPath   = runPath
-        
-    def shouldRetry(self):
-        if not self.log:            
-            if os.path.exists(self.logName):
-                self.log = open(self.logName,'r').readlines()
-
-        # if not self.elog:
-        #  if os.path.exists(self.elogName):
-        #     self.elog = open(self.elogName,'r').readlines()
-        
-        if self.commandNotFound():        return True
-        if self.missingJobOptionsExist(): return True
-        if self.missingSharedLibrary():   return True
-        
-        return False
-            
-
-    # -------------------------------------------------------------------------
-    # -------------------------------------------------------------------------
-    
-    def missingJobOptionsExist(self):
-        """
-        Looks for missing job options files in the log files. If Athena
-        complains of such a file, check to see if it really does not exist
-        """
-
-        patterns = ['^AthenaCommon.Include.IncludeErr',
-                    '^IOError: [Errno 2] No such file or directory:',
-                    '^IncludeError: include file']
-        # line = match(patterns, self.elog)
-        line = match(patterns, self.log)
-        if not line: return False
-        
-        tokens = line.split()
-        jos = match(['.py$'], tokens)
-
-        if not jos: return False
-        
-        # Note that this makes the assumption (for now, always borne out)
-        # that the 'missing' job options file is in the same project as the
-        # one trying to include it, and that these are both in the same project
-        # as the package under test.
-
-        if os.path.isabs(jos) and not os.path.exists(jos): return False
-        
-        joSearchPaths = getFromEnvLog(self.runPath,'JOBOPTSEARCHPATH',',')
-        if not joSearchPaths:
-            self.logger.warning('No search paths available in which to search for missing job options!')
-            self.logger.warning('(This is either because RTT could not find env.log file, or JOBOPTSEARCHPATH missing/empty in that file.)')
-            return False                                 
-
-        for p in joSearchPaths:
-            p = p.strip()
-            fullPath = RTTpath(p,jos)
-            if os.path.exists(str(fullPath)):
-                m  = 'Apparent connection error: Found the following job options with'
-                m += ' Athena include but which exist: %s' % str(jos)
-                self.logger.info(m)        
-                return True
-        return False
-
-
-    # -------------------------------------------------------------------------
-
-    def commandNotFound(self):
-        """
-        Checks for 'not found' commands such as athena, cmt, etc.
-        """
-
-        patterns = [
-            '^#CMT> Warning: package CMT * /afs/cern.ch/sw/contrib not found',
-            'athena.py: command not found',
-            'cmt: command not found',
-            'pool_insertFileToCatalog: command not found'
-            ]
-        
-        # lineFound = match(patterns, self.elog)
-        lineFound = match(patterns, self.log)
-        if lineFound:
-            m = 'Apparent connection error:\n%s' % lineFound
-            self.logger.info(m)
-            return True
-        return False
-
-    # -------------------------------------------------------------------------
-    
-    def missingSharedLibrary(self):
-        """
-        Looks for missing shared library files in the log files. If Athena
-        complains of such a file, check to see if it really does not exist
-        """
-        patterns = ['^System::doLoad: ERROR:']
-        # lines = allMatches(patterns, self.elog)
-        lines = allMatches(patterns, self.log)
-
-        if not lines: return False
-
-        soFiles = []
-        for l in lines:
-            #example line:
-            #System::doLoad: ERROR:libMuonAlgs.so: cannot open shared object file
-            #: No such file or directory'
-            #
-            tokens = l.split()
-            soFiles.extend([t[:-1] for t in tokens if t.endswith('.so:')])
-
-        # print soFiles
-        if not soFiles: return False
-
-        # make a list of libary paths
-        libDirs = getFromEnvLog(self.runPath,'LD_LIBRARY_PATH')
-        if not libDirs:
-            self.logger.warning('No lib paths in which to search for missing shared library!')
-            self.logger.warning('(This is either because RTT could not find env.log file, or LD_LIBRARY_PATH missing/empty in that file.)')
-            return False
-        
-        fullLibPaths = []
-        for lib in soFiles:
-            possiblePaths  = [os.path.join(libDir, lib) for libDir in libDirs]
-            foundLocations = [l for l in possiblePaths if os.path.exists(l)]
-
-            # looked everywhere on LD_LIBRARY_PAH, but did not find the
-            # library. It is genuinely not present so do not retry
-            if foundLocations: 
-
-                m  = 'Apparent afs connection error: '
-                m += 'Found the following so files reported'
-                m += ' missing but in fact present %s' % str(foundLocations)
-                self.logger.info(m)
-
-        
-                return True
-
-        return False
-
-
-
-class WorkerJobMinder(Minder):
-    
-    # ------------------------------------------------------------------------
-
-    def __init__(self, jDescriptor, logger, runTimer):
-        """
-        resDir     - results directory: recieves the logfile
-        status
-        """
-
-        # Minder __init__ needs a JobGroupDirectoryMaker to have been
-        # instantiated
-        factory = JobGroupDirectoryMakerFactory(logger)
-        self.jobGroupDirectoryMaker = factory.create(jDescriptor.paths,
-                                                     'Worker')
-
-        Minder.__init__(self, jDescriptor, logger, runTimer)
-
-        # important that the jobTimer is initialised to None for the
-        # logic of the testing of the job being in batch to work.
-        self.jobTimer       = None
-        self.jobWeight      = 1 # These jobs are counted in the Launcher stats.
-        self.popen          = None
-        self.subLog         = ""        
-        self.hasFired       = False
-        self.pid            = None
-
-        # extend the Minder checks to perform. Kit was obtained by base class
-        self.checks.append(self.jobGroupKit.logChecker)
-        self.checks.extend(self.jobGroupKit.getItem('checkers'))
-                
-        # Add the log file to the keep files
-        # directory depends on the submit command.
-        # The name is stored in the job Descriptor at submission time        
-        destLogName = self.fullResultsFileName(self.jDescriptor.log)
-        srcFile     = join(self.runPath,self.jDescriptor.log)
-
-        # add the log file as a keep file
-        infoString = "Job Log file"
-        self.keepFiles[srcFile] = {'keepFileString':destLogName,
-                                   'infoString':infoString}
-
-       
-        commander = CommanderFactory(self.logger).create( jDescriptor )
-        msg = 'Could not create a minder for descriptor %s' % (
-            jDescriptor.identifiedName)
-        if not commander: return
-        
-        self.submitCommand = commander.submitCommand(jDescriptor)
-        self.logger.debug( "submit command: "+self.submitCommand)
-
-    
-
-
-    # ------------------------------------------------------------------------
-
-    def fire(self):
-        """
-        Submits the job by spawning a subprocess.
-        The Popen objects captures std error, and std out
-        """
-        self.logger.debug('start fire()')
-        if self.hasFired:
-            msg  = 'Has already fired, being refired!??! '
-            msg += self.jDescriptor.identifiedName
-            self.logger.warning(msg)
-            return
-        os.chdir(self.jDescriptor.runPath)
-
-        self.checkQueue()
-        
-        m =  'fired %s  \n submit command %s'
-        m = m % (str(self.jDescriptor.identifiedName),
-                 str(self.submitCommand))
-        self.logger.debug(m)
-
-
-        self.popen=Popen4(self.submitCommand)
-        self.pid = self.popen.fromchild.readlines()[0].strip() # submit command prints process id to standard out
-        self.logger.debug('Minder: pid = %s' % self.pid)
-        self.hasFired=True
-        
-    #------------------------------------------------------------------------
-
-    def checkQueue(self):
-        return True
-        
-
-    #-----------------------------------------------
-
-    def setDone(self):
-        """
-        Sets the done flag. Also deletes the Popen object.
-        This is to prevent crashes due to 'too many files open'
-        os error when running many jobs.
-        """
-
-        m='Setting the done flag for job %s' % self.jDescriptor.identifiedName
-        self.logger.debug(m)
-                    
-        Minder.setDone(self)
-        del self.popen
-        self.popen = None
-
-    # ------------------------------------------------------------------------
-
-    # =======================================================
-    #
-    # Override of a state transition method
-    #
-    # =======================================================
-
-    def fromRunning(self):
-        if self.fromRunningTest():
-            self.fromRunningAction()
-
-    def fromRunningAction(self):
-        "Runs postscripts, performs checks and sets the status accordingly"
-
-        theJob = self.jDescriptor.identifiedName
-
-        self.logger.debug('Just before runPS for ' + str(theJob))
-        self.runPostScripts()
-        self.logger.debug('Just before runActions for ' + str(theJob))
-        self.runActions()
-        self.logger.debug('Just before runMoniActions for ' + str(theJob))
-        self.runMoniActions()
-        self.logger.debug('Just before runTests for ' + str(theJob))
-        self.runTests()
-        self.logger.debug('Just before performChecks for ' + str(theJob))
-        
-        # perform checks on the job (finished OK? regression?..)
-        fullStatusList = self.performChecks()
-
-        self.logger.debug('Just after performChecks for ' + str(theJob))
-        #reduce this list to a 2-member list and set the Status object
-        ssl = self.summaryCheckStatusList(fullStatusList)
-
-        self.logger.debug('ssl for job: ' + str(theJob) + ' is: ' + str(ssl))
-        self.setStatusList(ssl)
-
-        self.logger.debug('Just after setStatusList for ' + str(theJob))
-        #if not success everywhere in fullStatusList, get info for the user
-        self.handleCheckResults(ssl)
-
-        self.logger.debug('Just after handleCheckResults for ' + str(theJob))
-        
-        # copy files to be kept (log histos, ntuples...) to results dir
-        #if not self.postProcessOn:
-        self.copyKeepFiles()
-
-        self.logger.debug('Just after copyKeepFiles for ' + str(theJob))
-        
-        # make sure ALL files in run and results directory are
-        # accessible by all members of the UCL Atlas group
-        self.setProtection()
-
-        self.logger.debug('Leaving WorkerJobMinder::fromRunningAction ' + str(theJob))
-        
-#------------------------------------------------------------------------
-
-    def __str__(self):
-
-        s = Minder.__str__(self)
-        s += '----------- Worker Minder ---------------\n'
-        s += ' checks:              %s' % formatCollection(self.checks)
-        return s
-    
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|        Worker Job Minder  dump            |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-        
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-class LSFBatchJobMinder(WorkerJobMinder):
-    """
-    A JobMinder class for running on the LSF batch queues
-    """
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-        self.logger.debug( "initialising LSFBatchJobMinder")
-
-    # =======================================================
-    #
-    # Override of a state transition method
-    #
-    # =======================================================
-
-    def fromQueuedTest(self):
-        "Polls popen object to see if the submission command has terminated"
-
-        if self.popen == None:    return False
-        if self.popen.poll()==-1: return False
-
-        return True
-    
-    # -----------------------------------------------
-    
-    def fromRunningTest(self): 
-        " Running is complete if the logfile is found"
-
-        #        self.checkForTimeOut()
-        lf, found  = self.getLogFile()
-
-        return found
-
-    # -----------------------------------------------
-    
-    def fromRunningAction(self):
-        "Additional actions for this class"
-    
-        lf, found  = self.getLogFile()
-        self.logger.debug("found log file = " + lf)
-
-        dest = join(self.runPath, self.jDescriptor.log)
-        shutil.copy(lf, dest)
-        
-        WorkerJobMinder.fromRunningAction(self)
-        
-    # -----------------------------------------------
-    
-    def getLogFile(self):
-        logFile = ""
-        got_it = 0
-        fileList = os.listdir(self.runPath)
-        for object in fileList:
-            path = join(self.runPath,object)
-            if os.path.isdir(path):
-                if object.find("LSFJOB_")!=-1 :
-                    subdirFileList = os.listdir(path)
-                    for thing in subdirFileList:
-                        if thing=="STDOUT":
-                            logFile = join(self.runPath,object,thing)
-                            got_it = 1
-                            break
-
-        return (logFile,got_it)
-
-    # -----------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|     LSF Batch Job Minder  dump            |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-        
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-class LinuxInteractiveJobMinder(WorkerJobMinder):
-    
-    def __init__ (self, jDescriptor, logger,  runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-
-        self.logger.debug( "initialising LinuxInteractiveJobMinder")
-        
-
-    # =======================================================
-    #
-    # Override of a state transition method
-    #
-    # =======================================================
-
-    def fromQueuedTest(self):        
-        if self.popen == None: return False
-        return True
-
-    def fromRunningTest(self):
-        "Running is complete if the subprocess has completed"
-
-        if self.popen.poll()==-1: return False    #poll until process completes
-        
-        # Looks to see if the log has appeared
-        elif  self.jDescriptor.log  in os.listdir(self.runPath): return True
-
-        #return False
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('| Linux InteractiveJob Minder  dump         |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-            
-class PBSJobMinder(WorkerJobMinder):
-    
-    def __init__ (self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-
-        
-        self.logger.debug("initialising PBSJobMinder")    
-
-        log  = str(RTTpath(self.jDescriptor.runPath,self.jDescriptor.log))
-        # elog = str(RTTpath(self.jDescriptor.runPath,self.jDescriptor.elog))
-        sp   = self.jDescriptor.paths.shareArea
-        rp   = self.runPath
-        # self.retryTester = RetryTester(log, elog, logger, sp, rp)
-        self.retryTester = RetryTester(log, logger, sp, rp)
-        
-    # ------------------------------------------------------------------------
-
-    def stopJob(self):
-        if self.done: return
-        
-        try:
-            cmd = ['qdel ' + str(self.pid)]
-            sc       = ShellCommand(cmd, 100, None, True)
-        except Exception, e:
-            m  = 'Job: ' + str(self.jDescriptor.identifiedName)
-            m += ' threw an exception while issuing cmd: %s' % cmd[0] 
-            self.logger.warning(m)
-
-
-    # -----------------------------------------------
-    
-    def fromQueuedTest(self):
-        """Finds out from popen if queued on PBS
-        stores the submission log"""
-
-        self.logger.debug('WorkerJobMinder fromQueuedTest')
-        
-        if self.popen == None:    return False
-        if self.popen.poll()==-1: return False
-        
-        # handle the case where PBS does not allow us to submit the job
-        self.subLog = self.popen.fromchild.read()
-        self.logger.debug('Got self.subLog. It is: ' + str(self.subLog))
-
-        self.checkForPBSsubmissionError()
-        
-        return True
-
-    # -------------------------------------------------------------------------
-
-    def fromRunningTest(self):
-        "Running is complete if the log files have returned"
-        
-
-        # query if job is in batch queue
-        queryOK, inBatch = self.checkJobInBatch()
-
-        # give up if the query failed
-        if not queryOK:
-            m = 'Job %s: PBS query failed, setting state error'
-            m = m %  self.jDescriptor.identifiedName
-            self.logger.warning(m)
-
-            # on bad query return False so that RunningAction is not called
-            # and set the status to error to go to the error state functions.
-            self.setStatus('error')
-            return False 
-
-        # if the job is still running, test fails
-        if inBatch: return False
-
-        m = 'Job %s: has left batch'  %  self.jDescriptor.identifiedName
-        self.logger.debug(m)
-
-        # are the log files visible?
-        answer = self.areLogFilesInRunDir()
-        m = 'Job %s: log files found? %s'
-        m = m % (self.jDescriptor.identifiedName, str(answer))
-        self.logger.debug(m)
-
-        self.logger.debug(m)
-        dirSnapShot(self.runPath)
-        dirSnapShot2(self.runPath)
-        dirSnapShot3(self.runPath)
-        dirSnapShot4(self.runPath)
-
-        return answer
-    
-    # -------------------------------------------------------------------------
-
-    def fromRunningAction(self):
-        #add the error log to the keep file list. This exceptional
-        #handling of the error log is due to the fact that the error
-        #logs are only produced when the PBS queues are used,
-
-        # src = join(self.runPath, self.jDescriptor.elog)
-        # infoString = "Job Error log"
-        # self.keepFiles[src] = {
-        # 'keepFileString': self.fullResultsFileName(self.jDescriptor.elog),
-        # 'infoString':     infoString
-        # }
-
-        self.audit()
-        
-        WorkerJobMinder.fromRunningAction(self)
-        mustBeSet = {
-            'exitStatus': self.exitStatus,
-            'CPU time':self.cpuTime,
-            'Mem': self.mem,
-            'VMem':self.vmem,
-            'Wall time':self.wallTime
-            }
-        
-        [self.logger.info('Audit info: %s %s' % m) for m in mustBeSet.items()]
-
-    # -------------------------------------------------------------------------
-
-    def fromError(self):
-
-
-        if self.fromErrorRetryTest():
-            m= 'fromError() - retry %d' % self.errorStateCurRetry
-            self.logger.info(m)
-            self.fromErrorAction()
-        else:
-            m = 'fromError() - no retry: setting done' 
-            self.logger.info(m)
-            self.setDone()
-
-    # -------------------------------------------------------------------------
-
-    def fromErrorRetryTest(self):
-        
-        if self.retryTester.shouldRetry():
-            m ='fromErrorRetryTest() %d %d' %(self.errorStateCurRetry,
-                                              self.errorStateMaxRetry)
-            self.logger.debug(m)
-            if self.errorStateCurRetry < self.errorStateMaxRetry:
-                self.errorStateCurRetry += 1
-                return True
-
-        return False
-
-    # -------------------------------------------------------------------------
-    
-    def fromErrorAction(self):
-
-        # the presence of the log files signals the end of the running
-        # phase - remove the files,and ensure they are copied to the
-        # results web pages.
-
-        logs = [str(RTTpath(self.jDescriptor.runPath,self.jDescriptor.log))]
-        # str(RTTpath(self.jDescriptor.runPath,self.jDescriptor.elog))]
-        
-        for log in logs:
-            if os.path.exists(log):
-                dest = log+'_'+str(self.errorStateCurRetry)
-                m = 'Error retry: moving file %s to %s' % (log, dest)
-                self.logger.debug(m)
-                os.rename(log, dest)
-                self.makeKeepFileEntry(os.path.basename(dest),"Log file leading to a retry")
-                # self.keepFiles[dest] = {'keepFileString':self.fullResultsFileName(os.path.basename(dest)),'infoString':infoString}
-        
-
-
-        # re-fire the job
-        self.logger.info('fromErrorAction() - restarting the job' )
-        self.status.set('queued')
-        self.hasFired = False
-        self.jobTimer = None
-        self.fire()
-        
-    # -------------------------------------------------------------------------
-
-    def checkQueue(self):
-        #check if queue exists        
-
-        if self.jDescriptor.paths.mode == 'PBS':
-            testCmd = 'qstat -q ' + self.jDescriptor.batchQueue()
-            self.popen =Popen4(testCmd)
-            result  = self.popen.fromchild.read()
-
-            badQueue = result.find('Unknown queue')!=-1
-            if badQueue:
-                msg  = 'Exiting job because queue, '
-                msg += self.jDescriptor.batchQueue()
-                msg += ', does not exist for mode '
-                msg += self.jDescriptor.paths.mode
-                self.logger.error(msg)
-
-                self.jDescriptor.reporter.badQueueError(self.jDescriptor.batchQueue(),self.jDescriptor.mode)
-                self.jDescriptor.reporter.writeReport()
-                self.forceCloseDown()
-
-    # -------------------------------------------------------------------------
-    
-    def checkForPBSsubmissionError(self):        
-        if self.subLog.find('pbs_iff: error returned') != -1:
-            self.jDescriptor.reporter.batchMachineRefusesSubmissionError()
-            self.jDescriptor.reporter.writeReport()
-            self.forceCloseDown()
-        
-    # -------------------------------------------------------------------------
-
-    def checkJobInBatch(self):
-        'Create the RunTimer if job is no longer in batch queues.'
-
-        queryOK = True
-        inBatch = self.jobTimer == None
-
-        # if job timer has been set previously, the job has left batch
-        # the initial condition is that the job is in batch.
-        if not inBatch: return (queryOK, inBatch)
-
-        # if interogation of the batch system failed, flag this.
-        # note that if the query fails, the timer will not be set
-        queryOK, inBatch = self.isJobStillInBatch()
-        if not queryOK: return (queryOK, inBatch)
-
-        # the batch system was successfully queried, and the job was not there
-        # and this is the first time that this has happened: start the timer
-        # which is also the flag that the job is no longer in the batch system.
-        if not inBatch: self.setJobTimer()
-        return (queryOK, inBatch)
-
-    # -------------------------------------------------------------------------
-
-    def setJobTimer(self):
-        timerLength = 60*60 # seconds
-        msg  = 'PBS job ID: ' + str(self.pid)
-        msg += ' belonging to job ' + str(self.jDescriptor.identifiedName)
-        msg += ' is no longer in the PBS queues. Creating a job timer initialised with ' + str(timerLength) + ' seconds...'
-        self.logger.info(msg)
-        
-        # create the timer
-        self.jobTimer = RunTimer(timerLength, self.logger)
-
-    # -------------------------------------------------------------------------
-    
-    def pbsReplyProvesJobIsInQueue(self,reply):
-        # A positive reply is a 3 liner with the 3rd line starting with the job id
-        if len(reply)==3:
-            if reply[2].strip().startswith(self.pid.split('.')[0]):
-                return True
-            
-        return False
-
-    # -------------------------------------------------------------------------
-    
-    def pbsReplyProvesJobIsAbsentFromQueue(self,reply):
-        # A negative reply is a 1 liner 
-        if len(reply)==1:
-            if reply[0].strip().find('qstat: Unknown Job Id')!=-1:
-                return True
-            
-        return False
-
-    # -------------------------------------------------------------------------
-
-    def sleepTime(self,attempt):
-        if attempt <= 3:
-            return 300
-        else:
-            return 300 * (2**(attempt-3))
-
-    # -------------------------------------------------------------------------
-    
-    def isJobStillInBatch(self):
-        'Returns True is job still in batch, False otherwise.'
-        isFound = True
-        maxTry  = 10
-        attempt = 0
-        queryStat  = False
-        
-        while True:
-            attempt += 1
-            try:
-                pbsQuery = ['qstat ' + str(self.pid)]
-                sc       = ShellCommand(pbsQuery, 100, None, True)
-            except Exception, e:
-                m  = 'Job: ' + str(self.jDescriptor.identifiedName)
-                m += ' threw an exception while issuing qstat'
-                m += ' Attempt: %d of %d\n' % (attempt, maxTry)
-                m += ' Traceback:\n%s%s\n'% (exc2string(sys.exc_info()),str(e))
-                self.logger.warning(m)
-
-                try:
-                    self.printStatusCodesToLogger()
-                except Exception, e:
-                    m  = 'Job: %s threw an exception trying to print '
-                    m += ' qstat status codes! Attempt: %d'
-                    m = m % (str(self.jDescriptor.identifiedName), attempt)
-                    self.error(m)
-                    self.logger.error(exc2string(sys.exc_info()))
-                    self.logger.error(str(e))
-                                        
-                    self.sleepTime(attempt) # maybe things be all better in a bit
-            else:
-                pbsReply = sc.getReply()
-                
-                if self.pbsReplyProvesJobIsInQueue(pbsReply):
-                    queryStat = True
-                    isFound   = True
-                    return (queryStat,isFound)
-                elif self.pbsReplyProvesJobIsAbsentFromQueue(pbsReply):
-                    queryStat = True
-                    isFound   = False
-                    return (queryStat,isFound)
-                else: # reply is corrupt
-                    msg  = 'Attempt number: ' + str(attempt) +'\n'
-                    msg += 'PBS reply to command: ' + str(pbsQuery) + ' was corrupt! It was:\n '
-                    msg += str(pbsReply) + '\n'
-                    msg += self.getExtraInfoOnQstat() + '\n'
-                    self.logger.warning(msg)                    
-                    queryStat = False
-                    self.sleepTime(attempt) # maybe things be all better in a bit
-                    
-            if attempt==maxTry:
-                # give up and kill this job
-                msg  = 'Job: ' + str(self.jDescriptor.identifiedName)
-                msg += ' is being killed after %d' % maxTry
-                msg += ' failed attempts to interrogate qstat about it.'
-                self.logger.error(msg)
-                queryStat = False
-                isFound   = False
-                return (queryStat, isFound)
-    
-    # -------------------------------------------------------------------------
-
-    def getExtraInfoOnQstat(self):
-        extraInfoCmd = ['echo "ls -l /usr/bin/..."',
-                        'ls -l /usr/bin/',
-                        'echo ""',
-                        'echo "Echoing PATH variable"',
-                        'echo $PATH',
-                        'echo ""',
-                        'echo "which qstat..."',
-                        'which qstat']
- 
-        try:
-            sc = ShellCommand(extraInfoCmd)
-        except Exception, e:
-            reply = ['Exception was thrown trying to use ShellCommand to gather extra information on qstat.']
-            reply.append(exc2string(sys.exc_info()))
-            reply.append(str(e))
-        else:
-            reply = sc.getReply()
-
-        reply.append('Will sleep for 30 seconds now before next attempt to interrogate qstat.')
-
-        replyString = ''
-        for line in reply:
-            replyString += line + '\n'
-        return replyString
-
-    # -------------------------------------------------------------------------
-    
-    def printStatusCodesToLogger(self):
-        if self.statusInt:
-            wcoreStat   = os.WCOREDUMP(self.statusInt)
-            wifSigStat  = os.WIFSIGNALED(self.statusInt)
-            wifExitStat = os.WIFEXITED(self.statusInt)
-            wexitStat   = os.WEXITSTATUS(self.statusInt)
-            wstopStat   = os.WSTOPSIG(self.statusInt)
-            wtermStat   = os.WTERMSIG(self.statusInt)
-            del(self.statusInt)
-
-            msg  = '==============================================\n'
-            msg += 'EXCEPTION WAS THROWN TRYING TO QSTAT A JOB ID!\n'
-            msg += 'Job ID  : ' + str(self.pid) + '\n'
-            msg += 'Job name: ' + str(self.jDescriptor.identifiedName) + '\n'
-            msg += '----------------------------------------------\n'
-            msg += 'wcoreStat   : ' + str(wcoreStat)   + '\n'
-            msg += 'wifSigStat  : ' + str(wifSigStat)  + '\n'
-            msg += 'wifExitStat : ' + str(wifExitStat) + '\n'
-            msg += 'wexitStat   : ' + str(wexitStat)   + '\n'
-            msg += 'wstopStat   : ' + str(wstopStat)   + '\n'
-            msg += 'wtermStat   : ' + str(wtermStat)   + '\n'
-            msg += '==============================================\n'
-            self.logger.error(msg)
-        else:
-            msg  = '==============================================\n'
-            msg += 'EXCEPTION WAS THROWN TRYING TO QSTAT A JOB ID!\n'
-            msg += 'Job ID  : ' + str(self.pid) + '\n'
-            msg += 'Job name: ' + str(self.jDescriptor.identifiedName) + '\n'
-            msg += '----------------------------------------------\n'
-            msg += 'Unable to ask os functions as self.statusInt was not set!\n'
-            msg += '==============================================\n'
-            self.logger.error(msg)
-
-    # -------------------------------------------------------------------------
-    
-    def areLogFilesInRunDir(self):        
-        files = os.listdir(self.runPath)
-
-        #if self.jDescriptor.log in files and self.jDescriptor.elog in files:
-
-        # NOTE: we used to have the above when  Commander qsub command split the
-        # std out and err into 2 different files. Now we have joined them, so only
-        # need look for log file.
-        if self.jDescriptor.log in files:
-            m = "Found the log file %s in rundir %s"
-            m = m % (self.jDescriptor.log, self.runPath)            
-            self.logger.debug(m)
-
-            return True
-
-        return False
-
-    # -------------------------------------------------------------------------
-    def audit(self):
-
-        if self.jDescriptor.paths.site != 'UCLslc3': return
-
-        user = os.environ['USER']
-        fn = time.strftime("%Y%m%d", time.gmtime())
-        src = user+'@pc72.hep.ucl.ac.uk:/var/spool/pbs/server_logs/'+fn        
-        dest = os.path.join(str(self.runPath),fn)
-
-        sc = ShellCommand(['scp -1 '+src+' '+ dest],timeout=60)
-        self.logger.info(sc.getReply())
-
-        if not os.path.exists(dest):
-            m  = 'could not retrieve file %s: - ' % src
-            m += 'needs ssh keys to be set up for user %s' % user
-            self.logger.error(m)
-        else:
-            self.logger.info('Managed to retrieve from pc72 PBS log file %s ' % dest)
-
-            
-        pbslog = open(dest, 'r')
-        lines = pbslog.readlines()
-        pbslog.close()
-        self.logger.info('Found %d lines in the PBS log file. Will now delete it to save space.' % len(lines))
-        os.remove(dest)
-        
-        # sample line:
-        # 08/17/2006 15:57:05;0010;PBS_Server;Job;347428.pc72.hep.ucl.ac.uk;Exit_status=0 resources_used.cput=03:48:00 resources_used.mem=893492kb resources_used.vmem=969136kb resources_used.walltime=23:39:42
-
-        lines = [l for l in lines if l.find(self.pid)>-1]
-        lines = [l for l in lines if l.find('resources_used')>-1]
-        self.logger.info('Grepping of PBS log file found %d matching lines.' % len(lines))
-        if not lines:
-            self.logger.info('Found 0 matching lines in PBS log file. Returning.')
-            return
-        
-        line = lines[0]
-        tokens = line.split()
-
-        ok = True
-        for t in tokens:
-            if t.find('Exit_status=')>-1:
-                ttokens = t.split('=')
-                if len(ttokens) == 2: self.exitStatus = ttokens[1]
-                continue
-            if t.startswith('resources_used.'):
-                ttokens = t.split('.')
-                if len(ttokens) == 2:
-                    ttokens = ttokens[1].split('=')
-                    if ttokens[0] == 'cput':     self.cpuTime = convertToSeconds(ttokens[1])
-                    if ttokens[0] == 'mem':      self.mem = ttokens[1]
-                    if ttokens[0] == 'vmem':     self.vmem = ttokens[1]
-                    if ttokens[0] == 'walltime': self.wallTime = convertToSeconds(ttokens[1])
-                    
-
-        mustBeSet = [
-            self.exitStatus,
-            self.cpuTime,
-            self.mem,
-            self.vmem,
-            self.wallTime
-            ]
-
-        [self.logger.warning('Error finding audit information')
-         for m in mustBeSet if m=='Unknown']
-            
-
-    # -------------------------------------------------------------------------
-
-    def __str__(self):
-        
-        s = WorkerJobMinder.__str__(self)+'\n'
-        return s
-
-    # -------------------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|           PBS  Minder  dump               |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-    
-class DummyJobMinder(WorkerJobMinder):
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-
-    # ------------------------------------------------------------
-    
-    def fire(self):
-        self.logger.debug('end fire()')
-        self.setStatus('success')
-        self.hasFired = True
-
-    # ------------------------------------------------------------
-    
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|           Dummy  Minder  dump             |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-    
-class RunningDummyJobMinder(WorkerJobMinder):
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-        
-    def fire(self):
-        self.logger.debug('start fire()')
-        self.setStatus('running')
-        self.hasFired = True
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|           Running Dummy  Minder  dump     |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-    
-class QueuedDummyJobMinder(WorkerJobMinder):
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-        
-    def fire(self):
-        self.logger.debug('start fire()')
-        self.setStatus('queued')
-        self.hasFired = True
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|      Queued Dummy  Minder  dump           |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-
-# ==================================================================
-# ==================================================================
-# ==================================================================
-
-class TransparentDummyJobMinder(WorkerJobMinder):
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self,
-                                 jDescriptor,
-                                 logger,
-                                 runTimer)
-
-    def fire(self):
-        """Sets the sattus to the final status of the last job to use
-        the run directory"""
-        
-        self.logger.debug('start fire()')
-        statFile = open( join(self.jDescriptor.runPath, 'status.txt'), 'r')
-        status = statFile.read()
-        self.setStatus(status)
-
-        self.hasFired = True
-                         
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('| Transparant Dummy Job Minder  dump        |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-
-#--------------------------------------------------------------------------
-        
-
-# =============================================================================
-class ErrorMinder(WorkerJobMinder):
-    def __init__(self, jDescriptor, logger, runTimer):
-        WorkerJobMinder.__init__(self, jDescriptor, logger, runTimer)
-    def fromQueuedAction(self):
-        self.setStatus("error")
-    def fromError(self): self.setDone()
diff --git a/Tools/RunTimeTester/src/WorkerMinder.py.obsolete b/Tools/RunTimeTester/src/WorkerMinder.py.obsolete
deleted file mode 100755
index 543aa49cd8125e0c10b497abcdac013de57cba92..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WorkerMinder.py.obsolete
+++ /dev/null
@@ -1,163 +0,0 @@
-"""
-This file contains:
-an intermediate base class         -   WorkerJobMinder
-
-- Has a the command string.
-- has a 'fire' method to submit the jobs
-- has a popen method to monitor the subshell used to run the job.
-
-The base class also specifies common tasks that must be performed across job mode
-boundaries:
-
-- database manipulations for jobs which store reference files
-- various protection manipulations to allow other mebers of the same UNIX group
-  to manipulate files on future runs.
-- copying files to be stored at the end of the job
-- actions to be taken of there is a forced shutdown
-  (triggered by receiving a signal 15)
-PS
-"""
-
-import os, shutil, sys, string, copy
-# from popen2 import Popen4
-from ShellCommand import ShellCommand
-
-from os.path                   import join
-
-from formatCollection          import formatCollection
-from Minder2                   import Minder
-from Defs                      import RTTdefs
-
-# ------------------------------------------------------------------------
-
-class WorkerJobMinder(Minder):    
-
-    def __init__(self, argBag, jDescriptor):
-        """
-        resDir     - results directory: recieves the logfile
-        status
-        """
-
-        # Minder __init__ needs a JobGroupDirectoryMaker to have been
-        # instantiated
-        Minder.__init__(self, argBag, jDescriptor)
-
-        self.jobWeight      = 1 # These jobs are counted in the Launcher stats.
-        #self.popen          = None
-        self.subLog         = [] # to hold reply from the submission command
-        self.logChecker     = copy.copy(argBag.jobGroupKit.logChecker)
-        self.logResult      = 'unavailable' # should be cached in logChecker
-        
-        # extend the Minder checks to perform. Kit was obtained by base class
-        self.checks.extend(argBag.jobGroupKit.getItem('checkers'))
-        self.logChecker.logger = self.logger
-        for c in self.checks: c.logger = self.logger
-        # Add the log file to the keep files
-        # directory depends on the submit command.
-        # The name is stored in the job Descriptor at submission time        
-        destLogName = self.fullResultsFileName(self.log)
-        srcFile     = join(self.runPath,self.log)
-
-        # add the log file as a keep file
-        infoString = "Job Log file"
-        displayColor = "#cc3333"
-
-        self.keepFiles[srcFile] = {'keepFileString' : destLogName,
-                                   'infoString'     : infoString,
-                                   'displayColor'   : displayColor}
-
-       
-    # ------------------------------------------------------------------------
-
-    def collectScripts(self, scriptWriter):
-        Minder.collectScripts(self, scriptWriter)
-
-        script               = scriptWriter.makeScript()
-
-        dict = {'runScript': {'script':       script,
-                              'scriptName':  'runScript.sh',
-                              'cmd':          self.submitCommand # set in subclasses
-                              }
-                }
-        
-        self.scripts.update(dict)
-       
-    # ------------------------------------------------------------------------
-
-    def checkLog(self):
-        self.logResult        = RTTdefs.status(self.logChecker.check(self))
-        self.processingResult = self.logResult # determines is post processing should be done.
-        # log checker determines how many of the chained jobs succeeded.
-        self.chainSuccesses   = self.logChecker.nSuccesses
-        self.chainLength      = self.logChecker.chainLength
-    # ------------------------------------------------------------------------
-    
-    def fire(self, what):
-        """
-        Submits the job by spawning a subprocess.
-        The Popen objects captures std error, and std out
-        'what' is a a key in to tthe self.scripts dictionary
-        """
-
-        cmd = self.scripts[what]['cmd']
-
-        m =  'fire commandL\n: %s' % cmd
-        self.logger.debug(m)
-
-        sc = ShellCommand([cmd], self.logger)
-        self.subLog = sc.getReply()
-        # popen  = Popen4(cmd)
-        # self.subLog = popen.fromchild.readlines()
-
-
-    # ------------------------------------------------------------------------
-
-    def dataForMonitoring(self):
-        dict = Minder.dataForMonitoring(self)
-        dict['jobSuccess'] = self.isSuccess()
-        dict['jobFailure'] = self.isFailure()
-        return dict
-
-    # ------------------------------------------------------------------------
-
-    def isSuccess(self): return self.logResult=='success'
-
-    # ------------------------------------------------------------------------
-
-    def isFailure(self): return self.logResult=='error'
-
-    # ------------------------------------------------------------------------
-
-    def status(self):
-
-        if self.operatorCloseDown:return 'operatorCloseDown' 
-
-        if self.logResult == 'unavailable': return self.stateEngine.state.state
-
-        return self.logResult
-
-    # ------------------------------------------------------------------------
-
-    def doPostProcessing(self):
-        if self.processingResult == 'error': return False
-        return Minder.doPostProcessing(self)
-    # ------------------------------------------------------------------------
-
-    def __str__(self):
-
-        s = Minder.__str__(self)
-        s += '----------- Worker Minder ---------------\n'
-        s += ' checks:              %s' % formatCollection(self.checks)
-        return s
-    
-
-    def dump(self):
-        self.logger.debug('|-------------------------------------------|')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|        Worker Job Minder  dump            |')
-        self.logger.debug('|                                           |')
-        self.logger.debug('|-------------------------------------------|')
-
-        self.logger.debug(self.__str__())
-        
-# ==================================================================
diff --git a/Tools/RunTimeTester/src/WorkerMinderStateEngine.py.obsolete b/Tools/RunTimeTester/src/WorkerMinderStateEngine.py.obsolete
deleted file mode 100755
index 8b2491de141c3bb213fef9e2d11aaa85ab6496be..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/WorkerMinderStateEngine.py.obsolete
+++ /dev/null
@@ -1,30 +0,0 @@
-"""State Engine for WorkerJobMinders.
-
-overrides of the method fromRunning to include log checking.
-
-"""
-
-from MinderStateEngine  import MinderStateEngine
-
-class WorkerMinderStateEngine(MinderStateEngine):
-    
-    def __init__(self, workerminder, state='queued'):
-        MinderStateEngine.__init__(self, workerminder, state)
-        
-    # -----------------------------------------------
-
-    def fucntionsMinderMustHave(self):
-        return  ['fire', 'checkLog']
-
-    # -----------------------------------------------
-    
-    def fromQueued(self):
-        self.minder.fire('runScript')
-        self.state.set('running')
-
-    # -----------------------------------------------
-    
-    def fromRunningActions(self):
-        self.minder.checkLog()
-        MinderStateEngine.fromRunningActions(self)
-
diff --git a/Tools/RunTimeTester/src/XMLCollector.py.obslete b/Tools/RunTimeTester/src/XMLCollector.py.obslete
deleted file mode 100755
index 2b242cbcfc9e1023999039789e9bf8abdec0f500..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/XMLCollector.py.obslete
+++ /dev/null
@@ -1,314 +0,0 @@
-#####################################
-#
-# Store a collection of aml files in an XML file
-# This module acts as a mapping container for the
-# file (ie supports the necessary methods for a map
-#
-# UnTested.
-#
-###############################
-
-
-
-import sys, os, glob, string, shutil, StringIO
-import exceptions, types    
-
-from Tools2          import isType 
-from xml.dom.minidom import Node
-import xml.dom.ext
-
-
-from makeRTTDOM  import makeRTTDOM
-from Evaluate    import Evaluate, getText
-
-
-def keyToDict(node):
-    
-    dict = {}
-    keyNode = Evaluate('keys', node)[0]
-    keyElements = [c for c in keyNode.childNodes
-                   if c.nodeType ==  Node.ELEMENT_NODE]
-    
-    
-    for k in keyElements: dict[k.nodeName] = getText(k)
-    
-    
-    return dict
-
-
-def dictToKey(dict, elementCreator, textNodeCreator):
-    
-    keys = elementCreator('keys')
-
-    for k in dict.keys():
-        t = textNodeCreator(dict[k])
-        e = elementCreator(k)
-        e.appendChild(t)
-        keys.appendChild(e)
-
-    return keys
-
-class XMLCollectorIter:
-    def __init__(self, documentElement):
-        self.reset(documentElement)
-
-    def reset(self, documentElement):
-        self.doms = Evaluate('keyedSummary/RTTSummary', documentElement)
-        currentDomIndex = 0
-        maxDomIndex     = len(self.doms)
-        
-
-    def next(self):
-
-        if currentDomIndex == maxDomIndex: raise StopIteration
-
-        dom = self.doms[currentDomIndex]
-        currentDomIndex += 1
-
-        reurn dom
-
-
-    def __iter__(self): return self
-    
-        
-
-class XMLCollector:
-    def __init__(self, dbDir, fileName, logger):
-
-        self.dbDir      = dbDir
-        self.fileName   = os.path.join(self.dbDir, fileName)
-        self.logger     = logger
-
-        self.document = None
-        self.readIn()
-
-        if self.document is None:
-            self.logger.info('creating new ETTSummaries document')
-            self.document = makeRTTDOM('RTTSummaries')
-
-        self.iter            = XMLCollectorIter(self)
-        
-    def readIn(self):
-
-        if not os.path.exists(self.fileName): return
-
-        try:
-            dom = xml.dom.minidom.parse(self.fileName)
-        except:
-            self.logger.error('parse of %s failed' % self.fileName)
-        else:
-            self.document = dom
-
-
-    def matchDict(self, newDict):
-
-        oldSummaries = Evaluate('keyedSummary', self.document.documentElement)
-
-        oldSummaries = [(keyToDict(n), n) for n in oldSummaries]
-
-        def dictCmp(d1, d2):
-
-            if len(d1.keys()) != len(d2.keys()): return False
-
-            for k in d1.keys():
-                try:
-                    if d1[k] != d2[k]: return
-                except KeyError:
-                    return False
-
-            return True
-
-        matched = [n[1] for n in oldSummaries if dictCmp(n[0], newDict)]
-
-        return matched
-
-
-    def __len__(self):
-        return len(Evaluate('keyedSummary', self.document.documentElement))
-
-    def __getitem__(self, key):
-        if not (type.type(key) is types.DictType): raise Exception.TypeError
-        # remove any nodes with same key
-        self.matchDict(key)
-
-        if not matches: raise Exception.KeyError
-        
-        RTTSummaryNodes = Evaluate('RTTSummary', matches[0])
-
-        if not RTTSummaryNodes: raise Exception.KeyError
-        
-        buffer = StringIO.StringIO()
-        summary = xml.dom.ext.PrettyPrint(RTTSummaryNode, buffer)
-        
-        return buffer.getvalue()
-
-       
-
-    def __setitem__(self, key):
-        " key is a dict of name-value strings"
-        
-        if not (type.type(key) is types.DictType): raise Exception.TypeError
-
-        # make the node to insert
-        keyElement = dictToKey(key,
-                               self.document.elementCreator,
-                               self.document.textNodeCreator
-                               )
-
-        ksElement   = self.document.createElement('keyedSummary')
-        ksElement.appendChild(keyElement)
-        ksElement.appendChild(self.document.importNode(dom.documentElement,
-                                                       deep=1))
-
-        # remove any nodes with same key
-        matches = self.matchDict(key)
-
-        [self.document.documentElement.removeChild(m) for m in matches]
-
-        #insert new node
-        self.document.documentElement.appendChild(ksElement)
-
-        # this invalidates the iterator, as it is reset.
-        self.iter.reset(self.document.documentElement)
-
-    def __delitem__(self, key):
-        if not (type.type(key) is types.DictType): raise Exception.TypeError
-        matches =  self.matchDict(key)
-        [self.document.documentElement.removeChild(m) for m in matches]
-
-        
-    def __iter__(self): return self.iter
-
-    def __contains__(self, item):
-        if not (type.type(key) is types.DictType): raise Exception.TypeError
-
-        matches =  self.matchDict(key)
-        if matches: return True
-        return False
-
-
-class RTTXMLCollector(XMLCollector):
-    def __init__(self, dbDir, fileName, logger, maxVersion=20, frequency=1):
-        XMLCollector.__init__(self, dbDir, fileName, logger)
-
-        self.maxVersion = maxVersion
-        self.updateFrequency = freq
-        self.updateCount     = 0
-
-
-
-    def update(self, dom):
-        el = dom.documentElement
-
-        paths = [
-            'overview/build',
-            'overview/originalBranch',
-            'overview/site',
-            'overview/platform',
-            'overview/release',
-            'overview/releaseType',
-            'overview/runType',
-            'overview/nicosDate2'
-            ]
-
-
-
-        keys = []
-
-        for p in paths:
-            keys.append(Evaluate(p, el)[0])
-
-        XMLCollector.update(self, dom, keys)
-
-
-    def update(self, dom, keyNodes, force = False):
-
-        if not force:
-            self.updateCount += 1
-            if self.updateCount < self.updateFrequency: return
-            
-            self.updateCount = 0
-
-        self.__setitem__(keyNodes)
-
-
-    def version(self):
-
-        versionedFiles = glob.glob(self.fileName+';[0-9][0-9][0-9]')
-
-        versionNumbers = [int(f[-3:]) for f in versionedFiles]
-        
-        if len(versionNumbers) == 0:
-            ver = 0
-        elif len(versionNumbers)<self.maxVersion:
-            versionNumbers.sort()
-            ver = versionNumbers[-1]
-            ver += 1
-        else:
-            times = [(os.stat(f).st_ctime, f) for f in versionedFiles]
-            times.sort()
-            latest = times[-1][1]
-            ver = int(latest[-3:])+1
-            if ver>=self.maxVersion: ver = 0
-
-        return string.zfill(ver, 3 )
-
-
-        
-    def write(self):
-        ver = self.version()
-        vFileName = self.fileName+';'+ver
-        
-        if os.path.exists(self.fileName):shutil.copy(self.fileName, vFileName)
-
-        ofile = open(self.fileName,'w')
-        xml.dom.ext.PrettyPrint(self.document, ofile)
-        ofile.close()
-
-
-
-if __name__ == '__main__':
-
-
-    class Logger:
-        def debug(self, string): print string
-        def info(self, string): print string
-        def warning(self, string): print string
-        def error(self, string): print string
-
-    logger = Logger()
-    dbDir = '/local_disk/sherwood/RTT01'
-    fn = 'collectorTest.xml'
-
-
-    collector = XMLCollector(dbDir, fn, logger)
-
-    dom = xml.dom.minidom.parse('/local_disk/sherwood/RTT01/Results/opt/N.0.X/release/rel_5/RTTSummary.xml')
-
-    
-    collector.update(dom)
-    collector.write()
-
-
-
-    dict = {
-        'build': 'opt',
-        'originalBranch': 'N.0.X',
-        'site': 'CERN',
-        'platform': 'slc3',
-        'release': 'rel_5',
-        'runType': 'release'}
-    
- 
-    summary = collector.retrieve(dict)
-    assert summary is None
-
-
-    dict['releaseType'] =  'project'
-    dict['nicosDate2']  =  '1164350933'
-    
-    summary = collector.retrieve(dict)
-    # assert xml is not None
-
-    print 'result:'
-    print summary
diff --git a/Tools/RunTimeTester/src/XMLFileSplicer.py.obsolete b/Tools/RunTimeTester/src/XMLFileSplicer.py.obsolete
deleted file mode 100755
index cf9aa41e754bfc6a983ab6bcd033345474d247ff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/XMLFileSplicer.py.obsolete
+++ /dev/null
@@ -1,127 +0,0 @@
-"""
-class provding functions to add xml fragments to an existing XML string
-"""
-
-
-from xml.dom.minidom        import parseString
-
-def russianDolls(nodes, dom):
-    "nest a list of DOM nodes - innermost tag first"
-    
-    assert(nodes) # no empty lists
-
-    if len(nodes) == 1:
-        print 'returning single node'
-        return nodes[0]
-
-    
-
-    def dolls(node1, node2):
-        print 'dolls, node1, node2', node1, node2
-        node2.appendChild(node1)
-        return node2
-    
-    innerNode = nodes[0]
-    for node in nodes[1:]:
-        print node
-        innerNode = dolls(innerNode, node) 
-
-    return innerNode
-
-
-def makeElementWithText(text, tag, dom):
-    "creates a DOM node with tagName tag and a text child node"
-    
-    textNode = dom.createTextNode(str(text))
-    node     = dom.createElement(tag)
-    node.appendChild(textNode)
-    return node
-
-def dataToNodes(data, dom):
-    """
-    converts a list of pairs to a list of DOM nodes. The second element,
-    if not None, is treated as a text node.
-    """
-    nodes = []
-    for datum in data:
-        if datum[1]:
-            nodes.append(makeElementWithText(datum[1], datum[0], dom))
-        else:
-            nodes.append(dom.createElement(datum[0]))
-
-    return nodes
-
-
-
-class XMLFileSplicer:
-    def __init__(self, initialXML=None):
-        
-        if not initialXML:
-            initialXML  = '<?xml version="1.0"?>\n'
-            initialXML += '<!DOCTYPE runTimeTesterSummary SYSTEM'
-            initialXML += ' "http://www.hep.ucl.ac.uk/'
-            initialXML += 'atlas/AtlasTesting/DTD/RTTsummary.dtd">\n'
-            initialXML += '\n'
-            initialXML += '<runTimeTesterSummary>\n'
-            initialXML += '</runTimeTesterSummary>\n'
-
-        self.dom = parseString(initialXML)
-        
-    def spliceAtEnd(self, nestedData, innerData):
-        """
-        copies the data passed to the end of a DOM.
-        nestedData will form a series of nested nodes.
-        innerData be unnested children of the innner most node
-        produced from nestedData.
-
-        Quasi-exceptions:
-
-        innerNodes empty list: the nested node is added to document end
-        with no inner nodes
-        
-        nestedData empty list: all innerNodes are added to document end
-        
-        """
-
-        innerNodes = []
-
-        if innerData: innerNodes = dataToNodes(innerData, self.dom)
-
-        if nestedData:
-            nodes      = dataToNodes(nestedData, self.dom)
-            nodes.reverse()
-            nodes[0].childNodes.extend(innerNodes)
-            node = russianDolls(nodes, self.dom)
-            self.dom.documentElement.childNodes.append(node)
-        else:
-            self.dom.documentElement.childNodes.extend(innerNodes)
-            
-    def asString(self): return self.dom.toprettyxml('   ')
-
-if __name__ == "__main__":
-
-    good = "<a><b></b></a>"
-
-    try:
-        xfs = XMLFileSplicer(good)
-        print 'read good file OK'
-    except Exception, e:
-        print e
-        print 'error reading good file'
-
-
-    
-    nestedData = [
-        ('monitor',None),
-        ('innermonitor','aaa'),
-        ]
-    for i in range(3):
-        data = [
-            ('key','kkk'),
-            ('val', str(i))
-            ]
-
-        xfs.spliceAtEnd(nestedData, data)
-    
-    # print xfs.dom.toprettyxml('   ')
-    print xfs.asString()
diff --git a/Tools/RunTimeTester/src/XMLTools.py b/Tools/RunTimeTester/src/XMLTools.py
deleted file mode 100755
index 4a63deb7513b178af36dd34c434dc96815eb9ddb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/XMLTools.py
+++ /dev/null
@@ -1,39 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-
-
-def nodeToString(node):
-  """
-  return text content of this possed noed and
-  all its children
-  """
-  
-  text = getText([node])
-  for n in node.childNodes:
-    text += nodeToString(n)
-
-  return text
-    
-def getText(nodeList):
-  text = ''
-  for node in nodeList:
-    #    if node.nodeType == node.TEXT_NODE:
-    if isTextNode(node):      
-      text = text + node.data.strip()
-      text = text.strip()
-      
-  return text
-               
-
-def isTextNode(node):
-  return node.nodeType == node.TEXT_NODE
-
-
-def isElementNode(node):
-  return node.nodeType == node.ELEMENT_NODE
-
-def getElementNodes(nodeList):
-    return [n for n in nodeList if isElementNode(n)]
-
-
diff --git a/Tools/RunTimeTester/src/adjustPathEnvVar.py b/Tools/RunTimeTester/src/adjustPathEnvVar.py
deleted file mode 100644
index 1544695d77fdb645f6fd3e010e4131e895898cea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/adjustPathEnvVar.py
+++ /dev/null
@@ -1,21 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, string
-# -------------------------------------------------------------------------
-
-def adjustPathEnvVar(siteSpecificsObj):        
-    envPath   = os.environ['PATH']
-    
-    sitePaths = siteSpecificsObj.addToPath()
-    toRemove  = []
-    tokens    = []
-    if sitePaths:
-        tokens = sitePaths.split(':')
-        [toRemove.append(t) for t in tokens if envPath.find(t)>=0]
-        
-        [tokens.remove(t) for t in toRemove]
-        
-        if tokens:
-            tokens.append(envPath)
-            os.environ['PATH'] = string.join(tokens, ':')
-            
diff --git a/Tools/RunTimeTester/src/cfg2xml.py.obsolete b/Tools/RunTimeTester/src/cfg2xml.py.obsolete
deleted file mode 100755
index 6549507e4f2c3a13f3a3ffa39fe49a4a5d7feea3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/cfg2xml.py.obsolete
+++ /dev/null
@@ -1,38 +0,0 @@
-import sys, os
-from os.path import exists
-
-def cfg2xml(cfgFile):
-    
-    iFile = cfgFile+'.cfg'
-    oFile = cfgFile+'.xml'
-
-    if not exists(iFile):
-        print iFile+' does not exist: exiting'
-        sys.exit()
-
-    of = open(oFile,'w\n')
-    of.write('<?xml version="1.0"?>\n')
-    of.write('<jobList>\n')
-    for line in open(iFile).readlines():
-        if line[0] == '#': continue
-        
-        tokens = line.split()
-        of.write('  <athena>\n')
-        of.write('    <options>'+tokens[0]+'</options>\n')
-        of.write('    <package>'+tokens[1]+'</package>\n')
-        of.write('    <group>'+tokens[2]+'</group>\n')
-        of.write('    <queue>'+tokens[3]+'</queue>\n')
-        if len(tokens) == 5: of.write('    <dataset>'+tokens[4]+'</dataset>\n')
-        of.write('  </athena>\n')
-    of.write('</jobList>\n')
-
-
-if __name__ == '__main__':
-
-    if len(sys.argv) != 2:
-        print "Usage: cfg2xml configFile"
-        sys.exit()
-
-    cfgFile = sys.argv[1]
-    cfg2xml(cfgFile)
-    
diff --git a/Tools/RunTimeTester/src/chainStoreMaker.py.obsolete b/Tools/RunTimeTester/src/chainStoreMaker.py.obsolete
deleted file mode 100644
index 7e98e7834d1872281b0fd9aabdf384f17ef43751..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/chainStoreMaker.py.obsolete
+++ /dev/null
@@ -1,17 +0,0 @@
-from createDirectoryDictionary import createDirectoryDictionary
-from RTTConstructionFactoryHelpers import makeArgBag
-import os
-
-def chainStoreMaker(paths, chainName, enableMake):
-    argBag = makeArgBag(['release', 'originalBranch', 'runType', 'cmtConfig', 'topProject',
-                         'otherProject', 'isNightly'], paths)
-    argBag.base = paths.chainStoreBasePath
-    
-    
-    the_path = createDirectoryDictionary(argBag, withoutReleaseName=True)['packages']
-    
-    csPath =  os.path.join(paths.chainStoreBasePath, the_path, paths.package, chainName)
-    if enableMake:
-        if not os.path.exists(csPath):
-            os.makedirs(csPath)
-    return csPath
diff --git a/Tools/RunTimeTester/src/childrenToDescendants.py b/Tools/RunTimeTester/src/childrenToDescendants.py
deleted file mode 100755
index e6ec3e1bbedfd858bdd33e343c11fd9ad6da0bcf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/childrenToDescendants.py
+++ /dev/null
@@ -1,80 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Converts a tree dictionary with structure:
-{...
-parent: [child0,, child1...]
-...
-}
-
-to 
-{...
-parent: [descendant0, descendant1...]
-...
-}
-"""
-
-
-def combine(to, fro):
-    lto = len(to)
-    # print 'comnine0:', to, fro
-    [to.append(ffro) for ffro in fro if ffro not in to]
-    # print 'comnine1:', to
-    return len(to)>lto
-    
-def link(dict):
-    dead = []
-    dlen = len(dict)
-    npass = 0
-    while dlen>len(dead)+1:
-        for key in dict.keys():
-            if key in dead: continue
-            for ent in dict[key]:
-                if not combine(dict[key],dict[ent]):
-                    if ent not in dead: dead.append(ent)
-    
-        # print 'pass: %d ' % npass
-        # print 'dead: %s ' % dead
-        # print 'dict: %s'  % d
-        # npass += 1
-
-    
-
-if __name__ == '__main__':
-    d={
-        'a':['b', 'c'],
-        'b':['d', 'e'],
-        'c':['f', 'g'],
-        'e':['j', 'k'],
-        'f':['l', 'm'],
-        'g':['n', 'o'],
-        'h':[],
-        'i':[],
-        'j':[],
-        'k':[],
-        'l':[],
-        'm':[],
-        'n':[],
-        'o':[],
-        'd':['h', 'i']
-        }
-
-    link(d)
-    [d[ent].sort() for ent in d.keys()]
-    print d
-    assert d['a'] == ['b','c','d','e','f','g','h','i','j','k','l','m','n','o']
-    assert d['b'] == ['d','e','h','i','j','k']
-    assert d['c'] == ['f','g','l','m','n','o']
-    assert d['d'] == ['h','i']
-    assert d['e'] == ['j','k']
-    assert d['f'] == ['l','m']
-    assert d['g'] == ['n','o']
-    assert d['h'] == []
-    assert d['i'] == []
-    assert d['j'] == []
-    assert d['k'] == []
-    assert d['l'] == []
-    assert d['m'] == []
-    assert d['n'] == []
-    assert d['o'] == []
-    
-
diff --git a/Tools/RunTimeTester/src/cmtLines.py.obsolete b/Tools/RunTimeTester/src/cmtLines.py.obsolete
deleted file mode 100755
index 13f8268a848d9d59d4aad0134e0978035b88d302..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/cmtLines.py.obsolete
+++ /dev/null
@@ -1,143 +0,0 @@
-import os, os.path
-from popen2                     import Popen4
-from Factory_HardCoded          import HardCodedFactory
-from RTTSException              import RTTSException
-from Factory_UserStuffRetriever import UserStuffRetrieverFactory
-# -------------------------------------------------------------------------
-# import logging
-# logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-def cmtLinesProject(jDescriptor):
-    """
-    Returns lines of script needed to set up cmt if this is being
-    done from scratch(eg for nightly running)
-    Returns an empty string otherwise
-    (for running from checked-out code, where it is
-    assumed the user has performed the cmt operations.
-    """
-
-    
-    # is the package in the local package list declared in the top level
-    # RTT config file?
-    try:
-        if jDescriptor.paths.local:
-            hCMT, pCMT = checkConditionsForLocalRunning(jDescriptor)
-        else:
-            hCMT, pCMT = checkConditionsForReleaseRunning(jDescriptor)
-    except:
-        jDescriptor.logger.error("Could not find RTT CMT dirs")
-        raise
-            
-    jDescriptor.logger.debug(
-        'cmtLines setting up for batch running (project builds)')
-
-    release       = jDescriptor.paths.release
-    oriDir        = os.getcwd()
-    
-    string  = 'unset PYTHONPATH\n'
-    string += 'cd %s\n' % hCMT
-    string += 'source setup.sh -tag=AtlasOffline,%s,opt\n' % (release)
-    string += 'cd %s\n' % pCMT
-    string += 'source setup.sh\n'
-    # if jDescriptor.paths.local:
-    #     string += 'cmt broadcast gmake\n'
-    string += 'cd %s\n' % oriDir
-    return string
-
-
-def checkConditionsForReleaseRunning(desc):
-    """
-    check we have made the project build cmt requirements files
-    for now I am going to hardwire a lot of the code. This is less than
-    ideal as we are assuming what is done in other parts of the code
-    (UserStuffRetrieverProject)
-     Need to make an USR in order to refind the cmt requirements
-     files...
-    """
-    
-    homeRequirements    = desc.paths.rttCmtHomePath
-    projectRequirements = desc.paths.rttCmtProjPath
-    hExists             = os.path.exists(homeRequirements)
-    pExists             = os.path.exists(projectRequirements)
-    msg                 = ''
-    
-    if not hExists:
-        msg += 'cmtLinesProject no file %s\n' % homeRequirements
-
-
-    if not pExists:
-        msg += 'cmtLinesProject no file %s\n' % projectRequirements
-
-    if not (hExists and pExists):
-        desc.logger.error(msg)
-        raise RTTSException(msg)
-    
-    return (homeRequirements, projectRequirements)
-
-
-def conditionsForLocalRunning(desc):
-    desc.logger.debug('cmtLines setting up for local running')
-    """
-    Local packages can only be run if
-
-    1. mode = LinuxInteractive mode
-  
-    because the user has to set up use statements, typically in the
-    TestRelease requirements file.
-    """
-        
-    modeOK    = jDescriptor.paths.mode=='LinuxInteractive'
-
-    if not modeOK: 
-        msg  = 'You must have mode LinuxInteractive '
-        msg += 'if you want to run local packages '
-        msg += 'your current mode is '
-        msg += jDescriptor.paths.mode
-        desc.logger.error(msg)
-        raise RTTSException(msg)
-        
-    return checkConditionsForReleaseRunning(desc)
-
-def cmtLinesKit(jDescriptor):
-    rel = checkConditionsForKitRunning(jDescriptor)
-    string   ='source '+os.path.join(jDescriptor.paths.distArea, 'cmtsite/setup.sh')
-    string  +=' -tag=AtlasOffline,opt,%s'%rel+'\n'
-    string  +='cd '+os.path.join(jDescriptor.paths.distArea, 'AtlasOffline')
-    string  += '/'+jDescriptor.paths.release
-    string  +='/AtlasOfflineRunTime/cmt\n'
-    string  += 'source setup.sh\n'
-    return  string
-
-def checkConditionsForKitRunning(jDescriptor):
-    kitCMTPath = os.path.join(jDescriptor.paths.distArea, 'cmtsite/setup.sh')
-    if not os.path.exists(kitCMTPath):
-        msg = kitCMTPath +' does not exist\n'
-        jDescriptor.logger.error(msg)
-        raise RTTSException(msg)
-    
-    release = jDescriptor.paths.release
-    parts   = release.split('.')
-    
-    if len(parts) !=3 :
-         msg = 'Running against kits for nightlies not supported yet!\n'
-         raise RTTSException(msg)
-     
-    #if jDescriptor.paths.isNightly:
-    #    msg = 'Running against kits for nightlies not supported yet!\n'
-    #    raise RTTSException(msg)
-    
-    rel = jDescriptor.paths.release[1:] 
-    return rel
-    
-
-if __name__ =="__main__":
-    
-    runSetup=Popen4('/usr/local/bin/python2.3 /unix/atlas/RTTdev/UsefulTools/setCompilerVersion.py')
-    runSetup.wait()
-
-    sourceTemp=Popen4('source tempSetup.sh')
-
-    os.system("echo $CMTCONFIG")
-    
diff --git a/Tools/RunTimeTester/src/configureCMTDirs.py b/Tools/RunTimeTester/src/configureCMTDirs.py
deleted file mode 100644
index 445a3d99de5006b09d9a5530a0487302390bbc2a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/configureCMTDirs.py
+++ /dev/null
@@ -1,19 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from ShellCommand import ShellCommand
-
-def configureCMTDirs(localRTTRun, cmtConfigurer, logger):
-    # Not necessary if this is a local run
-    if localRTTRun:
-        logger.info('Local run, not configuring CMT dir')
-        return
-    
-    cmtConfigurer.makeHomeCMTDir()
-    
-    nlines = -1 # do not limit the number of lines
-    cmds = [cmtConfigurer.configure()]
-    logger.info('cmds to configure cmt dir:\n%s' %str(cmds))
-    shellCom = ShellCommand(cmds, logger, nlines)
-    reply = shellCom.getReply()
-    m = 'Reply on configuring CMT: %s' % str(reply)
-    logger.info(m)
diff --git a/Tools/RunTimeTester/src/createDirectoryDictionary.py b/Tools/RunTimeTester/src/createDirectoryDictionary.py
deleted file mode 100644
index 484a77374671deac4be090991be34200901e08f4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/createDirectoryDictionary.py
+++ /dev/null
@@ -1,54 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTSException import RTTInputError
-from RTTpath       import RTTpath
-
-def getSegs():
-    return   {'AtlasHLT'        : 'hlt',
-              'AtlasProduction' : 'offline',
-              'AtlasPoint1'     : 'point1',
-              'AtlasTier0'      : 'tier0',
-              'AtlasP1HLT'      : 'p1hlt'
-              }
-
-def findSeg(project): return  getSegs()[project]
-
-def createDirectoryDictionary(argBag, withoutReleaseName=False):
-    """Sets up a dictionary containing the various paths concerned with
-    results.
-    e.g. <resBase>/12.4.0/kit/x86_64_slc4_gcc34_opt
-    <resBase>/rel_0/bug/kit/x86_64_slc4_gcc34_opt
-    """
-        
-        
-    dict = {}
-    
-    path = RTTpath(argBag.base)
-    dict['base'] = str(path)
-    
-    if not withoutReleaseName:
-        path.postpend(argBag.release)                  # 12.4.0, rel_0
-
-    if argBag.isNightly: path.postpend(argBag.originalBranch) # dev or bugfix
-
-    path.postpend(argBag.runType)                  # kit or build
-    path.postpend(argBag.cmtConfig)                # target CMT config
-
-    segs = getSegs()
-    if not (argBag.topProject in segs or argBag.otherProject in segs):
-        msg = 'Unknown top/otherProject to path mapping \ntopProject:%s \notherProject:%s\nmapping:%s' % (
-            str(argBag.topProject), str(argBag.otherProject), str(segs))
-        raise RTTInputError(msg)
-
-    if argBag.topProject in segs:
-        pathSegment = segs[argBag.topProject]
-        if argBag.otherProject in segs:
-            pathSegment  += '_' + segs[argBag.otherProject]
-    else:
-        pathSegment = segs[argBag.otherProject]
-            
-    path.postpend(pathSegment)
-
-    dict['packages'] = str(path)
-    
-    return dict
diff --git a/Tools/RunTimeTester/src/createKits.py b/Tools/RunTimeTester/src/createKits.py
deleted file mode 100644
index d6dabbfab8390ee8105ebd449f55c0fefd1f073b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/createKits.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from kitFromConfFile   import kitFromConfFile
-from exc2string2       import exc2string2
-from JobGroupsParser   import makeTopLevelJobGroupKits
-from JobGroupKits      import JobGroupKits
-from RTTSException     import RTTCodingError
-
-def createKits(cvsPackages, jobGroupConfig, logger):
-    try:
-        userDefinedKits = kitFromConfFile(cvsPackages)
-    except Exception, e:
-        m = 'createKits: Exception %s\nTraceback:\n' % (str(e), exc2string2())
-        logger.critical(m)
-        raise RTTCodingError(m)
-
-    # ---------------------------------------------------------
-    # get dictionary for valid rtt supplied conf files
-    # ---------------------------------------------------------
-    rttDefinedKits = makeTopLevelJobGroupKits(jobGroupConfig)
-    
-    # ---------------------------------------------------------
-    # make a singleton kit distributor
-    # ---------------------------------------------------------
-    return JobGroupKits(rttDefinedKits, userDefinedKits)
-     
diff --git a/Tools/RunTimeTester/src/createRunDir.py.obsolete b/Tools/RunTimeTester/src/createRunDir.py.obsolete
deleted file mode 100755
index b81fbaf06b7b6b89bea1fbcd33a3bede88990734..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/createRunDir.py.obsolete
+++ /dev/null
@@ -1,19 +0,0 @@
-#===== RTT MODULES =====
-from Tools                   import mkdirIfNew, rmdirMkdir
-#===== PYTHON MODULES ==
-import os
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def createRunDir(jDescriptor):
-
-    """
-    Creates run directory (path calculated from release, build,
-    job options file name..)
-    """
-    logger.debug("entering createRunDir")
-
-    rmdirMkdir(jDescriptor.runPath, 0777)
diff --git a/Tools/RunTimeTester/src/cullJobDescriptors.py b/Tools/RunTimeTester/src/cullJobDescriptors.py
deleted file mode 100644
index 96007613a450dae42e186c3ce4a57325d668e736..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/cullJobDescriptors.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-def cullJobDescriptors(descs, isNightly, limitedNightlyVeto, rttPilotJobs):
-    if isNightly:
-        descs = [d for d in descs if not d.__dict__['numberedOnly']]
-    if limitedNightlyVeto:
-        descs = [d for d in descs if not d.__dict__['limitedNightlies']]
-    if not rttPilotJobs:
-        descs = [d for d in descs if not d.__dict__['rttPilotJob']]
-    return descs
-    
diff --git a/Tools/RunTimeTester/src/cullPackages.py b/Tools/RunTimeTester/src/cullPackages.py
deleted file mode 100644
index f6e84e86c7cefd2e0f799d56c8e254b6995e927a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/cullPackages.py
+++ /dev/null
@@ -1,18 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-def cullPackages(cvsPackages, logger):
-
-    packagesToRun = [p for p in cvsPackages if p.runMyJobs()]
-    nAll = len(cvsPackages)
-    nToRun = len(packagesToRun)
-    m  = 'About to make TestRuns for '
-    m += '%d packages\n' % (nToRun)
-    m += '%d packages out of %d were ' % ((nAll-nToRun), nAll)
-    m += 'dropped due to user selection\n'
-    m += 'retained packages:\n'
-    m += str([p.name for p in packagesToRun])
-
-    logger.info(m)
-
-    return packagesToRun
diff --git a/Tools/RunTimeTester/src/dataFiles2String.py.obsolete b/Tools/RunTimeTester/src/dataFiles2String.py.obsolete
deleted file mode 100755
index 27eb3e991b3e063c1061ba42e4b933e947b59a14..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/dataFiles2String.py.obsolete
+++ /dev/null
@@ -1,16 +0,0 @@
-import os
-
-def dataFiles2String(jDescriptor):
-
-    s = ''
-
-    for datasetList in jDescriptor.datasets.values():
-        for file in datasetList:
-            s += os.path.basename(file.name) + ' '
-
-                                 
-        
-    return s
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/dbgF.py.obsolete b/Tools/RunTimeTester/src/dbgF.py.obsolete
deleted file mode 100755
index 5d5868182c60ebc4bd73a901c82a90f285d35651..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/dbgF.py.obsolete
+++ /dev/null
@@ -1,12 +0,0 @@
-#! /usr/local/bin/python
-"""
-Formats and time stamps  debug statements
-"""
-
-from Timer            import timer
-
-
-def dbgF(obj, text):
-    line = timer()+" "+(obj.__class__).__name__+": "+text
-    print line
-    
diff --git a/Tools/RunTimeTester/src/dbgFf.py.obsolete b/Tools/RunTimeTester/src/dbgFf.py.obsolete
deleted file mode 100755
index d348caf023ec0e03986674e62f61bdf8aab7679e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/dbgFf.py.obsolete
+++ /dev/null
@@ -1,12 +0,0 @@
-#! /usr/local/bin/python
-"""
-Formats and time stamps  debug statements
-"""
-
-from Timer            import timer
-
-
-def dbgFf(fName, text):
-    line = timer()+" "+fName+": "+str(text)
-    print line
-    
diff --git a/Tools/RunTimeTester/src/dump_garbage.py.obsolete b/Tools/RunTimeTester/src/dump_garbage.py.obsolete
deleted file mode 100755
index d2c28180de165a882bdfc3dc533ecb8c364d6544..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/dump_garbage.py.obsolete
+++ /dev/null
@@ -1,23 +0,0 @@
-import gc
-
-def dump_garbage():
-
-    gc.collect()
-    print gc.garbage
-    msg ='Garbage:\n'
-    for x in gc.garbage:
-        s = str(x)
-        if len(s) > 80: s = s[:77]+'...'
-        msg+= '%s\n   %s' % (type(x), s)
-
-    return msg
-
-
-if __name__ == '__main__':
-    gc.enable()
-    gc.set_debug(gc.DEBUG_LEAK)
-    l = []
-    l.append(l)
-    del l
-
-    print dump_garbage()
diff --git a/Tools/RunTimeTester/src/exc2string.py b/Tools/RunTimeTester/src/exc2string.py
deleted file mode 100755
index 1190efa39c71af97b83b8407a7e0b0010ae28b98..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/exc2string.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string, sys, traceback
-
-def exc2string(list):
-    list   = sys.exc_info()
-    errmsg  = str(list[0])
-    errmsg += str(list[1])
-    errmsg += string.join(traceback.format_tb(list[2]))
-    return errmsg
-                
diff --git a/Tools/RunTimeTester/src/exc2string2.py b/Tools/RunTimeTester/src/exc2string2.py
deleted file mode 100755
index 30bdab75d2c0ffaf3124997b7fe32e650ca3428f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/exc2string2.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import string, sys, traceback
-
-def exc2string2():
-    list   = sys.exc_info()
-    errmsg  = str(list[0])
-    errmsg += str(list[1])
-    errmsg += string.join(traceback.format_tb(list[2]))
-    return errmsg
-                
diff --git a/Tools/RunTimeTester/src/extractDCubeBaseDir.py b/Tools/RunTimeTester/src/extractDCubeBaseDir.py
deleted file mode 100644
index d50212aa13ad544dd5d8d0963b95ceae2b4b6ec3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/extractDCubeBaseDir.py
+++ /dev/null
@@ -1,12 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Evaluate                     import Evaluate
-from xml.dom.minidom              import parse
-
-import os
-def extractDCubeBaseDir(dCubeCfgFile):
-    if not os.path.exists(dCubeCfgFile): return 'unknown'
-    
-    theDom = parse(dCubeCfgFile)
-    return Evaluate('ref_dir/text()', theDom.documentElement)
-
diff --git a/Tools/RunTimeTester/src/findInstalledDirs.py.obsolete b/Tools/RunTimeTester/src/findInstalledDirs.py.obsolete
deleted file mode 100755
index 04a6582abcba3beda61fba4307aef88e46c4896d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findInstalledDirs.py.obsolete
+++ /dev/null
@@ -1,96 +0,0 @@
-import os, os.path
-from os.path import join
-
-import hardCoded
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-def findInstallAreas(local, releasePath):
-    """
-    local - must be a boolean
-    releasePath - string that represents a path
-    returns a list of strings representing paths
-    
-    local = true: find local installed area from CMTPATH env variable
-    Always add the release installed area.
-    """
-    
-    paths = []
-    installArea = hardCoded.installAreaDirName()
-
-    if local:
-        if 'CMTPATH' in os.environ.keys():
-            cmtpaths = os.environ['CMTPATH'].split(':')
-            cmtpaths = [join(path, installArea) for path in cmtpaths]
-            paths.extend(cmtpaths)
-
-
-    paths.append(join(releasePath, installArea))
-    paths = [path for path in paths if os.path.exists(path)]
-    
-    logger.info('installed dirs'+str(paths))
-    return paths
-
-def findInstalledDirs(isLocal, releasePath, directory):
-    """
-    local - must be a boolean
-    releasePath - string that represents a path
-
-
-    Returns a list of strings representing full directory paths
-    hanging from Installed.
-    The installed directories are looked for in the directories given
-    in CMTPATH."""
-
-    paths = findInstallAreas(isLocal, releasePath)
-    paths = [join(path, directory) for path in paths]
-    paths = [path for path in paths if os.path.exists(path)]
-
-    logger.debug('local dirs included:'+str(paths))
-    return paths
-
-def findOptionsPaths(isLocal, releasePath):
-    """
-    Returns a list of paths to be used to find option files.
-    The most local path is given first, so in locating option files
-    the path should be traversed from beginning to end.
-    """
-
-    return findInstalledDirs(isLocal, releasePath, 'jobOptions')
-
-def findSharePaths(isLocal, releasePath):
-    """
-    Returns a list of paths that will act as a search list.
-    """
-    return findInstalledDirs(isLocal, releasePath, 'share')
-
-
-if __name__ == '__main__':
-    "Test function"
-
-    import sys
-    from formatCollection import formatCollection
-    
-    if  'CMTPATH' not in os.environ.keys():
-        print 'run home CMT setup.sh before this test'
-        sys.exit()
-    codeStatuses=['Installed', 'CheckedOut']
-    dirs = ['jobOptions', 'share']
-    release = '/afs/cern.ch/atlas/software/dist/9.0.2'
-
-
-    print '\n\njobOptions paths'
-    for status in codeStatuses:
-        print status
-        print formatCollection(findOptionsPaths(status,release))
-
-
-    
-    print '\n\nshare paths'
-    for status in codeStatuses:
-        print status
-        print formatCollection(findSharePaths(status,release))
-
-        
-        
diff --git a/Tools/RunTimeTester/src/findJobOptionsSearchPath.py b/Tools/RunTimeTester/src/findJobOptionsSearchPath.py
deleted file mode 100644
index cf2347cdd15e435cd5dc20b2d98e3b1b7a6d9231..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findJobOptionsSearchPath.py
+++ /dev/null
@@ -1,59 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from ShellCommand import ShellCommand
-
-from RTTSException import RTTCodingError
-from RTTSException import RTTInputError
-
-class FindJobOptionsSearchPathError(Exception):
-    def __init__(self, msg): self.msg = msg
-    def __str__(self): return str(self.msg)
-    
-def findJobOptionsSearchPath(cmds, paths, logger):
-
-    logger.debug('Calculating JOBOPTIONSSEARCHPATH')
-
-    # cmds.extend(['source setup.sh'])
-    cmds.extend(['echo $JOBOPTSEARCHPATH', 'echo $PATH'])
-    
-    nlines = -1 # do not limit the number of lines
-    shellCom = ShellCommand(cmds, logger, nlines)
-    reply = shellCom.getReply()
-    #m = 'Reply on setting up CMT to find JO searchpath: %s' % str(reply)
-    #logger.debug(m)
-    searchPath = ''
-    # the env variable path will be used e.g. to locate pre athena scripts
-    PATH       = '' 
-    
-    if len(reply)>1:
-        searchPath = reply[-2]
-        PATH = reply[-1]
-        # These paths are critical to all Athena jobs. As these are most
-        # of the RTT jobs, give up if the search path is not set.
-    else:
-        m = """Could not obtain env vars JOBOPTSEARCHPATH and PATH,
-        cmds sent to seek env variables:
-        %s
-        Reply:  %s
-        abandoning run """ % (str(cmds), str(reply))
-        logger.critical(m)
-        raise RTTInputError(m)
-    
-    # logger.debug('Found JOBOPTSEARCHPATH: %s' % searchPath)
-    # logger.debug('Found PATH: %s' % PATH)
-
-    def jobOpSearchPathSepChar():
-        if paths.isNightly: return ':'
-        if int(paths.release.split('.')[0]) < 13: return ','
-        return ':'
-    
-    
-    
-    searchPaths = {
-        'jobOptionsSearchPath': {'path':                    searchPath,
-                                 'separater':  jobOpSearchPathSepChar()},
-        'PATH':                 {'path':                           PATH,
-                                 'separater':                       ':'}
-        }
-    
-    return searchPaths
diff --git a/Tools/RunTimeTester/src/findNightlyRelease.py b/Tools/RunTimeTester/src/findNightlyRelease.py
deleted file mode 100755
index 4ed86b40dfcc1d2c2ea7453c0cfd95cd87ecc6b5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findNightlyRelease.py
+++ /dev/null
@@ -1,41 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, time
-import logging
-
-def findNightlyRelease(stampObject, logger):
-    """If release is set to 'nightly', find out what is the current release
-    and update the confDict dictionary accordingly."""
-
-    (done, release) = stampObject.findNightlyRelease()
-        
-    waitTime = 5 # mins
-    waitmsg  = 'NICOS nightly flag not currently available. '
-    waitmsg += 'Will check again in %s minutes.' % str(waitTime)
-        
-    if done:
-        msg  = 'Found the nightly release immediately: %s' % release
-        logger.info(msg)
-        return release
-
-    logger.info(waitmsg)
-    [logging.FileHandler.flush(handler) for handler in logger.handlers]
-
-    while not done:
-        waitTime = 15
-        # time.sleep(60*waitTime)
-        time.sleep(waitTime)
-        (done, release) = stampObject.findNightlyRelease()
-                
-        if not done:
-            if stampObject.giveUpWaitingForNightly():
-                msg = 'NICOS stamp file will not come, giving up. Over and out.'
-                logger.info(msg)
-                return None
-            logger.info(waitmsg)
-            [logging.FileHandler.flush(handler) for handler in logger.handlers]
-        else:
-            msg  = 'Found the nightly release: %s' % release
-            logger.info(msg)
-
-    return release
diff --git a/Tools/RunTimeTester/src/findOptionsPaths.py.obsolete b/Tools/RunTimeTester/src/findOptionsPaths.py.obsolete
deleted file mode 100755
index f506ce431b2a000077a37d708a1650ffaf72b7ac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findOptionsPaths.py.obsolete
+++ /dev/null
@@ -1,50 +0,0 @@
-import os, os.path
-from os.path import join
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def findOptionsPaths(codeStatus, releasePath):
-    """
-    Returns a list of paths to be used to find option files.
-    The most local path is given first, so in locating option files
-    the path should be traversed from beginning to end.
-
-    The function checks the codestatus variable to determine
-    if local Install areas might exist. If the variable is set to
-    'checkedOut', it slices $CMTPATH, and checks for the existence of
-    Install areas. If these are found it adds them to the list.
-    If the codestatus is 'Installed' the options are looked for in the release
-    jobOptions directory.
-    """
-
-    optionsPaths = []
-    
-    if codeStatus == 'CheckedOut':
-        if 'CMTPATH' in os.environ.keys():
-            cmtpaths = os.environ['CMTPATH'].split(':')
-            cmtpaths = [join(path,'InstallArea/jobOptions') for path in cmtpaths]
-            cmtpaths = [path for path in cmtpaths if os.path.exists(path)]
-            optionsPaths.extend(cmtpaths)
-        optionsPaths.append(join(releasePath, 'InstallArea/jobOptions'))
-    else:   
-        optionsPaths.append(join(releasePath, 'InstallArea/jobOptions'))
-
-    
-    return optionsPaths
-
-if __name__ == '__main__':
-    "Test function"
-
-    import sys
-    if  'CMTPATH' not in os.environ.keys():
-        print 'run home CMT setup.sh before this test'
-        sys.exit()
-    codeStatuses=['Installed', 'CheckedOut']
-    for status in codeStatuses:
-        print findOptionsPaths(status, '/afs/cern.ch/atlas/software/dist/8.4.0')
-                       
-        
-        
diff --git a/Tools/RunTimeTester/src/findPackages.py b/Tools/RunTimeTester/src/findPackages.py
deleted file mode 100755
index 9096efa7b0c985d283ce100dca987406228c80fd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findPackages.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path, string
-from Paths import Paths
-
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def findPackages(root, releasePath = '', dirs='src;cmt'):
-    """
-    findPackages returns a list of directories below a root directory 'root'
-    which:
-    1 contain the directories listed in the semi-colon separated string 'dirs'
-    2 contain a directory with the same name as the current one once any -xx is dropped
-      (ie Atlfast-xx-yy-zz needs to contain Atlfast)
-    """
-
-    class Bunch:
-        def __init__(self, **kwds):
-            self.__dict__.update(kwds)
-        
-    dirList=dirs.split(';')
-    arg = Bunch(dirList=dirList, results=[])
-    
-    def visit(arg, dirname, files):
-        for file in files:
-            fullname = os.path.normpath(os.path.join(dirname, file))
-            if os.path.isdir(fullname):
-                if file.find('-') != -1:
-                    dirList.append(file.split('-')[0])
-                    # nDirList.append(file.split('-')[0])
-                    alist = [dir for dir in dirList if dir in os.listdir(fullname)]
-                    if alist == dirList:
-                        arg.results.append(string.replace(dirname, releasePath,'')[1:])
-                        logger.debug(string.replace(dirname, releasePath,'')[1:])
-                        
-                    dirList.pop()
-                    
-    os.path.walk(root, visit, arg)
-
-    return arg.results
-
-if __name__ == '__main__':
-    import sys
-
-    
-    print sys.argv[1]
-    findPackages(sys.argv[1])
diff --git a/Tools/RunTimeTester/src/findRTThome.py b/Tools/RunTimeTester/src/findRTThome.py
deleted file mode 100755
index b6498ad516f29b247d5418a70aa3e4420c20e26e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/findRTThome.py
+++ /dev/null
@@ -1,41 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-import os
-
-def findRTThome():
-    userDictPetersMac={
-        'cfg': '/Users/peter/RTTdevelopment/RTTChain/cfg/cfg.xml',
-        'src': '/Users/peter/TestArea/RunTimeTester-00-02-54_2/src',
-        'base': '/Users/peter/TestArea/RunTimeTester-00-02-54_2/'
-        }
-
-
-    userDicts = {
-        'petersMac':     userDictPetersMac
-        }
-
-    here = 'pcucl02'
-    if os.environ.get('HOSTNAME', '').find('.ucl.') != -1: here = 'ucl'
-    
-    print 'printing HOST',os.environ.get('HOST', '')
-
-    if os.environ.get('HOST', '').find('peter-4.home') != -1: here = 'petersMac' 
-
-    user = os.environ.get('USER', '')
-    host = os.environ.get('HOSTNAME', '')
-
-    print here, user
-    userDict = userDicts[here]
-    cfg  = userDict['cfg']
-    src  = userDict['src']
-    base = userDict['base']
-    
-    return base, cfg, src
-
-    
-if __name__ == '__main__':
-    (base, cfg, src) = findRTThome()
-    print base
-    print cfg
-    print src
diff --git a/Tools/RunTimeTester/src/formatCollection.py b/Tools/RunTimeTester/src/formatCollection.py
deleted file mode 100755
index da19213c494cb88af961b3cf0d8ac10d37aaafaf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/formatCollection.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from string import replace
-def formatCollection(c):
-    return replace(str(c),',',',\n  ')
diff --git a/Tools/RunTimeTester/src/getInheritanceHeirarchy.py b/Tools/RunTimeTester/src/getInheritanceHeirarchy.py
deleted file mode 100755
index f9c62992bd9ffabd1493f7adfe3099153e877bfa..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/getInheritanceHeirarchy.py
+++ /dev/null
@@ -1,52 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-def getInheritanceHeirarchy(object):
-    classes = [object.__class__]
-    def getParents(classes):
-        total = []
-        for c in classes:
-            parents = c.__bases__
-            if parents:
-                total += parents
-                total += getParents(parents)
-        return total
-    
-    classes += getParents(classes)
-    return [c.__name__ for c in classes]
-
-def isWorkerMinder(minder):
-    if 'WorkerJobMinder' in getInheritanceHeirarchy(minder):
-        return True
-    return False
-
-def isWatcherMinder(minder):
-    if 'WatcherJobMinder' in getInheritanceHeirarchy(minder):
-        return True
-    return False
-
-def isAthenaJob(descriptor):
-    if 'AthenaJobDescriptor' in getInheritanceHeirarchy(descriptor):
-        return True
-    return False
-
-if __name__ == '__main__':
-
-    class A:
-        def __init__(self):
-            
-            self.i =1
-    class B(A):
-        def __init__(self):
-            A.__init__(self)
-        
-            self.j=2
-    class C(B):
-        def __init__(self):
-            B.__init__(self)
-            k=3
-
-    b = B()
-    c = C()
-
-    print getInheritanceHeirarchy(c)
-
diff --git a/Tools/RunTimeTester/src/getJobGroupKit.py b/Tools/RunTimeTester/src/getJobGroupKit.py
deleted file mode 100644
index cff6620f07bf0661b456caa7027b28033286eec9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/getJobGroupKit.py
+++ /dev/null
@@ -1,37 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from exc2string2     import exc2string2
-from RTTSException   import RTTCodingError
-from JobGroupKits    import JobGroupKits
-
-
-import logging
-logger = logging.getLogger('rtt')
-
-def getJobGroupKit(jobGroup, jobGroupKits):
-    
-    jobGroupKit = None
-    msg ='Could not obtain a kit with name '
-
-    try:
-        kit    = jobGroupKits.kit(jobGroup)
-    except:
-        msg = 'getJobGroupKit: Failed to  obtain a kit with name %s' % jobGroup
-        print msg
-    else:
-        return kit
-        
-    if not jobGroupKit:
-        msg = 'getJobGroupKit: Attempting to  obtain a kit with name %s' % jobGroup
-        print msg
-        jobGroup = 'Athena'+jobGroup
-        try:
-            kit = jobGroupKits.kit(jobGroup)
-        except:
-            msg = 'getJobGroupKit: Failed to obtain a kit with name %s\nTraceback:\n%s' % (jobGroup, exc2string2())
-            logger.error(msg)
-            raise RTTCodingError(msg)
-        else:
-            msg = 'getJobGroupKit: Succeeded in obtain a kit with name %s' % jobGroup
-            logger.debug(msg)
-            return kit
diff --git a/Tools/RunTimeTester/src/getTimes.py b/Tools/RunTimeTester/src/getTimes.py
deleted file mode 100644
index bb35798e1c472b9e93279426907a0c25dd409a9d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/getTimes.py
+++ /dev/null
@@ -1,6 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import time
-def getTimes():
-    formatedTime = time.strftime('%y/%m/%d@%H:%M', time.localtime())
-    return (formatedTime, time.time())
diff --git a/Tools/RunTimeTester/src/get_release_2017.py b/Tools/RunTimeTester/src/get_release_2017.py
deleted file mode 100644
index 57a4743fb0a8ef33d5539f1248b31fbd0dd8568e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/get_release_2017.py
+++ /dev/null
@@ -1,195 +0,0 @@
-"""get_release: returns the path to a build provided that such a build
-exists, and that it is recent."""
-
-import os
-import re
-from time import sleep
-from datetime import (datetime,
-                      time,
-                      timedelta)
-import PathConstants
-
-fn_re = re.compile(r'^(?P<year>20\d\d)-(?P<month>\d\d)-(?P<day>\d\d)T\d\d\d\d$')
-build_day_start = 20
-
-def fn2dt(fn): return datetime.strptime(fn, "%Y-%m-%dT%H%M")
-
-def fn2asetup_date(fn):
-    m = fn_re.match(fn)
-    return 'r%s-%s-%s' % (m.group('year'), m.group('month'), m.group('day'))
-             
-def _get_build_in_period(branch, period_start, period_end):
-    """
-    period_end - end of build persiod under conssideration:
-                 either build_day_start for today or for tomorrow
-                 depending on the start time of the RTT.
-
-    full dir eg 
-    /cvmfs/atlas-nightlies.cern.ch/repo/sw/master/2017-03-09T2245
-    """
-
-    # find if any branchs are recent
-    basedir = PathConstants.branchdir_2017
-    found = [fn for fn in os.listdir(os.path.join(basedir, branch))
-             if fn_re.match(fn)]
-
-    def is_recent(fn):
-        fn_dt = fn2dt(fn)
-        return  fn_dt >= period_start and fn_dt < period_end 
-
-
-    # print found
-    recent = [f for f in found if is_recent(f)]
-    if not recent:
-        print 'no file available for period %s - %s in %s' % (
-            str(period_start),
-            str(period_end),
-            os.path.join(PathConstants.branchdir_2017, branch)
-        )
-        return ''
-    # examine build for matching cmt config paramters
-    # start search form the most recent, to back in time.
-
-    print 'recent', recent
-    recent = sorted(
-        recent,
-        cmp=lambda x,y: cmp(fn2dt(x), fn2dt(y)),
-        reverse=True)
-    return recent[0]
-
-
-def _get_release(release_datetime):
-    """return the release (rel_X) for a datetime within the current
-    build persiod. The current build period starts at the build_start_time
-    of the same day if the rtt starts up at a time after the build_start_time,
-    otherwise the build_start_time of the previous day"""
-
-    
-    # python calls Sunday 6, Monday 0, ATLAS: Sunday 0, Monday 1
-    rel = release_datetime.weekday() + 1
-    if rel == 7: rel = 0
-
-    # but if the build starts after build_day_start, consider
-    # it to be the next day
-
-    # so, a build starting after build_day_start on sunday
-    # will be assigned 6 by python, converted to 0 above.
-    # further it will be reassigned to 1 as it starts after
-    # build_day_start
-    if release_datetime.hour >= build_day_start: rel += 1
-    if rel == 7: rel = 0
-
-    return 'rel_%d' % rel
-
-def get_period_start(dt):
-    
-    result = dt.replace(
-        hour=build_day_start,
-        minute=0,
-        second=0,
-        microsecond=0)
-
-    if dt.hour < build_day_start:
-        result = result - timedelta(days=1)
-
-    return result
-
-
-def get_release(branch, cmtconfig, check_datetime=None):
-    """Return a release (string matching rel_[0,6]) if a build occurs
-    during a period of interest, '' otherwise.
-    If the RTT check time is after build_day_start, the period of
-    interest starts at the same day at build_day_start, otherwise
-    at build_day_start of the previous day. It terminates 24 hours later.
-
-    If no such build exists, the function waits for it to arrive, or
-    the current time exits the persiod of interest.
-    """
-    if check_datetime is None: check_datetime = datetime.now()
-    
-    period_start = get_period_start(check_datetime)
-    print period_start
-    period_end = period_start + timedelta(days=1)
-
-    while True:
-        fn  = _get_build_in_period(branch,
-                                   period_start,
-                                   period_end)
-        if not fn:
-            print 'get_release_2017 no file available, sleeping'
-            sleep(60*5) # no file available, sleep 5 mins
-        else:
-            return (fn, _get_release(fn2dt(fn)), fn2asetup_date(fn))
-    
-if __name__ == '__main__':
-    import sys
-    branch = sys.argv[1]
-    cmtconfig = sys.argv[2]
-
-    # eg python get_release_2017.py master x86_64-slc6-gcc49-opt
-
-
-
-    print 'calling _get_build_in_period', branch, cmtconfig
-    period_start = get_period_start(datetime.now())
-    period_end = period_start + timedelta(days=1)
-    
-    fn = _get_build_in_period(branch, period_start, period_end)
-    print 'release file: -->%s<--' % fn
-
-    # the call to get release waits for a suitable build to appear,
-    # so may hang.
-    # fn, rel = get_release(datetime.now(), branch, cmtconfig)
-    # print 'ran get_release', branch, cmtconfig
-    # print fn, rel
-
-    print 'calling _get_release'
-
-    early = build_day_start - 1
-    late = build_day_start + 1
-
-    assert early in range(24)
-    assert late in range(24)
-    
-    sunday_early = datetime(year=2017,
-                              month=3,
-                              day=12,
-                              hour=early)
-
-    sunday_late = sunday_early.replace(hour=late)
-
-    assert _get_release(sunday_early) == 'rel_0'
-    assert _get_release(sunday_late) == 'rel_1'
-
-
-    print '\ntemp debug\n'
-    now = datetime.now()
-    now_1 = now - timedelta(days=1)
-    now_2 = now - timedelta(days=2)
-    now_3 = now - timedelta(days=3)
-
-    conf = 'x86_64-slc6-gcc62-opt'
-    branch = 'master'
-
-    print branch, conf
-    print get_release(branch, conf)
-    print get_release(branch, conf, now)
-    print get_release(branch, conf, now_1)
-    print get_release(branch, conf, now_2)
-    print get_release(branch, conf, now_3)
-
-
-    branch = '21.0'
-    print 
-    print branch, conf
-    print
-    print get_release(branch, conf)
-    print get_release(branch, conf, now)
-    print get_release(branch, conf, now_1)
-    print get_release(branch, conf, now_2)
-    print get_release(branch, conf, now_3)
-
-    print 'ok'
-
-    
-    
diff --git a/Tools/RunTimeTester/src/hardCoded.py.obsolete b/Tools/RunTimeTester/src/hardCoded.py.obsolete
deleted file mode 100755
index 8182250f5ed8a8efbef3c32c7e8f28096e4fff49..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/hardCoded.py.obsolete
+++ /dev/null
@@ -1,82 +0,0 @@
-from os.path import join
-
-def SLC3_string():
-    return 'slc3'
-
-def RH73_string():
-    return 'rh73'
-
-def afsDistBase(which):
-    base = {'release':'/afs/cern.ch/atlas/software/dist',
-            #'kit':'/afs/cern.ch/atlas/software/installed/rel/dist'
-           'kit':'/afs/cern.ch/atlas/software/releases'}
-
-    theBase = ''
-    try:
-        theBase = base[which]
-    except KeyError:
-        pass
-
-    return theBase
-
-def installAreaDirName():
-    return 'InstallArea'
-
-def nightliesAFSArea():
-    return join(afsDistBase('release'),'nightlies')
-
-def fullNightlyPathDict():
-    dictionary = {'N.X.0': join(afsDistBase('release'),nightlyDirDict()['N.X.0']),
-                  'N.0.X': join(afsDistBase('release'),nightlyDirDict()['N.0.X']),
-                  '10.0.X': join(afsDistBase('release'),nightlyDirDict()['10.0.X'])
-                  }
-
-    return dictionary
-
-def branchToWebPageTextMapping():
-    """Given a branch, one can get the text to be written on RTTpage1 in the branch header row.
-    An added feature is that if a branch does not feature in this dictionary, then it will
-    not be displayed on RTTpage1."""
-    
-    theDict = {'N.0.X':'11.0.X nightlies','10.0.X':'10.0.X nightlies','Numbered':'Numbered releases','N.X.0':'11.X.0 nightlies'}
-    return theDict
-
-def nightlyDirDict():
-    return {'N.X.0':'nightlies/rel', '10.0.X':'nightlies/rd', 'N.0.X':'nightlies/11.0.X'}
-
-def latestCopiedReleaseLinkName():
-    # NOTE: Alex does not build an SLC3 dbg for N.0.X; RTT should stop if we try and run that
-    dictionary = {(SLC3_string(),'opt','10.0.X'):'latest_copied_releaseSlcProd',
-                  (SLC3_string(),'opt','N.X.0'):'latest_copied_releaseSlcOpt',
-                  (SLC3_string(),'dbg','N.X.0'):'latest_copied_releaseSlcDbg',
-                  (SLC3_string(),'opt','N.0.X'):'latest_copied_releaseSlc110X',
-                  (SLC3_string(),'dbg','N.0.X'):'latest_copied_releaseSlc110XDbg',
-                  }
-    return dictionary
-
-def nicosNightlyWebPageURL():
-    urlPrefixDict = {('N.X.0','opt'):'http://atlas.web.cern.ch/Atlas/GROUPS/SOFTWARE/OO/dist/nightlies/atlaswwwSlcOpt/',
-                     ('10.0.X','opt'):'http://atlas.web.cern.ch/Atlas/GROUPS/SOFTWARE/OO/dist/nightlies/atlaswwwSlcProd/',
-                     ('N.0.X','opt'):'http://atlas.web.cern.ch/Atlas/GROUPS/SOFTWARE/OO/dist/nightlies/nicos_web_areaSlc110X',
-                     ('N.0.X','dbg'):'http://atlas.web.cern.ch/Atlas/GROUPS/SOFTWARE/OO/dist/nightlies/nicos_web_areaSlc110XDbg'}
-
-    return urlPrefixDict
-
-def nicosNumberedWebPageURL(rel,build):
-    prefix = 'http://atlas.web.cern.ch/Atlas/GROUPS/SOFTWARE/OO/dist/'
-    suffix = {'opt':'/Log/slcoptwww/','dbg':'/Log/slcwww/'}
-
-    return prefix + str(rel) + suffix[build]
-
-    
-def nightlyStampDir(branch):
-    return join(afsDistBase('release'),nightlyDirDict()[branch]) 
-
-def pathFragmentToReleaseNotes(release):
-    return join(release,'cmt/ReleaseNotes.txt')
-
-def AFS_release_string():
-    return 'AFS_Release'
-
-def AFS_kit_string():
-    return 'AFS_Kit'
diff --git a/Tools/RunTimeTester/src/installBaseDirectories.py b/Tools/RunTimeTester/src/installBaseDirectories.py
deleted file mode 100644
index f0c0b3b2f6f674534711ea95bdc1dccefffe0586..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/installBaseDirectories.py
+++ /dev/null
@@ -1,15 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from Factory_BaseDirectoryMaker import BaseDirectoryMakerFactory
-from exc2string2                import exc2string2
-
-def installBaseDirectories(paths, logger):
-    try:
-        factory = BaseDirectoryMakerFactory(logger)
-        directoryMaker = factory.create(paths)
-        directoryMaker.makeDirs()
-    except Exception, e:        
-        m = 'Fatal error while making directories: %s\n' % str(e)
-        m += exc2string2()
-        logger.critical(m)
-
diff --git a/Tools/RunTimeTester/src/installDB.py b/Tools/RunTimeTester/src/installDB.py
deleted file mode 100644
index 9cc34bfc62f18a0512403963675f22d435208f3c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/installDB.py
+++ /dev/null
@@ -1,10 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, anydbm
-def installDB(dbName, logger):
-    if os.path.exists(dbName): return
-    
-    cfile = anydbm.open(dbName, 'c')
-    cfile.close()
-    statusText = 'moni db created'
-    logger.info(statusText)
diff --git a/Tools/RunTimeTester/src/installStartTimeStamp.py b/Tools/RunTimeTester/src/installStartTimeStamp.py
deleted file mode 100644
index 45b0bc15595461c157ac81b4440fa150f4b6df4c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/installStartTimeStamp.py
+++ /dev/null
@@ -1,7 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-def installStartTimeStamp(paths, startTime2, logger):
-    os.system('touch %s' % os.path.join(paths.resultsDirs['packages'], 'rtt_run_start_%s' % startTime2))
-    logger.debug('Created start time stamp file in %s' % paths.resultsDirs['packages'])
-
diff --git a/Tools/RunTimeTester/src/installWebFiles.py b/Tools/RunTimeTester/src/installWebFiles.py
deleted file mode 100644
index 8c41885618cc4322a4e8ab6ae52c7c796e70edd2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/installWebFiles.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-import os.path
-import shutil
-
-def installWebFiles(RTTLibDir, resultsBasePath):
-    webDir = os.path.join(RTTLibDir, 'web')
-    necessaryFiles = [os.path.join(webDir, f) for f in os.listdir(webDir) if os.path.isfile(os.path.join(webDir, f))]
-
-    for aFile in necessaryFiles:
-        if not os.path.exists(os.path.join(resultsBasePath, os.path.basename(aFile))):
-            # self.mainLogger.debug('Installing missing web file: %s' % os.path.basename(aFile))
-            src = aFile
-            dst = resultsBasePath
-            shutil.copy(src, dst)
-                
diff --git a/Tools/RunTimeTester/src/isAthenaJob.py.obsolete b/Tools/RunTimeTester/src/isAthenaJob.py.obsolete
deleted file mode 100755
index abcc4d4b82548cb6edd1df4ffa31a38a069de983..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/isAthenaJob.py.obsolete
+++ /dev/null
@@ -1,10 +0,0 @@
-# functions to determine if a job is an Athena job
-from getInheritanceHeirarchy import getInheritanceHeirarchy
-
-def isAthenaDescriptor(desc):
-    baseClasses = getInheritanceHeirarchy(desc)
-    if 'AthenaJobDescriptor' in baseClasses: return True
-    return False
-
-def isAthenaMinder(minder):
-    return isAthenaDescriptor(minder.jDescriptor)
diff --git a/Tools/RunTimeTester/src/jobGroupKitFactory.py.OBSOLETE b/Tools/RunTimeTester/src/jobGroupKitFactory.py.OBSOLETE
deleted file mode 100755
index 99a5e02fafcd8d15f94861db84960096238769b5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/jobGroupKitFactory.py.OBSOLETE
+++ /dev/null
@@ -1,36 +0,0 @@
-from  JobGroupKits import * 
-import sys
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger()
-# -------------------------------------------------------------------------
-
-def  jobGroupKitFactory(group):
-    """ provide an instance of a JobGroupKit:
-    You know the name of a group - and you need the kit to give you
-    pieces relevant to that group"""
-
-    kits = {
-        "AthenaHelloWorld":      HelloWorldKit,
-        "AthenaAtlfast":         AtlfastKit,
-        "AthenaRecExCommon":     RecExCommonKit,
-        "AthenaJiveXML":         JiveXMLKit,
-        "AthenaegammaRec":       EgammaKit,
-        "AthenaTriggerRelease":  TriggerReleaseKit,
-        "AthenaTrigReleaseID":   TrigReleaseIDKit,
-        "AthenaTrigReleaseCalo": TrigReleaseCaloKit,
-        "RuleChecker":           RuleCheckerKit,
-        "QAmetrics":             QAmetricsKit,
-        "AthenaegammaRecWatcher":   EgammaWatcherKit,
-        "AthenaAtlfastWatcher":  AtlfastWatcherKit,
-        "PBSTest":               PBSTestKit,
-        "KitValidation":         KitValidationKit
-        "Monitor":               MonitorKit
-        }
-
-    if group in kits.keys():
-        return kits[group]()
-    else:
-        logger.error('Unknown group, could not instantiate a JobGroupKit: '+group)
-        sys.exit(0)
diff --git a/Tools/RunTimeTester/src/kitFromConfFile.py b/Tools/RunTimeTester/src/kitFromConfFile.py
deleted file mode 100755
index e5b7468a22a21d23d8101732458fc82c0c51d83d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/kitFromConfFile.py
+++ /dev/null
@@ -1,58 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-#----- Python modules -----
-
-import os.path, sys, xml.dom.minidom
-
-#----- RTT modules -----
-
-from exc2string       import exc2string
-from JobGroupsParser  import makeKits
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-from MethodTracer        import MetaTracer
-# uncomment to monitor all method calls (shows args + return values)
-# __metaclass__ = MetaTracer
-
-
-def kitFromConfFile(packages):
-    """
-    returns all kits from constructed given a dictionary of
-    (jobgroup: pathToConfFile) from any legal config file format.
-    
-    a router method that calls kit returning methods depending
-    on whether the conf file  is RTT or Unified style.
-    The paths is only needed by the RTT style - soon
-    to be supressed!
-    """
-    kits = []
-    
-    # loop over all CVSPackage objects
-    for package in packages:
-        try:
-            if package.noError():
-                newKits = makeUnifiedConfigKits(package.packageXMLfile)
-                kits.extend(newKits)
-        except Exception, e:
-            m  = "Unable to make job group kits for package %s\n" % str(package.name)
-            m += "Traceback: %s\n%s" % (str(e),
-                                        exc2string(sys.exc_info()))
-            logger.error(m)
-            package.setErrorMessage(str(e))                
-    return kits
-    
-#-------------------------------------------------------------------------
-
-def makeUnifiedConfigKits(packageXMLfile):
-    logger.debug('Making job group kits from package XML file: %s' % packageXMLfile.path)
-
-    groups = packageXMLfile['rtt/jobGroups']
-    if groups: 
-        return makeKits(groups[0]) # should be only one at most in list, pass it in
-    return {}
-
-if __name__ == '__main__':
-    pass
diff --git a/Tools/RunTimeTester/src/listFiles.py.obsolete b/Tools/RunTimeTester/src/listFiles.py.obsolete
deleted file mode 100755
index fd8922d8a7f46185fda58f46b09339451e26cf15..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/listFiles.py.obsolete
+++ /dev/null
@@ -1,34 +0,0 @@
-#Python cookbook 4.18
-
-import os.path, fnmatch
-
-def listFiles(root, patterns='*', recurse=True, return_folders=False):
-    pattern_list = patterns.split(';')
-
-    class Bunch:
-        def __init__(self, **kwds): self.__dict__.update(kwds)
-
-    arg = Bunch(recurse=recurse, pattern_list=pattern_list, return_folders=return_folders, results = [])
-
-    def visit(arg, dirname, files):
-
-        for name in files:
-            
-            fullname= os.path.normpath(os.path.join(dirname, name))
-            #print fullname
-            if arg.return_folders or os.path.isfile(fullname):
-                for pattern in arg.pattern_list:
-                    if fnmatch.fnmatch(name, pattern):
-                        arg.results.append(fullname)
-                        print 'listDir: found ', fullname
-                        break
-                    
-        if not arg.recurse: files[:] = []
-
-    #print root
-    #print pattern_list
-    os.path.walk(root, visit, arg)
-
-    print 'listFiles complete'
-    print str(arg.results)
-    return arg.results
diff --git a/Tools/RunTimeTester/src/listTransfer.py.obsolete b/Tools/RunTimeTester/src/listTransfer.py.obsolete
deleted file mode 100644
index 0f8e11db6d5c70e706df2246e189948519bed1ed..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/listTransfer.py.obsolete
+++ /dev/null
@@ -1,6 +0,0 @@
-def listTransfer(dest, kit, label):
-    
-    kitlist = kit.getItem(label) # list might be None
-    orilist = getattr(dest, label, [])
-    if kitlist: orilist.extend(kitlist)
-    setattr(dest, label, orilist)
diff --git a/Tools/RunTimeTester/src/logAnalyser.py.obsolete b/Tools/RunTimeTester/src/logAnalyser.py.obsolete
deleted file mode 100755
index 17bd908fc9063bb22f0f246419c36997a014e25c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/logAnalyser.py.obsolete
+++ /dev/null
@@ -1,89 +0,0 @@
-#! /usr/bin/python
-import os, sys
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-class LogAnalyser:
-    """
-    reads in a list of strings, and files for the presence
-    of these strings. Write counts to an html file
-    """
-    def __init__(self):
-        
-        
-        self.strings={
-            'INFO Application Manager Finalized succes'                     :0,
-            'FATAL Application configuration failed'                        :0,
-            'WARNING Failed to load modules: GaudiAlg'                      :0,
-            'libGaudiAlg.so: cannot open shared object file: No such file'  :0,
-            'RecExCommon/CBNT_jobOptions.txt missing'                       :0,
-            'RecExCommon/RecExCommon_jobOptions.txt missing'                :0,
-            'FATAL EventSelector not found'                                 :0,
-            'JobOptionsSvc       ERROR'                                     :0,
-            'xKalmanppAthena/xKalman_jobOptions.txt missing'                :0
-        }
-
-        print "seeking for the following strings:"
-        for string in self.strings.keys():
-            print string
-
-        self.nLogFiles = 0
-            
-
-    def findStrings(self, dir=os.curdir):
-        for file in os.listdir(dir):
-            if os.path.isfile(file):
-                if file.find('.log')==-1: continue
-                self.nLogFiles = self.nLogFiles+1
-                for line in open(file, 'r').readlines():
-                    for string in self.strings:
-                        if not line.find(string)==-1:
-                            self.strings[string]=self.strings[string]+1
-
-                        
-    def printPage(self):             
-        html1 = """ 
-        <!DOCTYPE html PUBLIC
-        "-//W3C//DTD XHTML 1.0 Strict//EN"
-        "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">\n
-        <html>
-        <head>
-        </head>
-        <body>
-        <center>
-        <h2>String Counts - """+os.getcwd()+"""</h2>
-        </center>
-        <p>
-        <center><table BORDER=\"1\" CELLSPACING=\"1\" CELLPADDING=\"4\">\n
-        Number of log files processed: """+str(self.nLogFiles)+"""\n
-        <tr><td>String</td><td>Occurences</td></tr>
-        """
-        
-        html2=""
-        for key in self.strings.keys():
-            line = "<tr><td>"+key+"</td><td>"
-            line = line+str(self.strings[key])+"</td></tr>\n"
-            html2=html2+line
-            
-        html3="""
-        </table>
-        </p>
-        </body>
-        </html
-        """
-        ofile=open('counts.html','w', 0777)
-        ofile.write(html1)
-        ofile.write(html2)
-        ofile.write(html3)
-        ofile.close()
-
-    
-if __name__ == "__main__":
-    stringsFileName = "./logStrings.txt"
-    la = LogAnalyser()
-    la.findStrings()
-    la.printPage()
-    
diff --git a/Tools/RunTimeTester/src/makeLocalLog.py b/Tools/RunTimeTester/src/makeLocalLog.py
deleted file mode 100644
index 53365aba66c308c7b844fc58c98dac03af52d249..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/makeLocalLog.py
+++ /dev/null
@@ -1,46 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from LoggerSetup     import createFormatter, createLogFile, createFileHandler
-import logging
-
-def makeLocalLog(logFileDir, errorlog, warninglog, infolog, debuglog, name):
-
-    formatter = createFormatter()
-    
-    errorLogName   = createLogFile(logFileDir,
-                                   errorlog,
-                                   errorlog+'\n') 
-    
-    warningLogName = createLogFile(logFileDir,
-                                   warninglog,
-                                   warninglog+'\n') 
-    
-    infoLogName    = createLogFile(logFileDir,
-                                   infolog,
-                                   infolog+'\n')
-    
-    debugLogName   = createLogFile(logFileDir,
-                                   debuglog,
-                                   debuglog+'\n')
-    
-
-    handlers = []
-    handlers.append(createFileHandler(debugLogName,
-                                      logging.DEBUG,
-                                      formatter))
-    handlers.append(createFileHandler(errorLogName,
-                                      logging.ERROR,
-                                      formatter))
-    handlers.append(createFileHandler(warningLogName,
-                                      logging.WARNING,
-                                      formatter))
-    handlers.append(createFileHandler(infoLogName,
-                                      logging.INFO,
-                                      formatter))
-    
-    logger = logging.getLogger(name)
-    [logger.removeHandler(h) for h in logger.handlers] 
-    [logger.addHandler(h) for h in handlers] 
-    
-    logger.setLevel(logging.DEBUG)
-    return logger
diff --git a/Tools/RunTimeTester/src/makeRTTDOM.py b/Tools/RunTimeTester/src/makeRTTDOM.py
deleted file mode 100755
index 78ad2b2468ef6f1e6fcc3ad4e026943f7f0c34c3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/makeRTTDOM.py
+++ /dev/null
@@ -1,11 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from  xml.dom.minidom import getDOMImplementation
-
-def makeRTTDOM(docname):        
-        impl            = getDOMImplementation()
-        docType         = impl.createDocumentType(docname, '', '')
-        namespace       = None
-        document        = impl.createDocument(namespace, docname, docType)
-        
-        return document
diff --git a/Tools/RunTimeTester/src/makeWatcherJobs.py.hold b/Tools/RunTimeTester/src/makeWatcherJobs.py.hold
deleted file mode 100755
index 53a058f9cc0dcb7b51da9822c1fd5f87e87d4970..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/makeWatcherJobs.py.hold
+++ /dev/null
@@ -1,53 +0,0 @@
-from Paths         import Paths
-#from JobsConfigReader import JobsConfigReader
-from JobsXMLReader import jobs
-from sys import exit
-
-def makeJobs(paths):
-    """
-    Function that creates a list of jobdescriptors.
-    """
-
-    print 'in makeJobs****'
-    # confReader = JobsConfigReader(paths)
-    descList = jobs(paths)
-    for desc in descList:
-        try:
-            createRunDir(desc)
-        except Exception e:
-            #later on this should be reported to a Global Reporter	
-            # for now simply remove this job from the list
-            errmsg = exc2string(sys.exc_info())
-            dbgF(self, errmsg)
-            badDesc
-
-        jobMinders=[]
-        commander = commanderFactory( self.paths.mode )
-        for jd in descList:
-            submitCmd = commander.submitCommand(jd)
-            dbgF(self, "submit command: "+submitCmd)
-            jm=jobMinderFactory(paths.mode, jd, submitCmd)
-            
-            jobMinders.append(jm)               
-
-        return jobMinders
-    
-    # confederatedDescs = [desc in descList if desc.confederation]
-    # dict = {}
-    # for desc in confederatedDescs:
-    #     dict.setdefault(desc.confederation, []).append(desc)
-
-    # for key in dict.keys():
-    #     descs = :dict[key]
-    #     fGroup = descs[0].jobGroup
-    #    badGroups = [desc.jobGroup for desc in descs if desc.jobGroup != fGroup]
-        
-    #     if badGroups:
-    #         dbgFf(__name__,'Job federation has mixed groups, exiting '+str(badGroups))
-    #         exit(0)
-
-    #     fGroup = fGroup+'Watcher'
-    #     descriptor = descriptorFactory(fGroup)
-    #     minder = minderFactory(paths.mode, descriptor, '')
-    #     minder.jobsToWatch = 
-        
diff --git a/Tools/RunTimeTester/src/memstat.py b/Tools/RunTimeTester/src/memstat.py
deleted file mode 100755
index 7161aeb3b90ec1c28f7d1159c6fdd337dc83cb4e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/memstat.py
+++ /dev/null
@@ -1,24 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-# a function to obtain process memeory usage
-
-def memstat():
-    memfile = open('/proc/self/status', 'r')
-    lines = memfile.readlines()
-    memfile.close()
-
-    pid = 0
-    for line in lines:
-        if line.startswith('Pid'): pid = int(line.split()[1])
-        #break
-    
-    lines = [l for l in lines if l.startswith('Vm')]
-    msg = 'Memory usage for process %d:\n' % pid
-    for l in lines:
-        msg += l
-
-    # memfile = open('/proc/meminfo', 'r')
-    # msg += memfile.read()
-    # memfile.close()
-    
-    return msg
diff --git a/Tools/RunTimeTester/src/moduleImportFinder.py.obsolete b/Tools/RunTimeTester/src/moduleImportFinder.py.obsolete
deleted file mode 100755
index 142a96ef7ace682240f8491c86a42f36298e2ddb..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/moduleImportFinder.py.obsolete
+++ /dev/null
@@ -1,107 +0,0 @@
-import os
-from formatCollection import formatCollection
-
-class Imports:
-    def __init__(self, maxRecursion=25):
-        files = os.listdir('.')
-        files = [f for f in files if f.endswith('.py')]
-        # print files
-        self.dict = {}
-        for file in files:
-            self.dict[file] = self.findFrom(file)
-
-        self.indent = ''
-        self.istep = '   '
-        self.recurs = 0
-        self.maxrecurs = maxRecursion
-        # print self.dict
-    def findFrom(self,file):
-        # print 'findFrom: handling file %s' % file
-        lines = open(file, 'r').readlines()
-        lines = [l.strip() for l in lines]
-        def fromSelector(l):
-            if len(l.split())<2: return False
-            if l.split()[0] != 'from': return False
-            if l.find('import')<0: return False
-            return True
-        
-        def importSelector(l):
-            if len(l.split())<2: return False
-            if not l.startswith('import'): return False
-            return True
-        
-
-        linesFrom = [l for l in lines if l.startswith('from')]
-        linesFrom = [l for l in linesFrom if fromSelector(l)]
-        froms     = [l.split()[1] for l in linesFrom]
-
-        linesImport = [l for l in lines if l.startswith('import')]
-        linesImport = [l for l in linesImport if importSelector(l)]
-        linesImport = [l[len('import '):] for l in linesImport]
-
-        imports = []
-        [imports.extend(l.split(',')) for l in linesImport]
-        imports = [imp.strip() for imp in imports]
-        imports = [imp for imp in imports if os.path.exists(os.path.join(
-            os.getcwd(), imp+'.py'))]
-        
-        # print formatCollection(lines)
-        froms.extend(imports)
-        return froms
-
-    def whoImports(self, file):
-        if file.endswith('.py'):
-            file = file[:-3]
-        imports = [a for a in self.dict.keys() if file in self.dict[a]]
-        # print '\n\nthe following imports %s:\n' % file
-        return  imports
-
-    def ifunc(self,p):
-        self.recurs += 1
-        # print self.recurs
-        if self.recurs>self.maxrecurs:
-            return
-        self.indent += self.istep
-        print self.recurs, self.indent,p
-        p2 = self.whoImports(p)
-        # print p2
-        if not p2: print self.indent,'===end'
-        for pp in p2:
-                self.ifunc(pp)
-        self.indent = self.indent[:-len(self.istep)]
-        self.recurs -= 1
-
-    def chain(self, file):
-        print '\nChain for import of %s\n' % file
-        pp = self.whoImports(file)
-        for p in pp:
-            print ' '
-            self.ifunc(p)
-            self.recurs = 0
-
-    def chain2(self, file):
-        '\nChain for import of %s' % file
-        p1  = self.whoImports(file)
-        self.indent = ''
-        self.istep = '    '
-        for p in p1:
-            print self.indent,p
-            p2  = self.whoImports(p)
-            for pp in p2:
-                self.indent += self.istep
-                print self.indent, pp
-                p3 = self.whoImports(pp)
-                for ppp in  p3:
-                    self.indent += self.istep
-                    print self.indent, ppp
-                    self.indent = self.indent[:-len(self.istep)]
-                self.indent = self.indent[:-len(self.istep)]
-                    
-            
-        
-if __name__ == '__main__':
-    importee = 'DirMaker'
-    maxRecursion = 1
-    imps = Imports(maxRecursion)
-    imps.chain(importee)
-
diff --git a/Tools/RunTimeTester/src/nicosDates.py.obsolete b/Tools/RunTimeTester/src/nicosDates.py.obsolete
deleted file mode 100644
index e0fde9403bc20d9e2a625f3dc62f6d642b2462a9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/nicosDates.py.obsolete
+++ /dev/null
@@ -1,13 +0,0 @@
-from NICOS import NICOS
-
-def nicosDates(paths):
-    nicosDate          = 'n/a'
-    nicosDateFormatted = 'n/a'
-    
-    nicos = NICOS(paths)
-    nDate = nicos.getReleaseDate()
-    if nDate.isValid():
-        nicosDate=str(nicos.getReleaseDate().absTime())
-        nicosDateFormatted=str(nicos.getReleaseDate())
-
-    return (nicosDate, nicosDateFormatted)
diff --git a/Tools/RunTimeTester/src/outputXSL.py.obsolete b/Tools/RunTimeTester/src/outputXSL.py.obsolete
deleted file mode 100755
index 53faced75c415f11041edc36eef829ed21ff247a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/outputXSL.py.obsolete
+++ /dev/null
@@ -1,33 +0,0 @@
-import os
-import os.path
-import page0DefaultFiles
-
-filesToWrite = {
-    'currentStatusN0X.js':       page0DefaultFiles.currentStatusN0X,
-    'currentStatusNX0.js':       page0DefaultFiles.currentStatusNX0,
-    'messageOfTheDay.js':        page0DefaultFiles.messageOfTheDay,
-    'page0.html':                page0DefaultFiles.page0,
-    'page0.noTopFrame.html':     page0DefaultFiles.page0_noTopFrame,
-    'status.html':               page0DefaultFiles.status,
-    'page0.xsl':                 page0DefaultFiles.page0XSL,
-    'page1.xsl':                 page0DefaultFiles.page1XSL,
-    'page2.xsl':                 page0DefaultFiles.page2XSL,
-    'page3.xsl':                 page0DefaultFiles.page3XSL,
-    'page4.xsl':                 page0DefaultFiles.page4XSL,
-    'RTTpage1_LinksAtTopOfPage.js': page0DefaultFiles.RTTpage1_LinksAtTopOfPage
-    }
-
-def writeOut((fn, text), resBase):
-    fullPath = os.path.join(resBase, fn)
-    if os.path.exists(fullPath): return
-    of = open(fullPath,'w')
-    of.write(text)
-    of.close()
-    
-
-
-
-if __name__ == '__main__':
-    resBaseDir = '/afs/cern.ch/atlas/project/RTT/data/brinick/Results/'
-    # resBaseDir = '/afs/cern.ch/atlas/project/RTT/Results/'
-    [writeOut(item, resBaseDir) for item in filesToWrite.items()]
diff --git a/Tools/RunTimeTester/src/page0DefaultFiles.py.obsolete b/Tools/RunTimeTester/src/page0DefaultFiles.py.obsolete
deleted file mode 100755
index ac3db9caa5fb2d9b26afb682115985aac5400fac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/page0DefaultFiles.py.obsolete
+++ /dev/null
@@ -1,974 +0,0 @@
-currentStatusN0X = """
-/************************************************/
-/*                                              */
-/*              EXPERT CORNER                   */
-/*                                              */
-/*                                              */
-/*  (If you don't want to change the way that)  */
-/*  (messages are displayed, rather you just )  */
-/*  (want to edit messages, then skip down to)  */
-/*  (the USER CORNER                         )  */
-/*                                              */
-/************************************************/
-
-var runs = new Array();
-
-/******************************************/
-/* A simple Javascript class for messages */
-/******************************************/
-
-function Run(date,rel,branch,comment){
-   this.theDate     = date;
-   this.theRel      = rel;
-   this.theBranch   = branch;
-   this.theComment  = comment;
-   runs.push(this);
-}
-
-/***************************************************/
-/* Functions to do with displaying on the web page */
-/***************************************************/
-
-function titleStyle(){
-   return 'font-family:sans-serif;font-weight:bold;font-size:10pt;color:#88f;background-color:#ccf';
-}
-
-function textStyle(){
-   return 'font-family:sans-serif;font-size:8pt';
-}
-
-/* HTML code for displaying a message*/
-function display(numbRunsToDisplay,title){
-   htmlCode  = '<div align="center">';
-   htmlCode += '<table width="100%" border="0" cellspacing="0" cellpadding="0" bgcolor="#dddddd">';
-   htmlCode += '<tr><td colspan= "4" align="center" style="' + titleStyle() + '">' + title + '</td></tr>'
-   htmlCode += '<tr style="font-weight:bold"><td width="20%">NICOS Date</td><td width="10%">Release</td><td width="10%">Branch</td><td>Comment</td></tr>';
-
-   for (var i=runs.length-1;i >= (runs.length - numbRunsToDisplay);i--){
-      run = runs[i];
-      htmlCode += '<tr style="' + textStyle() + '">';
-      htmlCode += '<td>' + run.theDate + '</td>';
-      htmlCode += '<td>' + run.theRel + '</td>';
-      htmlCode += '<td>' + run.theBranch + '</td>';
-      htmlCode += '<td>' + run.theComment + '</td>';   
-      htmlCode += '</tr>';
-   }
-
-   htmlCode += '</table>';
-   htmlCode += '</div>';
-
-   document.write(htmlCode);
-}
-
-
-/************************************************/
-/*                                              */
-/*                USER CORNER                   */
-/*                                              */
-/*                                              */
-/************************************************/
-
-new Run('---','rel_XXX','N.0.X','This JavaScript file located in %s');
-
-/***********************************************************/
-/*        ADD A NEW RUN ABOVE THIS BOX                     */
-/***********************************************************/
-
-/* Show the last how many runs ? */
-display(numbOfRunsToDisplay = 7, title = ":: Current status of N0X runs ::");
-"""
-
-#########################################################################
-#########################################################################
-
-currentStatusNX0  = """
-/************************************************/
-/*                                              */
-/*              EXPERT CORNER                   */
-/*                                              */
-/*                                              */
-/*  (If you don't want to change the way that)  */
-/*  (messages are displayed, rather you just )  */
-/*  (want to edit messages, then skip down to)  */
-/*  (the USER CORNER                         )  */
-/*                                              */
-/************************************************/
-
-var runs = new Array();
-
-/******************************************/
-/* A simple Javascript class for messages */
-/******************************************/
-
-function Run(date,rel,branch,comment){
-   this.theDate     = date;
-   this.theRel      = rel;
-   this.theBranch   = branch;
-   this.theComment  = comment;
-   runs.push(this);
-}
-
-/***************************************************/
-/* Functions to do with displaying on the web page */
-/***************************************************/
-
-function titleStyle(){
-   return 'font-family:sans-serif;font-weight:bold;font-size:10pt;color:#88f;background-color:#ccf';
-}
-
-function textStyle(){
-   return 'font-family:sans-serif;font-size:8pt';
-}
-
-/* HTML code for displaying a message*/
-function display(numbRunsToDisplay,title){
-   htmlCode  = '<div align="center">';
-   htmlCode += '<table width="100%" border="0" cellspacing="0" cellpadding="0" bgcolor="#dddddd">';
-   htmlCode += '<tr><td colspan= "4" align="center" style="' + titleStyle() + '">' + title + '</td></tr>'
-   htmlCode += '<tr style="font-weight:bold"><td width="20%">NICOS Date</td><td width="10%">Release</td><td width="10%">Branch</td><td>Comment</td></tr>';
-
-   for (var i=runs.length-1;i >= (runs.length - numbRunsToDisplay);i--){
-      run = runs[i];
-      htmlCode += '<tr style="' + textStyle() + '">';
-      htmlCode += '<td>' + run.theDate + '</td>';
-      htmlCode += '<td>' + run.theRel + '</td>';
-      htmlCode += '<td>' + run.theBranch + '</td>';
-      htmlCode += '<td>' + run.theComment + '</td>';   
-      htmlCode += '</tr>';
-   }
-
-   htmlCode += '</table>';
-   htmlCode += '</div>';
-
-   document.write(htmlCode);
-}
-
-
-/************************************************/
-/*                                              */
-/*                USER CORNER                   */
-/*                                              */
-/*                                              */
-/************************************************/
-
-/* This is all the messages we've ever posted.    */
-/* Add your new one down below after all this lot.*/
-
-new Run('---','rel_XXX','N.X.0','This JavaScript file located in %s');
-
-/***********************************************************/
-/*        ADD A NEW RUN ABOVE THIS BOX                     */
-/***********************************************************/
-
-/* Show the last how many runs ? */
-display(numbOfRunsToDisplay = 7, title = ":: Current status of NX0 runs ::");
-
-"""
-
-#########################################################################
-#########################################################################
-
-messageOfTheDay  = """
-/************************************************/
-/*                                              */
-/*              EXPERT CORNER                   */
-/*                                              */
-/*                                              */
-/*  (If you don't want to change the way that)  */
-/*  (messages are displayed, rather you just )  */
-/*  (want to edit messages, then skip down to)  */
-/*  (the USER CORNER                         )  */
-/*                                              */
-/************************************************/
-
-var messages = new Array();
-
-/******************************************/
-/* A simple Javascript class for messages */
-/******************************************/
-
-function Message(date,text,isImportant){
-   this.theDate     = date;
-   this.theText     = text;
-   this.isImportant = isImportant;
-   messages.push(this);
-}
-
-/***************************************************/
-/* Functions to do with displaying on the web page */
-/***************************************************/
-
-/* Display style for the message text */
-function textStyle(isImportant){
-   style = 'font-family:arial;font-size:10pt;';
-   if (isImportant){
-       style = style + 'font-weight:bold;color:red';
-   }
-   else{
-       style = style + 'color:black';
-   }
-
-   return style;
-}
-
-/* Display style for the date*/
-function dateStyle(){
-   return 'font-family:arial;font-size:10pt;color:black;font-weight:bold';
-}
-
-function titleStyle(){
-   return 'font-family:sans-serif;font-weight:bold;font-size:10pt;color:#88f;background-color:#ccf';
-}
-
-/* HTML code for displaying a message*/
-function display(numbMessagesToDisplay,title){
-   htmlCode  = '<div align="center">';
-   htmlCode += '<table width="100%" border="0" cellspacing="0" cellpadding="0" bgcolor="#dddddd">';
-   htmlCode += '<tr><td colspan= "4" align="center" style="' + titleStyle() + '">' + title + '</td></tr>'
-
-   for (var i=messages.length-1;i >= (messages.length - numbMessagesToDisplay);i--){
-      message = messages[i];
-      htmlCode += '<tr style="' + dateStyle() + '">';
-      htmlCode += '<td>' + message.theDate + '</td>';
-      htmlCode += '</tr>';
-      htmlCode += '<tr style="' + textStyle(message.isImportant) + '">';
-      htmlCode += '<td>' + message.theText + '</td>';
-      htmlCode += '</tr>';
-   }
-
-   htmlCode += '</table>';
-   htmlCode += '</div>';
-
-   document.write(htmlCode);
-
-}
-
-/************************************************/
-/*                                              */
-/*                USER CORNER                   */
-/*                                              */
-/*                                              */
-/************************************************/
-
-new Message('---','The JavaScript file that created this is located in %s',true);
-
-
-/***********************************************************/
-/*        ADDING A NEW MESSAGE                             */
-/***********************************************************/
-/* Add your latest message above this comment box          */
-/* Skeleton code is:                                       */
-/*   new Message('theDate','theMessage',false);            */
-/*                                                         */
-/*  where false means the message gets displayed normal    */
-/*        true means message displayed in red and bold     */
-/***********************************************************/
-
-/* Show the last how many messages ? */
-display(numberOfMessagesToDisplay = 1,title=":: Message of the Day ::");
-"""
-
-#########################################################################
-#########################################################################
-
-page0 = """
-<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Frameset//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-frameset.dtd">
-<html xmlns="http://www.w3.org/1999/xhtml" lang="en" xml:lang="en">
-  <head>
-    <meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
-    <title>Home | The ATLAS RunTimeTester Results</title>
-  </head>
-
-  <frameset rows="25%,*">
-    <frame src="status.html" name="top_frame"/>    
-    <frame src="page1.xml" name="main_frame"/>
-  </frameset>
-
-</html>
-"""
-
-#########################################################################
-#########################################################################
-
-page0_noTopFrame = """
-<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Frameset//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-frameset.dtd">
-<html xmlns="http://www.w3.org/1999/xhtml" lang="en" xml:lang="en">
-  <head>
-    <meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
-    <title>Home | The ATLAS RunTimeTester Results</title>
-  </head>
-
-  <frameset cols="15%,*">
-      <frame src="page0.xml" />    
-      <frame src="" name="main_frame"/>
-  </frameset>
-</html>
-"""
-                                        
-#########################################################################
-#########################################################################
-
-RTTpage1_LinksAtTopOfPage = """
-/*********************/
-/***** Base URLs *****/
-/*********************/
-
-uclHEP    = 'http://www.hep.ucl.ac.uk/atlas/AtlasTesting/';
-cernRTT   = 'http://cern.ch/atlasrtt/';
-atlasComp = 'http://atlas-computing.web.cern.ch/atlas-computing/';
-
-/*********************/
-/***** Style CSS *****/
-/*********************/
-
-styleClass  = '<style>';
-styleClass += 'a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}';
-styleClass += 'a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}';
-styleClass += '</style>';
-
-/*********************/
-/***** The links *****/
-/*********************/
-
-link00 = '<a target="main_frame" class="pageLink" href="page1.xml">All RTT results</a>';
-link0 = '<a target="main_frame" class="pageLink" href="weeklyTable/weekly.php">Summary of the week</a>';
-link1 = '<a target="_top" class="pageLink" href="https://savannah.cern.ch/bugs/?group=atlas-infra">Report RTT bugs</a>';
-link2 = '<a target="_top" class="pageLink" href="' + uclHEP + '">RTT documentation</a>';
-link3 = '<a target="_top" class="pageLink" href="' + atlasComp + 'projects/releases/releases.php">Release Status</a>';
-link4 = '<a target="_top" class="pageLink" href="' + atlasComp +'links/distDirectory/nightlies/global/">NICOS</a>';
-link5 = '<a target="main_frame" class="pageLink" href="' + cernRTT + 'DatasetManager.report">Dataset Manager Report</a>';
-
-document.write(
-   styleClass
- + '<div align="center">'
- + link00 + ' | '
- + link0 + ' | '
- + link1 + ' | '
- + link2 + ' | '
- + link3 + ' | '
- + link4 + ' | '
- + link5
- + '</div><br />' 
-);
-"""
-
-#########################################################################
-#########################################################################
-
-status = """
-<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">
-<html>
-<head>
-<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
-<style type="text/css">body{background-color:#eeeeff}</style>
-<title></title>
-</head>
-
-<body>
-<script src="RTTpage1_LinksAtTopOfPage.js" type="text/JavaScript"></script>
-<table width="100%">
-<tr><td colspan="2"><script src="messageOfTheDay.js" type="text/JavaScript"></script></td></tr>
-</table>
-</body>
-</html>
-"""
-
-#########################################################################
-#########################################################################
-
-page0XSL = \
-"""<?xml version="1.0" encoding="UTF-8"?>
-<xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
-<xsl:template match="RTTpage0">
-  <html>
-  <style>
-   div.header{font-family:arial;font-size:10pt;font-weight:bold;background-color:#ccccff;color:#8888ff}
-   div.text{font-family:arial;font-size:10pt;color:black}
-   a:link{font-family:arial;font-size:10pt;color:#333333;text-decoration:underline}
-   a:visited{color:gray;font-family:arial;font-size:10pt;}
-   a:hover{color:orange;text-decoration:none}
-   .button{font-family:sans-serif;font-weight:bold;font-size:10pt;background-color:#ccccff;border:1px #aaaaff solid;cursor:pointer;}
-  </style>
-  <body bgcolor="#eeeeff">
-      <div class="header" align="center">:: Top Frame ::</div>
-      <p align="center">
-          <button class="button" onclick="top.location.href='page0.noTopFrame.html'">Hide</button>
-          <button class="button" onclick="top.location.href='page0.html'">Show</button>
-      </p>         
-      <div class="header" align="center">:: Available Results ::</div>
-      <p>
-      <xsl:choose>
-         <xsl:when test="count(page1)=0">
-             <div class="text">No results available.</div>
-         </xsl:when>
-         <xsl:otherwise>
-            <xsl:for-each select="page1">
-               <xsl:sort select="normalize-space(page1Name)" />
-               <xsl:variable name="page1URL"><xsl:value-of select="normalize-space(relPage0Page1Path)" /></xsl:variable>
-               <a href="{$page1URL}" target="main_frame"><xsl:value-of select="page1Name" /></a><br />
-            </xsl:for-each>
-         </xsl:otherwise>
-       </xsl:choose>
-      </p>
-      <div class="header" align="center">:: Old Monolithic Results ::</div>
-      <a href="http://www.hep.ucl.ac.uk/atlas/RTTmonolithic/RTTpage1.html" target="main_frame">Prior 11.2.0</a>          
-  </body>  
-</html>
-</xsl:template>
-</xsl:stylesheet>
-"""
-
-#########################################################################
-#########################################################################
-
-page1XSL = \
-"""<?xml version="1.0" encoding="UTF-8"?>
-<xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
-<xsl:template match="RTTpage1">
-  <html>
-  <style>
-  .level1TableHeader{font-family:arial;font-size:16pt;font-weight:bold;color:black;background-color:#8888ff}
-  .level2TableHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;background-color:#aaaaff}
-  .tableEntry{font-family:arial;font-size:14pt;color:#333333;background-color:#bbbbff}
-   a.tableLink{font-family:arial;font-size:14pt;color:#333333;text-decoration:none;border-bottom:1px solid}
-   a.tableLink:hover{font-family:arial;font-size:14pt;color:orange;text-decoration:none;border-bottom:1px solid}
-   a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-   a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-   a.permaLink{font-family:arial;font-size:7pt;color:#777777;text-decoration:underline;float:right}
-   a.permaLink:hover{color:black;text-decoration:none}
-   .sectionHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;}
-   .buildHeader{font-family:arial;font-size:14pt;font-weight:bold;background-color:#ddddff;color:black;}
-  </style>
-  <body bgcolor="#eeeeff">
-  <center>
-     <h1>All available RTT runs</h1><br />
-
-
-     <table  BORDER="1" CELLSPACING="0" CELLPADDING="4">
-        <tr class="level2TableHeader">
-           <td align="center" valign="top" colspan="1">Release</td>
-           <td align="center" valign="top" colspan="1">Branch</td>
-           <td align="center" valign="top" colspan="1">Platform</td>
-           <td align="center" valign="top" colspan="1">Project</td>
-           <td align="center" valign="top" colspan="1">Started<font color="yellow"> (GMT)</font></td>
-           <td align="center" valign="top" colspan="1">Completed<font color="yellow"> (GMT)</font></td>
-           <td align="center" valign="top" colspan="1">NICOS</td>
-           <td align="center" valign="top" colspan="1">Status</td>
-        </tr>
-
-        <xsl:for-each select="overview">
-           <xsl:sort select="normalize-space(release)" />
-           <xsl:sort select="normalize-space(originalBranch)" />
-           <xsl:sort select="normalize-space(targetCMTCONFIG)" />
-           <xsl:sort select="normalize-space(topProject)" />
-
-           <xsl:variable name="page2url"><xsl:value-of select="normalize-space(relPage1Page2Path)" /></xsl:variable>
-
-           <tr class="tableEntry">
-              <td align="center" valign="top" colspan="1"><a class="tableLink" href="{$page2url}"><xsl:value-of select="normalize-space(release)" /></a></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="normalize-space(originalBranch)" /></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="targetCMTCONFIG" /></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="normalize-space(topProject)" /></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="startTime1" /></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="endTime1" /></td>
-              <td align="center" valign="top" colspan="1"><xsl:value-of select="nicosDate1" /></td>
-              <td align="center" valign="top" colspan="1"><font size="1"><b><xsl:value-of select="statusText" /></b></font></td>	
-           </tr>
-
-        </xsl:for-each>
-
-     </table>
-
-  </center>
-  </body>
-  </html>
-</xsl:template>
-</xsl:stylesheet>
-"""
-
-#########################################################################
-#########################################################################
-
-page2XSL = \
-"""<?xml version="1.0" encoding="UTF-8"?>
-<xsl:stylesheet version="1.1" xmlns:xsl="http://www.w3.org/1999/XSL/Transform" xmlns:rtt="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-<xsl:template match="RTTSummary">
-  <html>
-  <style>
-  .pageHeader{font-family:arial;font-size:20pt;font-weight:bold;color:black;}
-  .level1TableHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;background-color:#8888ff}
-  .level2TableHeader{font-family:arial;font-size:10pt;font-weight:bold;color:black;background-color:#aaaaff}
-  .tableEntry{font-family:arial;font-size:10pt;color:#333333;background-color:#bbbbff}
-   a.tableLink{font-family:arial;font-size:10pt;color:#333333;text-decoration:none;border-bottom:1px solid}
-   a.tableLink:hover{font-family:arial;font-size:10pt;color:orange;text-decoration:none;border-bottom:1px solid}
-   a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-   a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-   a.permaLink{font-family:arial;font-size:7pt;color:#777777;text-decoration:underline;float:right}
-   a.permaLink:hover{color:black;text-decoration:none}              
-   .sectionHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;}
-  </style>
-  <body bgcolor="#eeeeff">
-  <a href="" class="permaLink">Permalink</a>
-  <center>
-
-  <xsl:comment>******************************************</xsl:comment>
-  <xsl:comment>* CREATE THE TOP OF THE PAGE INFORMATION *</xsl:comment>
-  <xsl:comment>******************************************</xsl:comment>
-
-  <div class="pageHeader">
-        <xsl:value-of select="normalize-space(overview/releaseName)" />/<xsl:value-of select="normalize-space(overview/runType)" /> run ---
-        <xsl:value-of select="normalize-space(overview/targetCMTCONFIG)" /> --- <xsl:value-of select="normalize-space(overview/topProject)" />
-  </div>
-
-  <br />
-
-  <xsl:comment>******************************************</xsl:comment>
-  <xsl:comment>* OUTPUT THE TOOLS AND LIBTOOLS TAGS     *</xsl:comment>
-  <xsl:comment>******************************************</xsl:comment>
-
-  <div class="sectionHeader">
-       RTT src code: <xsl:value-of select="overview/rttSrcVersion" /><br />
-       RTT library tools: <xsl:value-of select="overview/rttLibVersion" /><br />
-  </div>
-
-  <br />
-
-  <xsl:comment>******************************************</xsl:comment>
-  <xsl:comment>* NUMBER OF PACKAGES FOUND BY THE RTT    *</xsl:comment>
-  <xsl:comment>******************************************</xsl:comment>
-
-  <div class="pageHeader">Packages found by the RTT: <xsl:value-of select="count(package)" /></div><br />
-
-  <xsl:comment>******************************************</xsl:comment>
-  <xsl:comment>* LINK TO THE TEXT RUN SUMMARY FILE      *</xsl:comment>
-  <xsl:comment>******************************************</xsl:comment>
-  <a class="pageLink" href="RTTSummary.txt">RTT developer run diagnostics</a><br /><br />
-
-  <xsl:comment>**************************************</xsl:comment>
-  <xsl:comment>* CREATE THE MAIN TABLE OF PACKAGES  *</xsl:comment>
-  <xsl:comment>**************************************</xsl:comment>
-
-    <table border="1" cellspacing="0" cellpadding="4">
-      <tr class="level1TableHeader" align="center" valign="top">
-         <td>Package</td><td>Status</td><td>ATN jobs</td><td>RTT jobs</td>
-      </tr>
-
-      <xsl:comment>**********************************************</xsl:comment>
-      <xsl:comment>* LOOP OVER EACH PACKAGE TAG IN THE XML FILE *</xsl:comment>
-      <xsl:comment>**********************************************</xsl:comment>
-
-      <xsl:variable name="atnParentNode" select="ATNSummary" />
-
-      <xsl:for-each select="package">
-         <xsl:sort select="normalize-space(packageName)" />
-         <xsl:variable name="page3URL"><xsl:value-of select="normalize-space(relPage2Page3Path)" /></xsl:variable>
-
-         <tr class="tableEntry" align="center" valign="top">
-
-           <td><a class="tableLink" href="{$page3URL}"><xsl:value-of select="packageName"/></a><br /><xsl:value-of select="packageTag"/></td>
-           <td><xsl:value-of select="normalize-space(packageStatus)" /></td>               
-           <td>
-               <table border="0" cellspacing="0" cellpadding="4">                   
-                   <tr class="level2TableHeader" align="center" valign="top"><td>Run by</td><td>Done Jobs</td><td>Jobs OK</td></tr>
-                        <xsl:choose>
-                           <xsl:when test="count(minder[normalize-space(jobGroup)='AthenaATN'])=0">
-                              <tr><td align="center">RTT</td><td align="center">---</td><td align="center">---</td></tr>
-                           </xsl:when>
-                           <xsl:otherwise>
-                              <tr>
-                                 <td align="center">RTT</td>
-                                 <td align="center">
-                                     <xsl:value-of select="minder[normalize-space(jobGroup)='AthenaATN']/normalize-space(chainLength)" />/<xsl:value-of select="minder[normalize-space(jobGroup)='AthenaATN']/normalize-space(chainLength)" />
-                                 </td>
-                                 <td align="center">
-                                     <xsl:value-of select="minder[normalize-space(jobGroup)='AthenaATN']/normalize-space(chainSuccesses)" />/<xsl:value-of select="minder[normalize-space(jobGroup)='AthenaATN']/normalize-space(chainLength)" />
-                                 </td>
-                              </tr>
-                           </xsl:otherwise>
-                        </xsl:choose>
-                        <xsl:variable name="currentPackageName"><xsl:value-of select="normalize-space(packageName)" /></xsl:variable>
-                        <tr>
-                            <td align="center">ATN</td>
-                            <td align="center">
-                               <xsl:choose>
-                                   <xsl:when test="count($atnParentNode/test[packageName=$currentPackageName])=0">---</xsl:when>
-                                   <xsl:otherwise><xsl:value-of select="count($atnParentNode/test[packageName=$currentPackageName])" />/<xsl:value-of select="count($atnParentNode/test[packageName=$currentPackageName])" /></xsl:otherwise>
-                               </xsl:choose>
-                            </td>
-                            <td align="center">
-                               <xsl:choose>
-                                   <xsl:when test="count($atnParentNode/test[normalize-space(packageName)=$currentPackageName])=0">---</xsl:when>
-                                   <xsl:otherwise><xsl:value-of select="count($atnParentNode/test[normalize-space(packageName)=$currentPackageName and normalize-space(testStatus)='OK'])" />/<xsl:value-of select="count($atnParentNode/test[packageName=$currentPackageName])" /></xsl:otherwise>
-                               </xsl:choose>
-                            </td>
-                        </tr>
-               </table> 
-           </td>
-           <td align="center">
-   
-               <xsl:comment>***************************************************</xsl:comment>
-               <xsl:comment>* CREATE THE TABLE OF JOB GROUPS FOR THIS PACKAGE *</xsl:comment>
-               <xsl:comment>***************************************************</xsl:comment>
-
-
-               <table border="0" cellspacing="0" cellpadding="4">
-                  <tr class="level2TableHeader" align="center" valign="top"><td>Done Jobs</td><td>Jobs OK</td><td>Tests OK</td></tr>
-                   
-                  <xsl:choose>
-                     <xsl:when test="count(minder[not(normalize-space(jobGroup)='AthenaATN')])=0">
-                        <tr>
-                           <td align="center">---</td>
-                           <td align="center">---</td>
-                           <td align="center">---</td>
-                        </tr>
-                     </xsl:when>
-                     <xsl:otherwise>
-                        <tr>
-                           <td align="center">
-                               <xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN') and done])" />/<xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN')])" />
-                           </td>
-                           <td align="center">
-                               <xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN') and done and normalize-space(status)='success'])" />/<xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN') and done])" />
-                           </td>
-                           <td align="center"><xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN') and done and normalize-space(status)='success' and normalize-space(postProcessingResult)='success'])" />/<xsl:value-of select="count(minder[not(normalize-space(jobGroup)='AthenaATN') and done and normalize-space(status)='success'])" />
-                           </td>
-                        </tr>
-                     </xsl:otherwise>
-                  </xsl:choose>
-
-               </table>
-   
-           </td>
-         </tr>
-   
-      </xsl:for-each>
-
-    </table>
-    <br />
-    <a class="pageLink" href="../../../../../page1.xml">Back to home page</a>
-  </center>
-               
-  </body>
-  </html>
-</xsl:template>
-</xsl:stylesheet>
-"""
-
-#########################################################################
-#########################################################################
-
-page3XSL = \
-"""<?xml version="1.0" encoding="UTF-8"?>
-<xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
-<xsl:template match="RTTpage3">
-  <html>
-  <head>
-
-  <style>
-  .level1TableHeader{font-family:arial;font-size:14pt;font-weight:bold;color:black;background-color:#8888ff}
-  .level2TableHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;background-color:#aaaaff}
-  .tableEntry{font-family:arial;font-size:10pt;color:#333333;background-color:#bbbbff}
-   a.permaLink{font-family:arial;font-size:7pt;color:#777777;text-decoration:underline;float:right}
-   a.permaLink:hover{color:black;text-decoration:none}              
-   a.tableLink{font-family:arial;font-size:10pt;color:#333333;text-decoration:none;border-bottom:1px solid}
-   a.tableLink:hover{font-family:arial;font-size:10pt;color:orange;text-decoration:none;border-bottom:1px solid}
-   a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-   a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-  </style>
-
-  </head>
-  <body bgcolor="#eeeeff">
-    <a href="" class="permaLink">Permalink</a>              
-    <center>
-
-       <xsl:comment>******************************************</xsl:comment>
-       <xsl:comment>* CREATE THE TOP OF THE PAGE INFORMATION *</xsl:comment>
-       <xsl:comment>******************************************</xsl:comment>
-
-       <h2>
-        <xsl:value-of select="normalize-space(overview/releaseName)" />/<xsl:value-of select="normalize-space(overview/runType)" /> run ---
-        <xsl:value-of select="normalize-space(overview/targetCMTCONFIG)" /> --- <xsl:value-of select="normalize-space(overview/topProject)" />: <xsl:value-of select="normalize-space(package/packageName)" />
-       </h2>
-       <br /><br />
-
-       <xsl:comment>******************************************</xsl:comment>
-       <xsl:comment>* CREATE SUMMARY STATS TABLE             *</xsl:comment>
-       <xsl:comment>******************************************</xsl:comment>
-
-       <table cellpadding="4" cellspacing="0" border="1">
-           <tr class="level1TableHeader">
-                <td align="center" valign="top" colspan="1">Total jobs</td>
-                <td align="center" valign="top" colspan="1">Done jobs</td>
-                <td align="center" valign="top" colspan="1">Jobs completing OK</td>
-                <td align="center" valign="top" colspan="1">Jobs with all post-run tests OK</td>
-           </tr>
-           <tr class="tableEntry">
-                <td align="center" valign="top" colspan="1"><xsl:value-of select="package/nJobsInPackage" /></td>
-                <td align="center" valign="top" colspan="1"><xsl:value-of select="package/nJobsDoneInPackage" /></td>
-                <td align="center" valign="top" colspan="1"><xsl:value-of select="package/nJobsSuccessInPackage" /></td>
-                <xsl:choose>
-                    <xsl:when test="package/nTestsInPackage=0">
-      	                <td align="center" valign="top" colspan="1">n/a</td>
-                    </xsl:when>
-                    <xsl:otherwise>                     
-      	                <td align="center" valign="top" colspan="1"><xsl:value-of select="package/nTestsSuccessInPackage" /></td>
-                    </xsl:otherwise>
-                </xsl:choose>
-           </tr>
-       </table>
-
-       <xsl:comment>******************************************</xsl:comment>
-       <xsl:comment>* CREATE MAIN TABLE                      *</xsl:comment>
-       <xsl:comment>******************************************</xsl:comment>
-
-       <br />
-
-           <table border="1" cellspacing="0" cellpadding="4">
-              <tr class="level1TableHeader" align="center" valign="top">
-                <td align="center" valign="top" colspan="1">Name</td>
-                <td align="center" valign="top" colspan="1">Group</td>
-                <td align="center" valign="top" colspan="1">Datasets</td>
-                <td align="center" valign="top" colspan="1">Job</td>
-                <td align="center" valign="top" colspan="1">Combined Checks</td>
-                <td align="center" valign="top" colspan="1">Ref. Release</td>
-              </tr>
-
-              <xsl:for-each select="package/minder">
-                 <xsl:sort select="jobGroup" />
-                 <xsl:sort select="identifiedName" />              
-                 <tr class="tableEntry" align="center" valign="top">
-                    <td align="center" valign="top" colspan="1">
-                          <xsl:variable name="page4url"><xsl:value-of select="normalize-space(relPage3Page4Location)" /></xsl:variable>
-
-                          <a class="tableLink" href="{$page4url}"><xsl:value-of select="normalize-space(jobName)" /><xsl:value-of select="normalize-space(jobID)" /></a>
-                    </td>
-                    <td align="center" valign="top" colspan="1">
-			 <xsl:value-of select="jobGroup" />
-                    </td>
-                    <td align="center" valign="top" colspan="1">
-                        <xsl:choose>
-                           <xsl:when test="count(datasets/dataset) &gt; 0">
-                               <xsl:for-each select="datasets/dataset">
-                                  <xsl:value-of select="current()" /><br />
-                               </xsl:for-each>         
-                           </xsl:when>
-                           <xsl:otherwise>
-                               n/a
-                           </xsl:otherwise>
-                        </xsl:choose>
-                    </td>
-                    <td align="center" valign="top" colspan="1">
-                        <xsl:choose>
-                           <xsl:when test="normalize-space(status)='running'">
-                               <font color="brown">running</font>
-                           </xsl:when>
-                           <xsl:when test="normalize-space(status)='queued'">
-                               <font color="brown">queued</font>
-                           </xsl:when>
-                           <xsl:when test="normalize-space(status)='success'">
-                               <font color="green">success</font>
-                           </xsl:when>
-                           <xsl:otherwise>
-                               <font color="red"><xsl:value-of select="normalize-space(status)" /></font>
-                           </xsl:otherwise>
-                        </xsl:choose>     
-                    </td>
-                    <td align="center" valign="top" colspan="1">
-                      <xsl:choose>
-                         <xsl:when test="normalize-space(postProcessingResult)='success'">
-                              <font color="green">success</font>
-                         </xsl:when>
-                         <xsl:when test="normalize-space(postProcessingResult)='error'">
-                              <font color="red">error</font>
-                         </xsl:when>
-                         <xsl:otherwise>
-                              <font color="black"><xsl:value-of select="normalize-space(postProcessingResult)" /></font>
-                         </xsl:otherwise>
-                      </xsl:choose>
-                   </td>
-                   <td align="center" valign="top" colspan="1"><xsl:value-of select="normalize-space(refRelease)" /></td>
-                 </tr>
-              </xsl:for-each>
-           </table>
-
-       <br />
-       <a class="pageLink" href="../../../../../../page1.xml">Back to home page</a>
-
-    </center>
-  </body>
-  </html>
-</xsl:template>
-</xsl:stylesheet>
-"""
-
-#########################################################################
-#########################################################################
-
-page4XSL = \
-"""<?xml version="1.0" encoding="UTF-8"?>
-<xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
-<xsl:template match="RTTpage4">
-  <html>
-  <head>
-
-  <style>
-  .level1TableHeader{font-family:arial;font-size:14pt;font-weight:bold;color:black;background-color:#8888ff}
-  .level2TableHeader{font-family:arial;font-size:12pt;font-weight:bold;color:black;background-color:#aaaaff}
-  .tableEntry{font-family:arial;font-size:10pt;color:#333333;background-color:#bbbbff;font-weight:bold}
-   a.permaLink{font-family:arial;font-size:7pt;color:#777777;text-decoration:underline;float:right}
-   a.permaLink:hover{color:black;text-decoration:none}              
-   a.tableLink{font-family:arial;font-size:10pt;color:#333333;text-decoration:none;border-bottom:1px solid}
-   a.tableLink:hover{font-family:arial;font-size:10pt;color:orange;text-decoration:none;border-bottom:1px solid}
-   a.pageLink{font-family:arial;font-size:10pt;color:#777777;text-decoration:underline}
-   a.pageLink:hover{font-family:arial;font-size:10pt;color:black;text-decoration:none}
-  </style>
-
-  </head>
-
-  <body bgcolor="#eeeeff">
-    <a href="" class="permaLink">Permalink</a>
-              
-    <center>
-
-       <xsl:comment>******************************************</xsl:comment>
-       <xsl:comment>* CREATE THE TOP OF THE PAGE INFORMATION *</xsl:comment>
-       <xsl:comment>******************************************</xsl:comment>
-
-       <h2>
-          <xsl:value-of select="normalize-space(overview/releaseName)" />/<xsl:value-of select="normalize-space(overview/runType)" /> run ---
-          <xsl:value-of select="overview/targetCMTCONFIG" /> --- <xsl:value-of select="normalize-space(overview/topProject)" />
-       </h2>
-       <h3> <xsl:value-of select="normalize-space(minder/jobName)" /><xsl:value-of select="normalize-space(minder/jobID)" /></h3>
-
-       <xsl:comment>******************************************</xsl:comment>
-       <xsl:comment>* CREATE MAIN TABLE                      *</xsl:comment>
-       <xsl:comment>******************************************</xsl:comment>
-
-       <br />
-           <table border="0" cellspacing="0" cellpadding="0">
-
-              <tr class="tableEntry" align="center" valign="top">
-                    <td>
-
-       <table cellpadding="4" cellspacing="0" border="0">
-           <tr class="level2TableHeader"><td colspan="2">Summary</td></tr>
-           <tr class="tableEntry">
-               <td align="right">Job Status:</td>
-               <td align="left">
-                   <xsl:choose>
-                      <xsl:when test="normalize-space(minder/status)='error'">
-                          <font color="red">error</font>
-                      </xsl:when>
-                      <xsl:when test="normalize-space(minder/status)='success'">
-                          <font color="green">success</font>
-                      </xsl:when>
-                      <xsl:otherwise> 
-                          <xsl:value-of select="minder/status" />
-                      </xsl:otherwise> 
-                   </xsl:choose>
-               </td>
-           </tr>
-           <xsl:choose>
-               <xsl:when test="count(minder/testResults/testResult)=0">
-                   <tr class="tableEntry"><td align="right">Tests:</td><td align="left">no tests</td></tr>
-               </xsl:when>
-               <xsl:otherwise>
-                   <xsl:for-each select="minder/testResults/testResult">
-                        <tr class="tableEntry"><td align="right">Test <xsl:value-of select="testName" />:</td>
-                           <xsl:choose>
-                              <xsl:when test="testStatus=0"><td align="left"><font color="green">success</font></td></xsl:when>
-                              <xsl:otherwise><td align="left"><font color="red">error</font></td></xsl:otherwise>
-                           </xsl:choose>
-                        </tr>
-                    </xsl:for-each> 
-               </xsl:otherwise>
-           </xsl:choose>
-           <xsl:choose>
-               <xsl:when test="count(minder/checkResults/checkResult)=0">
-                   <tr class="tableEntry"><td align="right">Checks:</td><td align="left">no checks</td></tr>
-               </xsl:when>
-               <xsl:otherwise>
-                   <xsl:for-each select="minder/checkResults/checkResult">
-                     <tr class="tableEntry"><td align="right">Check: </td>
-                           <xsl:choose>
-                              <xsl:when test="current()='success'"><td align="left"><font color="green">success</font></td></xsl:when>
-                              <xsl:otherwise><td align="left"><font color="red">error</font></td></xsl:otherwise>
-                           </xsl:choose>
-                        </tr>
-                    </xsl:for-each> 
-               </xsl:otherwise>
-           </xsl:choose>              
-              
-           <tr class="level2TableHeader"><td colspan="2">Classification</td></tr>
-
-           <xsl:choose>
-              <xsl:when test="not(minder/displayClass/node())">
-                 <tr><td class="tableEntry">None available</td></tr>
-              </xsl:when>
-              <xsl:otherwise>
-                 <tr class="tableEntry"><td align="right" valign="top">Class:</td><td align="left" valign="top"><xsl:value-of select="minder/displayClass" /></td></tr>
-                 <tr class="tableEntry"><td align="right" valign="top">Process:</td><td align="left" valign="top"><xsl:value-of select="minder/displayProcess" /></td></tr>
-                 <tr class="tableEntry"><td align="right" valign="top">Component:</td><td align="left" valign="top"><xsl:value-of select="minder/displayComponent" /></td></tr>
-
-              </xsl:otherwise>
-           </xsl:choose>
-              
-           <tr class="level2TableHeader"><td colspan="2">Documentation</td></tr>
-           <tr><td class="tableEntry">
-              <xsl:choose>
-              <xsl:when test="minder/jobDocString/node()"><xsl:value-of select="minder/jobDocString" /></xsl:when>
-              <xsl:otherwise>None available</xsl:otherwise>
-              </xsl:choose>
-           </td></tr>
-              
-           <tr class="level2TableHeader"><td colspan="2">Auditor stats</td></tr>
-           <xsl:choose>
-               <xsl:when test="minder/cpuTime and not(minder/cpuTime = 'Unknown')">
-                  <tr class="tableEntry"><td align="right" valign="top">CPU time:</td><td align="left" valign="top"><xsl:value-of select="minder/cpuTime" />s</td></tr>
-                  <tr class="tableEntry"><td align="right" valign="top">CPU time 2000:</td><td align="left" valign="top"><xsl:value-of select="minder/cpuTime2000" />s</td></tr>
-                  <tr class="tableEntry"><td align="right" valign="top">Walltime:</td><td align="left" valign="top"><xsl:value-of select="minder/wallTime" />s</td></tr>
-              </xsl:when>
-              <xsl:otherwise>
-                 <tr class="tableEntry"><td align="right" valign="top">CPU time:</td><td align="left" valign="top">Unknown</td></tr>
-                 <tr class="tableEntry"><td align="right" valign="top">CPU time 2000:</td><td align="left" valign="top">Unknown</td></tr>
-                 <tr class="tableEntry"><td align="right" valign="top">Walltime:</td><td align="left" valign="top">Unknown</td></tr>
-              </xsl:otherwise>
-           </xsl:choose>
-       <tr class="tableEntry"><td align="right" valign="top">Memory:</td><td align="left" valign="top"><xsl:value-of select="minder/mem" /></td></tr>
-            <tr class="tableEntry"><td align="right" valign="top">V-Memory:</td><td align="left" valign="top"><xsl:value-of select="minder/vmem" /></td></tr>
-            <tr class="tableEntry"><td align="right" valign="top">Batch Status:</td><td align="left" valign="top"><xsl:value-of select="minder/batchStatus" /></td></tr>
-       </table>
-
-                    </td>
-                    <td width="10" style="background-color:#ccccff"></td>
-                    <td align="left" valign="top" colspan="1">
-                         <xsl:choose>
-                             <xsl:when test="count(minder/deliveredkeepfiles/keepfile)=0">
-                                  <xsl:choose>
-                                      <xsl:when test="not(minder/errors/node())">No files available</xsl:when>
-                                      <xsl:otherwise><font color="#aa2222">This job was rejected by the RTT. The associated error message is:</font><br /><br /><xsl:value-of select="minder/errors" /></xsl:otherwise>
-                                  </xsl:choose>
-                             </xsl:when>
-                             <xsl:otherwise>
-                                  <xsl:for-each select="minder/deliveredkeepfiles/keepfile">
-                                       <xsl:sort select="current()" />
-                                       <xsl:variable name="keepFileLink"><xsl:value-of select="current()" /></xsl:variable>
-                                       <xsl:variable name="infoString"><xsl:value-of select="@info" /></xsl:variable>
-                                       <xsl:variable name="displayColor"><xsl:value-of select="@displayColor" /></xsl:variable>
-                                       <a class="tableLink" href="{$keepFileLink}" title="{$infoString}" style="color:{$displayColor}"><xsl:value-of select="current()" /></a><br />
-                                  </xsl:for-each>
-                             </xsl:otherwise>
-                         </xsl:choose>
-                    </td>
-                 </tr>
-           </table>
-
-       <br />
-       <a class="pageLink" href="../../../../../../../../../page1.xml">Back to home page</a> | <a class="pageLink" href="../../../../page2.xml">Back to packages page</a>
-
-    </center>
-  </body>
-  </html>
-</xsl:template>
-</xsl:stylesheet>
-"""
diff --git a/Tools/RunTimeTester/src/pathToProjectCMTDir.py b/Tools/RunTimeTester/src/pathToProjectCMTDir.py
deleted file mode 100644
index 85e26c09720157e4e489967c1a75dabbb0da0c4f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/pathToProjectCMTDir.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-def pathToProjectCMTDir(project, projectCMTDirs, useRunTime=False):
-    if useRunTime: return projectCMTDirs[project]['RunTime']
-    return projectCMTDirs[project]['Release']
-
-    # -------------------------------------------------------------------------------
-    
-def allProjectCMTDirs(projectCMTDirs, useRunTime=False):
-    releaseVals = []
-    runTimeVals = []
-    for releaseRunTimeDict in projectCMTDirs.values():
-        releaseVals.append(releaseRunTimeDict['Release'])
-        runTimeVals.append(releaseRunTimeDict['RunTime'])
-
-    if useRunTime:
-        return runTimeVals
-    return releaseVals
-    
-def allProjects(projectCMTDirs):
-    return projectCMTDirs.keys()
-
diff --git a/Tools/RunTimeTester/src/qaMetrics2OnePage.py.obsolete b/Tools/RunTimeTester/src/qaMetrics2OnePage.py.obsolete
deleted file mode 100755
index df361db59e449ca896cfbb9832309d42d32c43d8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/qaMetrics2OnePage.py.obsolete
+++ /dev/null
@@ -1,59 +0,0 @@
-import os
-from os.path import join
-
-
-def qaMetrics2OnePage(iDir, oDir):
-    " make a page of links to all the QAmetrics results html files"
-
-    oFileName = 'qaResultsOnePage.html'
-    ofn = join(oDir,oFileName)
-    ofile = open(ofn, 'w')
-
-    ofile.write('<html>\n')
-
-    for file in os.listdir(oDir):
-        if file.find('QAmetrics') != -1 and file.find('metrics.html') != -1:
-            line = '<a href=\"'+file+'\">'+file+'</a><br/>\n'
-            ofile.write(line)
-                        
-    ofile.write('</html>\n')
-
-    return oFileName
-
-def myArgs():
-	""" Returns directories containing viol files and results"""
-
-	iDir = '/Users/peter/Scratch'
-	oDir ='./Results'
-
-	if len(sys.argv) <= 1 or len(sys.argv) > 3:
-		print """Assuming defaults:
-		Viol files in """+iDir+""" 
-		Results in """ +oDir
-	elif len(sys.argv) > 2:
-		iDir = sys.argv[2]
-		
-	if len(sys.argv) > 1 :
-		oDir = sys.argv[1]
-
-
-	if not os.path.exists(iDir):
-		print 'No Viol directory:stopping '
-		sys.exit()
-	if not os.path.exists(oDir):
-		print 'No Results directory: stopping '
-		sys.exit()
-
-
-	return (iDir, oDir)
-
-if __name__ == '__main__':
-	"""
-	usage:
-	python  qaMetricsToOnePage.py iDir, oDir:
-        assumes results in oDir, input htmlfiles in iDir
-	
-	if arg(s) are missing, defaults are printed and used.
-	"""
-	iDir, oDir = myArgs()
-	qaMetrics2OnePage(iDir, oDir)
diff --git a/Tools/RunTimeTester/src/queryServer.py.obsolete b/Tools/RunTimeTester/src/queryServer.py.obsolete
deleted file mode 100755
index 1a8e09f1a01c715f5ca6b469ac13b0b2057e289e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/queryServer.py.obsolete
+++ /dev/null
@@ -1,49 +0,0 @@
-import xmlrpclib, sys
-from exc2string2 import exc2string2
-def getServer():
-    try:
-        transport =  xmlrpclib.Transport()
-        print transport
-        # server = xmlrpclib.ServerProxy('http://localhost:8000', transport, 'UTF-8', False,True)
-        server = xmlrpclib.ServerProxy('http://localhost:8001', transport, 'UTF-8', True, True)
-        print 'Successfully connected to server.'
-        return server
-    except:
-        print 'Could not connect to the webPageWriter %s' % exc2string2()
-        return None
-
-
-def menu(server):
-    while True:
-        print '0 - exit'
-        print '1 - tables'
-        print '2 - number of doms'
-
-        try:
-            ind = raw_input()
-        except:
-            print 'input error'
-
-        try:
-            ind = int(ind)
-        except:
-            continue
-        if ind not in range(3):
-            print 'index out of range',ind
-            continue
-
-        if ind == 0:
-            break
-
-        if ind == 1:
-            server.dumpTable()
-            continue
-
-        if ind == 2:
-            print server.countDoms()
-            continue
-        
-if __name__ == '__main__':
-
-    server = getServer()
-    menu(server)
diff --git a/Tools/RunTimeTester/src/readTopLevelConfigFile.py b/Tools/RunTimeTester/src/readTopLevelConfigFile.py
deleted file mode 100644
index 8b8c51bdcf5edd18e27edf9487d088e246690128..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/readTopLevelConfigFile.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from exc2string2      import exc2string2
-from RTTConfigParser  import RTTConfiguration
-from RTTSException    import RTTInputError
-
-def readTopLevelConfigFile(confFilePath, logger):
-    try:
-        rttConfig = RTTConfiguration(confFilePath)
-    except Exception, e:
-        m  = 'Exception thrown while reading  top level RTT Config file: %s:\nTraceBack:\n %s' % (str(e), exc2string2())
-        logger.critical(m)
-        raise RTTInputError(m)
-
-
-    
-    if rttConfig.configIsValid():
-        msg = 'Top level configuration %s is valid ' % str(rttConfig)
-        logger.debug(msg)
-        return rttConfig
-
-    
-    msg = 'Top level configuration %s is invalid: %s' % (str(rttConfig), rttConfig.text())
-    logger.critical(msg)
-    raise RTTInputError(msg)
-                
diff --git a/Tools/RunTimeTester/src/releaseToNum.py b/Tools/RunTimeTester/src/releaseToNum.py
deleted file mode 100755
index d65a887b8c22283050b0d210311e9479237150df..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/releaseToNum.py
+++ /dev/null
@@ -1,27 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-def releaseToNum(str):
-    """
-    Converts the release string x.y.z to a number
-    (eg "8.3.0"-> 830)
-    """
-    num=0
-    nightly = False
-
-    if str.find('rel') != -1: # nightly
-        nightly = True
-        return [nightly, num]
-    
-    cnumbers=str.split('.')
-    cnumbers.reverse()
-    powerTen=1
-    if len(cnumbers) != 3:
-        print 'releaseToNum: bad release: '+str
-    for cnum in cnumbers:
-        num=num+powerTen*int(cnum)
-        powerTen=powerTen*10
-
-    return [nightly, num]
-
-if __name__ == '__main__':
-    print releaseToNum('8.3.0')
diff --git a/Tools/RunTimeTester/src/remakeWebPages.py.obsolete b/Tools/RunTimeTester/src/remakeWebPages.py.obsolete
deleted file mode 100755
index 814a050963910d1d0cc0db83730f7a2db35788cf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/remakeWebPages.py.obsolete
+++ /dev/null
@@ -1,22 +0,0 @@
-from anydbm import open
-from WWebPageWriter import CERNWebPageWriter
-
-moniDBLocation = '/afs/cern.ch/atlas/project/RTT/prod/moniFile.db'
-resBaseDir = '/afs/cern.ch/atlas/project/RTT/Results'
-
-db = open(moniDBLocation, 'r')
-
-
-webPageWriter = CERNWebPageWriter(resBaseDir)
-items = db.items()
-
-print 'found %d summaries' % len(items)
-
-argDict = {
-    'keyDocList':        items,
-    'cmtTargetToUpdate': [],
-    'packageToUpdate':   [],
-    'mindersToUpdate':   []
-    }
-
-webPageWriter.update(argDict)
diff --git a/Tools/RunTimeTester/src/requirementsWriter.py.obsolete b/Tools/RunTimeTester/src/requirementsWriter.py.obsolete
deleted file mode 100755
index 893c0f77bc3a9aa34775ada839ad0070a2a2d5e8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/requirementsWriter.py.obsolete
+++ /dev/null
@@ -1,84 +0,0 @@
-import os, os.path, string
-
-from Tools import extractReleasePart
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-import Reporter
-globalReporter = Reporter.getGlobalReporter()
-# -------------------------------------------------------------------------
-
-def requirementsWriter_Project(runPath, packages=[]):
-    "returns the 'home directory' requirements file"
-
-    distArea = '/afs/cern.ch/atlas/software/dist'
-
-    script  = 'set CMTSITE CERN\n'
-    script += 'set SITEROOT /afs/cern.ch\n'
-    script += 'macro ATLAS_DIST_AREA '+distArea+'\n'
-    script += 'macro ATLAS_TEST_AREA '+runPath+'\n'
-    script += 'use AtlasLogin AtlasLogin-* $(ATLAS_DIST_AREA)\n'
-    for package in packages:
-        head, tail = os.path.split(package)
-        # script += 'use '+tail+' '+tail+'* '+head+'\n'
-        
-    print script
-    return script 
-
-# -------------------------------------------------------------------------
-
-# def requirementsWriter(releasePath, runPath, runType, packages=[]):
-def requirementsWriter(distArea, release, runPath, runType, packages=[]):
-    "returns the 'home directory' requirements file"
-
-    # releasePath is full path to release area e.g.
-    # /afs/cern.ch/atlas/software/dist/8.4.0 or
-    # /afs/cern.ch/atlas/software/dist/nightlies/rd/atlrel_1
-
-    
-
-    releasePath = '/'+release
-
-    if runType == 'kit':
-        script  = 'macro ATLAS_DIST_AREA '+ '"'+distArea+'"\n'
-        script += 'macro TestArea "'+runPath+'"\n'
-        
-        script += 'macro ATLAS_RELEASE "'+releasePath+'"\n'
-        script += 'path_prepend PATH ${ATLASSETTINGSROOT}/scripts\n'
-        script += 'set INSTALLAREA ${CMTINSTALLAREA}\n'
-        
-        script += 'path_prepend CMTPATH $(TestArea)\n'
-        script += 'path_remove CMTPATH "$(ATLAS_DIST_AREA)"\n'
-        script += 'path_append CMTPATH "$(ATLAS_DIST_AREA)$(ATLAS_RELEASE)"\n'
-        script += '$(ATLAS_DIST_AREA)/$(ATLAS_RELEASE)/External\n'
-
-        for package in packages:
-            head, tail = os.path.split(package)
-            script += 'use '+tail+' '+tail+'* '+head+'\n'
-
-    else: # not a kit => is a release
-
-        script  = 'set CMTSITE CERN\n'
-        script  = 'set SITEROOT /afs/cern.ch\n'
-        script += 'macro ATLAS_DIST_AREA '+ '"'+distArea+'"\n'
-        script += 'macro TestArea "'+runPath+'"\n'
-        script += 'macro ATLAS_RELEASE "'+releasePath+'"\n'
-        script += 'use AtlasSettings v* $(ATLAS_DIST_AREA)\n'
-        script += 'path_prepend CMTPATH $(TestArea)\n'
-        script += 'path_remove CMTPATH "$(ATLAS_DIST_AREA)"\n'
-        script += 'path_append CMTPATH "$(ATLAS_DIST_AREA)$(ATLAS_RELEASE)"\n'
-        
-    
-        for package in packages:
-            head, tail = os.path.split(package)
-            script += 'use '+tail+' '+tail+'* '+head+'\n'
-
-        logger.debug(script)
-        return script
-   
-        
-        print script
-        return script 
-
diff --git a/Tools/RunTimeTester/src/rmTree.py.obsolete b/Tools/RunTimeTester/src/rmTree.py.obsolete
deleted file mode 100755
index d9ab10e72ef6e0fe48cc2073d59f29d8cd7e28d0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/rmTree.py.obsolete
+++ /dev/null
@@ -1,22 +0,0 @@
-#!/usr/bin/python
-import os, sys
-
-from os.path import join, isdir, isfile, islink
-
-def rmTree(top):
-    for file in os.listdir(top):
-        path=join(top, file)
-        if islink(path):
-            os.remove(path)
-        elif isfile(path):
-            # os.chmod(path, 0777)
-            os.remove(path)
-        elif isdir(path):
-            rmTree(path)
-
-    os.rmdir(top)
-
-if __name__ == "__main__":
-
-    for dname in sys.argv[1:]: rmTree(dname)
-                          
diff --git a/Tools/RunTimeTester/src/rttClassLister.py.obsolete b/Tools/RunTimeTester/src/rttClassLister.py.obsolete
deleted file mode 100755
index b7c89e4350f6cfa2c86398c7e873f21033159a2d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/rttClassLister.py.obsolete
+++ /dev/null
@@ -1,69 +0,0 @@
-import glob, imp, sys, os, types, string
-
-
-def loadModule(f):
-
-    try:
-        return sys.modules[f]
-    except KeyError:
-        pass
-    
-    fp, pathname, description = imp.find_module(f)
-    # print 'module', f
-    # print 'pathname',pathname
-    # print 'description',description
-    try:
-        return imp.load_module(f, fp, pathname, description)
-    finally:
-        if fp: fp.close()
-
-
-def rttClassLister(srcPath):
-    outText = ''
-    nloadGood = 0
-    nloadBad  = 0
-    classes = [] # a list of all classes instantiated accumulated over all runs
-    files = glob.glob(os.path.join(srcPath,'*.py'))
-    # files = ['./SrcPaths.py']
-    # files = [f for f in files if f.find('Paths')>=0]
-    for f in files:
-        f = os.path.basename(f)
-        mf = f[:-3]
-        msg =  'loading: %s:' % mf
-        msg = string.ljust(msg,50)
-        outText += msg
-        try:
-            module = loadModule(mf)
-            outText +=  'OK\n'
-            nloadGood += 1
-        except Exception, e:
-            outText +=  'ERROR\n'
-            nloadBad += 1
-            continue
-
-        # dictionary has a string: class
-        # where the string is the class name, and the class has the
-        # full module qualifictation
-
-        # find the classes for the module
-        nclasses = [i for i  in module.__dict__.keys() if
-                    isinstance(module.__dict__[i], types.ClassType)]
-
-        # remove those that do not come from the RTT
-        toRemove = []
-        nonRTTmodules = ['os', 'xml', 'popen']
-        for nonRTTmodule in nonRTTmodules:
-            toRemove =  [n for n in nclasses
-                         if str(module.__dict__[n]).startswith(nonRTTmodule)]
-        [nclasses.remove(n) for n in toRemove]
-        classes.extend(nclasses)
-
-    temp = {}
-    for a in classes: temp[a]=None
-    classes = temp.keys()
-    classes.sort()
-    return (classes, outText, nloadGood, nloadBad)
-
-if __name__ == '__main__':
-    startPath = os.getcwd()
-    print rttClassLister(startPath )
diff --git a/Tools/RunTimeTester/src/setMaxMem.py b/Tools/RunTimeTester/src/setMaxMem.py
deleted file mode 100755
index 90e9bcbda9503b334ad01ddfa889538991a83d88..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/setMaxMem.py
+++ /dev/null
@@ -1,20 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTSException import RTTSystemResponseError
-
-import commands
-
-def setMaxMem():
-    cmd = 'ulimit -Hv'
-    sc, maxmem  = commands.getstatusoutput(cmd)
-    if sc != 0:
-        raise RTTSystemResponseError('bad status %d from %s %s' %  (sc, cmd, maxmem))
-    
-    cmd = 'ulimit -v %s' % maxmem
-    sc, out = commands.getstatusoutput(cmd)
-    if sc != 0: raise RTTSystemResponseError('bad status %d from %s %s' % (sc, cmd, out))
-
-    return maxmem
-if __name__ == '__main__':
-   
-    print setMaxMem()
diff --git a/Tools/RunTimeTester/src/setupCMTForNonLocalRun.py.obsolete b/Tools/RunTimeTester/src/setupCMTForNonLocalRun.py.obsolete
deleted file mode 100644
index ac1e8cd5f68b93772aa2f477b8ff151ff842eb50..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/setupCMTForNonLocalRun.py.obsolete
+++ /dev/null
@@ -1,19 +0,0 @@
-from ShellCommand           import ShellCommand
-from Factory_CMTConfigurer  import CMTConfigurerFactory
-
-def setupCMTForNonLocalRun(parametersForCMT, logger):
-    # Not necessary if this is a local run
-    if not parametersForCMT.localRttRun:
-        factory = CMTConfigurerFactory(logger)
-        cmtConfigurer = factory.create(parametersForCMT)
-        cmtConfigurer.makeHomeCMTDir()
-        
-        nlines = -1 # do not limit the number of lines
-        cmds = [cmtConfigurer.configure()]
-        shellCom = ShellCommand(cmds, logger, nlines)
-        reply = shellCom.getReply()
-        m = 'Reply on configuring CMT: %s' % str(reply)
-        logger.debug(m)
-        
-        statusText = 'initialise - made CMT directories'
-        logger.info(statusText)
diff --git a/Tools/RunTimeTester/src/setupLoggerForDebug.py.obsolete b/Tools/RunTimeTester/src/setupLoggerForDebug.py.obsolete
deleted file mode 100755
index 1cbf3815f23565dfbdc6a5917c4d972008f82c36..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/setupLoggerForDebug.py.obsolete
+++ /dev/null
@@ -1,33 +0,0 @@
-import logging
-
-def setupLoggerForDebug(level=None):
-    # Set up logger, streams and formatter
-    msgFmt    = '%(asctime)s %(filename)s, %(lineno)s, %(levelname)s %(message)s'
-    tmeFmt    = '%d%b%y %H:%M:%S'
-    formatter = logging.Formatter(msgFmt, tmeFmt)
-
-    levels = {
-        'DEBUG':     logging.DEBUG,
-        'INFO':      logging.INFO,
-        'WARNING':   logging.WARNING,
-        'ERROR':     logging.ERROR,
-        'CRITICAL':  logging.CRITICAL
-        }
-    if level in levels:
-        level=levels[level]
-    else:
-        print 'Unknown level %s\n' %level
-        print 'Possible levels: ', str(levels.keys())
-        print 'setting level to DEBUG'
-        
-        level = logging.DEBUG
-        
-    logStreamHandler = logging.StreamHandler()
-    logStreamHandler.setFormatter(formatter)
-    # logStreamHandler.setLevel(logging.INFO) # change output level here
-    logStreamHandler.setLevel(level)
-    
-    logger = logging.getLogger('rtt')
-    logger.addHandler(logStreamHandler)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
diff --git a/Tools/RunTimeTester/src/setupLoggerForTestSuite.py.obsolete b/Tools/RunTimeTester/src/setupLoggerForTestSuite.py.obsolete
deleted file mode 100755
index f8f4a48cf0b4d33c03946350c2eef42648a29086..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/setupLoggerForTestSuite.py.obsolete
+++ /dev/null
@@ -1,56 +0,0 @@
-import logging, sys, os
-from exc2string         import exc2string
-from Tools              import changePerm
-
-def setupLoggerForTestSuite(level=None):
-    # Set up logger, streams and formatter
-    msgFmt    = '%(asctime)s %(filename)s, %(lineno)s, %(levelname)s %(message)s'
-    tmeFmt    = '%d%b%y %H:%M:%S'
-    formatter = logging.Formatter(msgFmt, tmeFmt)
-
-    levels = {
-        'DEBUG':     logging.DEBUG,
-        'INFO':      logging.INFO,
-        'WARNING':   logging.WARNING,
-        'ERROR':     logging.ERROR,
-        'CRITICAL':  logging.CRITICAL
-        }
-    if level in levels:
-        level=levels[level]
-    else:
-        print 'Unknown level %s\n' %level
-        print 'Possible levels: ', str(levels.keys())
-        print 'setting level to DEBUG'
-        
-        level = logging.DEBUG
-
-
-
-
-    fn = 'testSuite.log'
-    print 'Making a logger for TestSuite %s' % fn
-    try:
-        open(fn, 'w')
-        changePerm(fn,0666)
-    except Exception, e:
-        print 'Could not open a logger file for TestSuite %s' % fn
-        print exc2string(sys.exc_info())
-  
-    fileHandler = logging.FileHandler(fn)
-    fileHandler.setLevel(level)
-    fileHandler.setFormatter(formatter)
-    
-    # logSreamHandler is not used ....
-    logStreamHandler = logging.StreamHandler()
-    logStreamHandler.setFormatter(formatter)
-    # logStreamHandler.setLevel(logging.INFO) # change output level here
-    # logStreamHandler is not used ....
-    logStreamHandler.setLevel(level)
-    
-    logger = logging.getLogger('rtt')
-    # logger.addHandler(logStreamHandler)
-    logger.addHandler(fileHandler)
-    print 'setupLoggerForTestSuote handlers %s' % str(logger.handlers)
-    logger.setLevel(logging.DEBUG) # Do Not change output level here
-
-    print logger.handlers
diff --git a/Tools/RunTimeTester/src/setupRunDir.py.OBSOLETE b/Tools/RunTimeTester/src/setupRunDir.py.OBSOLETE
deleted file mode 100755
index 7877e7298868dd802cda85a44aa6fc830decb273..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/setupRunDir.py.OBSOLETE
+++ /dev/null
@@ -1,184 +0,0 @@
-#===== RTT MODULES =====
-from requirementsWriter      import requirementsWriter
-from jobGroupKitFactory      import jobGroupKitFactory
-from scriptWriterFactory     import scriptWriterFactory
-from PathCompleter           import PathCompleter
-from releaseToNum            import releaseToNum
-from dataFiles2String        import dataFiles2String
-from getInheritanceHeirarchy import getInheritanceHeirarchy
-#===== PYTHON MODULES ==
-import os, shutil, anydbm
-from os.path              import join, splitext
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger()
-# -------------------------------------------------------------------------
-
-
-#--------------------------------------------------------------------------
-
-def setupRunDir(jMinder):
-
-    """
-    Creates run directory (obtained from the descriptor)
-
-    Uses JobGroupKits to establish what is needed to be copied to it.
-    Uses ScriptWriter to write a run script to it.
-    """
-    logger.info("entering setupRunDir")
-
-    jDescriptor = jMinder.jDescriptor
-    runPath     = jDescriptor.runPath
-
-    try:
-        # createRunDir(desc)
-        rmdirMkdir(runPath)
-        logger.debug('making '+runPath)
-    except Exception, e:
-        # later on this should be reported to a Global Reporter	
-        # for now simply remove this job from the list
-        errmsg = exc2string(sys.exc_info())
-        logger.error( errmsg)
-        descList.remove(desc)
-
-
-    os.chdir(runPath)
-        
-    logger.debug("writing requirements to  "+os.getcwd())
-    open("requirements","w").write(requirementsWriter(jDescriptor.paths.release, runPath))
-    #
-    # If the following two lines were done in separate os.system calls, the
-    # second would fail as it would not know about the env variables of the
-    # first.
-    #
-    # os.system(
-    #    "source /afs/cern.ch/sw/contrib/CMT/v1r16p20040701/mgr/setup.sh;cmt config")os.system(
-    #os.system(
-    #   "source /afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;cmt config")
-
-    nightly,relNum = releaseToNum(jDescriptor.paths.release)
-    if nightly or relNum >=870:
-        os.system(
-            "source /afs/cern.ch/sw/contrib/CMT/v1r16p20040901/mgr/setup.sh;cmt config")
-    else:
-        os.system(
-            "source /afs/cern.ch/sw/contrib/CMT/v1r14p20031120/mgr/setup.sh;cmt config")
-        
-
-    logger.debug("   asking ScriptWriter to write run script")
-    
-    runScriptWriter = scriptWriterFactory(jDescriptor)
-    scr = runScriptWriter.runScript()
-
-    
-    logger.debug("copying auxilliary files")  
-    # shutil.copy(join(jDescriptor.paths.sharePath, self.jobOptions), runPath)
-    kit = jobGroupKitFactory(jDescriptor.jobGroup)
-    for file in kit.auxFiles(jDescriptor.paths):
-        shutil.copy(file, runPath)
-        fileName = os.path.basename(file)
-        os.chmod(join(runPath,fileName),0777)
-            
-    if jDescriptor.asciiRefFiles or jDescriptor.binRefFiles:
-        retrieveRefFiles2(jDescriptor)
-
-    aux = open(join(jDescriptor.paths.sharePath, "auxFiles.txt"), "r")
-    pc = PathCompleter()
-    for line in aux.readlines():
-        tokens  = line.split()
-        package = tokens[0]
-        file    = tokens[1]
-        fn      = pc.fileFromReleasedFile(package, file, jDescriptor.paths)
-        shutil.copy(fn, runPath)
-        
-        fileName = os.path.basename(fn)
-        os.chmod(join(runPath,fileName),0777)
-        
-        logger.debug("copied "+fn+" to "+runPath)
-        
-    for  item in kit.stringsForRunDir(jDescriptor):
-        filename = item[0]
-        string = item[1]
-        file = open(join(runPath, filename), 'w')
-        file.write(string)
-        file.close()
-
-
-    print getInheritanceHeirarchy(jDescriptor)
-    
-    if 'AthenaJobDescriptor' in getInheritanceHeirarchy(jDescriptor):
-        athenaSetUp(jDescriptor)
-            
-    logger.info("returning from setupRunDir()")
-
-
-
-
-def athenaSetUp(jDescriptor):
-    # dataLink = createDataLink(runPath,jDescriptor)
-    # Create the Zebra.P link only of there is DC1 data for the job
-    if jDescriptor.hasDataSet:
-        if jDescriptor.isDC1data:
-            logger.debug("   creating soft link to data set: "+jDescriptor.DC1dataSet)
-            dataLink = 'ZEBRA.P'
-            if os.path.exists(dataLink): os.remove(dataLink)
-            os.symlink(jDescriptor.DC1dataSet, dataLink)       
-            
-
-#--------------------------------------------------------------------------
-
-#return the appropriate run directory name for the soft link to the data,
-#depending on whether the latter is DC1 or DC2
-def createDataLink(runPath,jDescriptor):
-    if jDescriptor.isDC1data:
-        link = "ZEBRA.P"
-        logger.debug("createDataLink: " + jDescriptor.DC1dataSet + " is DC1 data")
-    else:
-        logger.debug("createDataLink: " + jDescriptor.DC1dataSet + " is DC2 data")
-    return link
-
-def writeRefFiles(dbHandle, jDescriptor):
-    files = jDescriptor.asciiRefFiles
-    files.extend(jDescriptor.binRefFiles)
-
-                
-    for file in files:
-        key = [unicode(jDescriptor.refRelease),
-               unicode(jDescriptor.paths.build),
-               unicode(jDescriptor.name),
-               dataFiles2String(jDescriptor),
-               unicode(file)]
-        
-        refFileName = join(jDescriptor.runPath,'refFile_'+file)
-        
-            
-        try:
-            refFile = dbHandle[key.__repr__()]
-            logger.debug("Found reference file in DB")
-            logger.debug("DataBase at: "+jDescriptor.paths.refDB)    
-            logger.debug("key = "+key.__repr__())
-            open(refFileName, 'w').write(refFile)
-        except KeyError:
-            logger.debug("Warning: NO reference file found in DB")
-            logger.debug("DataBase at: "+jDescriptor.paths.refDB)    
-            logger.debug("key = "+key.__repr__())
-            reporter = jDescriptor.reporter
-            reporter.failedDBKeyFind(jDescriptor.paths.refDB,key.__repr__())
-        
-#--------------------------------------------------------------------------
-
-def retrieveRefFiles2(jDescriptor):
-    try:            
-        dbHandle = anydbm.open(jDescriptor.paths.refDB, 'r')
-        writeRefFiles(dbHandle,jDescriptor)
-        
-    except anydbm.error:
-        reporter = jDescriptor.reporter
-        reporter.failedDBConnect(jDescriptor.paths.refDB)
-        msg = "Could not open reference file dbm "+jDescriptor.paths.refDB
-        logger.warning(msg)
-        
-#--------------------------------------------------------------------------
-def noRunDir(jMinder):
-    logger.info('Not setting up run directory!')
-    pass
diff --git a/Tools/RunTimeTester/src/simpleDataForMonitoring.py b/Tools/RunTimeTester/src/simpleDataForMonitoring.py
deleted file mode 100644
index a9ca9364e2771c2f2b24a5efecc4eab0ea6c185c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/simpleDataForMonitoring.py
+++ /dev/null
@@ -1,25 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-def simpleDataForMonitoring(minder):
-    """
-    return a dictionay of values for monitoring. repeats data in DOM
-    document, but is faster to access.
-    """
-        
-    dict =   {
-        'name':              minder.identifiedName,
-        'nTests':            0,
-        'done':              minder.isDone(),
-        'operatorCloseDown': minder.closedDownByOperator(),
-        'nTestsSuccess':     0,
-        'nTestsFailure':     0,
-        'nRetries':          0,
-        'jobSuccess':        False,
-        'jobFailure':        False,
-        'ppSuccess':         False,
-        'ppFailure':         False
-        }
-    
-    
-    return dict
-
diff --git a/Tools/RunTimeTester/src/sizeOf.py.obsolete b/Tools/RunTimeTester/src/sizeOf.py.obsolete
deleted file mode 100755
index 779a3e309c52a507ebadb2e0d5f17aa12446babf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/sizeOf.py.obsolete
+++ /dev/null
@@ -1,113 +0,0 @@
-import sys
-from types import *
-def mergeDict(src, dest):
-      for k in src.keys():
-            dest.setdefault(k,0)
-            dest[k] += src[k]
-      
-
-def sizeOf(obj, logger):
-
-      objType  = type(obj)
-      
-      typeDict = {
-      BooleanType:         0,
-      BufferType:          0,
-      BuiltinFunctionType: 0,
-      BuiltinMethodType:   0,
-      ClassType:           0,
-      CodeType:            0,
-      ComplexType:         0,
-      DictProxyType:       0,
-      DictType:            0,
-      DictionaryType:      0,
-      EllipsisType:        0,
-      FileType:            0,
-      FloatType:           0,
-      FrameType:           0,
-      FunctionType:        0,
-      GeneratorType:       0,
-      InstanceType:        0,
-      IntType:             0,
-      LambdaType:          0,
-      ListType:            0,
-      LongType:            0,
-      MethodType:          0,
-      ModuleType:          0,
-      NoneType:            0,
-      NotImplementedType:  0,
-      ObjectType:          0,
-      SliceType:           0,
-      StringType:          0,
-      StringTypes:         0,
-      TracebackType:       0,
-      TupleType:           0,
-      TypeType:            0,
-      UnboundMethodType:   0,
-      UnicodeType:         0,
-      XRangeType:          0,
-      }
-      hasStorage = [type(()), type([]), type({}), type('s')]
-      for c in hasStorage:
-            key = str(c)+'_storage'
-            typeDict[key]=0
-
-      if objType not in typeDict.keys():
-            logger.warning('sizeOf: unknown type, will not be counted: %s' % str(objType))
-            return {}
-      typeDict[objType] += 1
-      # for item in typeDict.items():
-      #      print '%30s   %5d' %  item
-
-      #sys.exit(0)
-
-      containers = [type(()), type([])]
-      if objType in containers:
-            typeDict[str(objType)+'_storage'] += len(obj)
-            
-            for e in obj:
-                  mergeDict(sizeOf(e, logger), typeDict)
-
-
-      if hasattr(obj, '__dict__'):
-            mergeDict(sizeOf(obj.__dict__, logger), typeDict)
-
-      if objType == type('a'):
-            typeDict[str(objType)+'_storage'] += len(obj)
-                                         
-      if objType == type({}):
-            typeDict[str(objType)+'_storage'] += 2*len(obj)
-            [mergeDict(sizeOf(v, logger), typeDict) for v in obj.values()]
-            
-                  
-            
-            
-                  
-      
-      return typeDict
-
-if __name__ == '__main__':
-      class B:
-            def __init__(self):
-                  self.b = 1
-                  #self.bs = 'bbbbbbbbb'
-      class A:
-            def __init__(self):
-                  self.i = 0
-                  self.j = 4.2
-                  self.k = 'aaaa'
-                  self.dict = {}
-                  # self.bclass = B
-                  self.binst  = B()
-                  self.tup    = ()
-                  self.list   = []
-                  self.ty     = type('a')
-      from PseudoLogger import PseudoLogger
-      d = sizeOf(A(), PseudoLogger() )
-      print '-------final-----'
-      for k in d.keys():
-            if d[k]>0:
-                  print '%40s   %5d' % ( k, d[k])
-      print '-------final-----'
- 
-    	
diff --git a/Tools/RunTimeTester/src/sortValues.py b/Tools/RunTimeTester/src/sortValues.py
deleted file mode 100644
index dfbdb95565d1068005f41cb602441df071757d53..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/sortValues.py
+++ /dev/null
@@ -1,7 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-def sortValues(dict):
-    sortedKeys = dict.keys()
-    values = []
-    [values.append(dict[k]) for k in sortedKeys]
-    return values
diff --git a/Tools/RunTimeTester/src/srcFileCounter.py.obsolete b/Tools/RunTimeTester/src/srcFileCounter.py.obsolete
deleted file mode 100755
index 7f79814df2673730047a6440456dc61a6ed56927..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/srcFileCounter.py.obsolete
+++ /dev/null
@@ -1,152 +0,0 @@
-from os.path           import join
-from getopt            import getopt, GetoptError
-from shellCommand      import shellCommand
-from formatCollection  import formatCollection 
-from glob              import glob
-import sys, os, string
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-smallLogger = logging.getLogger('rtt.standalone')
-# -------------------------------------------------------------------------
-
-def test():
-    lines = []
-    lines.append(' library -s/absolute/somewhere -a=option xxxx yyyy\n')
-    lines.append('a line that does not interest us\n')
-    lines.append(' library -s/absolute/somewhereElse -a=option aaa\n')
-    doit(lines)
-
-
-def decodeFileNames(tokens):
-
-    # remove all options except -s (source directory)
-    options = [token for token in tokens if token.startswith('-')]
-    bOptions = [op for op in options if not op.startswith('-s')]
-    [tokens.remove(op) for op in bOptions]
-
-    # assume -s dir file file file -s dir
-    # ie occaisional souirce directory, followed by file name
-    # is no source directory before file name assume soure directory default
-
-    if (tokens[-1].startswith('-s')):
-        msg = 'last token is a directory name \n %s' % formatCollection(tokens)
-        logger.error(msg)
-        return []
-    
-    logger.debug('tokens after removing options: %s' % str(tokens))
-    dflag = False
-    files = []
-    for token in tokens:
-
-        sdir = 'src'
-        if token.startswith('-s'):
-            dflag = True
-            continue
-        
-        if dflag:
-            dflag = False
-            sdir  = token
-            continue
-        
-    
-        files.append(os.path.join(sdir, token))
-
-    return files
-            
-def removeFiles(files):
-    removeSuffixes = ['.java']
-
-    rmfiles = []
-    for suffix in removeSuffixes:
-        rmfiles.extend([file for file in files if file.endswith(suffix)])
-
-    files = [file for file in files if file not in rmfiles]
-    return files
-   
-def doit(lines, cmtFilePath):
-    """
-    Find out what cmt says to act on by issuing 'cmt show constitunts'
-    in the cmt directory. We take the sourcee files to b those which go into
-    library - and so we keep only 'library' contituents.
-    """
-    # files = '' # the final list of files
-    files = []
-    nSrcFiles = 0
-    for line in lines:
-        tokens = string.split(line) # split line into tokens
-        tokens = [string.strip(token) for token in tokens] # clean-up whitespace
-        if tokens[0] != 'library': continue # a library?
-        # library name - so start from the third
-        # accumulate the list of files from the different libraries
-        files.extend(decodeFileNames(tokens[2:]))
-        
-        files = removeFiles(files) # remove files RuleChecker cannot handle
-        files = [os.path.normpath(os.path.join(cmtFilePath, '..', file)) for file in files]
-
-    # files might have been specified with wild cards in requiremnts file
-    expandGlob = []
-    [expandGlob.extend(list) for list in [glob(file) for file in files]]
-    nSrcFiles += len(expandGlob)
-
-    smallLogger.info( ( 'no of src files =  %d') % nSrcFiles)
-    smallLogger.info( ( 'src files = \n %s') % formatCollection(expandGlob))
-    return nSrcFiles, expandGlob
-
-
-def srcFileCounter(pathToCMTFile, paths):
-    lines = [
-        'cd %s' % paths.workDirs['base'],
-        'source setup.sh',
-        'cd '+pathToCMTFile,
-        'cmt show constituents'
-        ]
-    files=shellCommand(lines)
-    (nSrcFiles, fileList) = doit(files, pathToCMTFile)
-    return nSrcFiles, fileList
-    
-
-
-if __name__ == '__main__':
-    """
-    To run cd to a cmt directory with a requriements file,
-    and type
-    python srcFileCounter.py <pathToCMTFile>
-
-    if no path is given, the script runs in demo mode
-    """
-
-    from setupLoggerForDebug import setupLoggerForDebug
-    from findRTThome     import findRTThome
-
-    import Paths
-
-    setupLoggerForDebug()
-
-    (base, cfg, src) = findRTThome()
-
-    try:
-        opts,args = getopt(sys.argv[1:], 't')
-    except GetoptError:
-        print 'unknown option(s)', opt, args
-
-
-    testMode = True
-    if len(args)==0:
-        test()
-        sys.exit(0)
-        
-    pathToCMTfile = args[0]
-    if not os.path.exists(pathToCMTfile):
-        print 'unknown path '+pathToCMTfile
-        sys.exit(0)
-
-    logger.debug('path tp cmt file %s' % pathToCMTfile)
-
-    paths = Paths.fromScratch(cfg, base)
-    (nFiles, fileList) = srcFileCounter(pathToCMTfile, paths)
-
-    print 'srcFileCounter found',nFiles,'src files'
-    print 'files found \n %s' % fileList
-
-
diff --git a/Tools/RunTimeTester/src/superParser.py b/Tools/RunTimeTester/src/superParser.py
deleted file mode 100755
index 8a57fddc5ddd06e7c899ad7ad981d30e56f8031e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/superParser.py
+++ /dev/null
@@ -1,53 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-from exc2string2            import exc2string2
-from Factory_JobsXMLReader  import JobsXMLReaderFactory
-from JobMaker               import JobMaker
-
-from RTTSException          import RTTCodingError
-
-# -------------------------------------------------------------------------
-
-def superParser(paths, descFactory, minderFactory, logger):
-    """Returns a list of all minders (RTT, RTTpilot and ATN)"""
-    
-
-    jobListElement  = paths.packageXMLfile['rtt/jobList[0]']
-    try:
-        atnElement  = paths.packageXMLfile['atn[0]']
-    except IndexError:
-        atnElement = None
-            
-    factory  = JobsXMLReaderFactory(logger)
-
-    def makeMinders((mode, element), paths):
-        xmlReader = factory.create(mode, element, paths, descFactory)
-
-        try:
-            jobMinders = JobMaker(xmlReader, paths,  descFactory, minderFactory, logger).jobMinders
-            return jobMinders
-        except:
-            msg  = 'Exception while parsing package xml file '
-            msg += 'mode: %s package %s \n' % (mode, paths.package)
-            msg += 'Traceback:\n%s' % exc2string2() 
-            raise RTTCodingError(msg)
-
-        return []
-
-    # Always prcocess RTT jobs
-    modes = {paths.mode: jobListElement} # to make RTT and RTTpilot jobs
-
-    # process ATN if requested
-    if paths.atnPilotJobs:
-        if atnElement: modes['ATN'] = atnElement 
-        
-    jobMinders = []
-    [jobMinders.extend(makeMinders(m, paths)) for m in modes.items()]
-    
-    # remove pilot jobs if they are NOT requested
-    if not paths.rttPilotJobs:
-        jobMinders = [j for j in jobMinders if not j.rttPilotJob]
-
-    return jobMinders
-
diff --git a/Tools/RunTimeTester/src/testXMLwithDTD.py b/Tools/RunTimeTester/src/testXMLwithDTD.py
deleted file mode 100755
index 01e7ee719713c8cb0f41f712add778b0dfa13e47..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/testXMLwithDTD.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-
-if __name__ == '__main__':
-    import os, getopt
-
-    from xml.parsers.xmlproc import xmlproc, xmlval
-
-    opts, args = getopt.getopt(sys.argv[1:], '')
-    
-    if len(args) != 1:
-        print 'must supply the full path of the xml file'
-        sys.exit(0)
-
-    fn = args[0]
-    if not os.path.exists(fn):
-        print '%s: inexistant!' % fn
-        sys.exit(0)
-
-    try:
-        validator = xmlval.XMLValidator()
-        validator.parse_resource(fn)
-        print '%s: valid OK.' % fn
-    except Exception, e:
-        print '%s: NOT valid: FAIL.' % fn
diff --git a/Tools/RunTimeTester/src/updateSingleFile.py.obsolete b/Tools/RunTimeTester/src/updateSingleFile.py.obsolete
deleted file mode 100755
index 3816e43ffb63176474aca6a51974456fafcdc039..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/updateSingleFile.py.obsolete
+++ /dev/null
@@ -1,36 +0,0 @@
-import xmlrpclib, time, os.path
-from exc2string2 import exc2string2
-
-def doit(aPath):
-    try:
-        transport =  xmlrpclib.Transport()
-        # print transport
-        # server = xmlrpclib.ServerProxy('http://localhost:8000', transport, 'UTF-8', False,True)
-        server = xmlrpclib.ServerProxy('http://localhost:8001', transport, 'UTF-8', True, True)
-        # print 'Successfully connected to server.'
-    except:
-        print 'Could not instantiate ServerProxy\n %s' % exc2string2()
-        return
-
-    try:
-        server.areYouAlive()
-    except:
-        print '%s: server not responding' % time.ctime()
-        return
-
-    if os.path.exists(aPath):
-        server.updateSingleSummary(aPath)
-    else:
-        print '%s does not exist. Doing nothing.' % aPath
-    
-if __name__ == '__main__':
-    #########################################################################
-    # BEWARE THE FOLLOWING FILES POINT AT PROD
-    # DANGER OF DESTROYING PROD
-    #########################################################################
-    thePath = '/afs/cern.ch/atlas/project/RTT/Results/rel_3/dev/build/i686-slc4-gcc34-opt/offline/RTTSummary.xml'
-    #thePath = '/afs/cern.ch/atlas/project/RTT/Results/rel_2/bugfix/build/i686-slc4-gcc34-dbg/hlt/RTTSummary.xml'
-    # path = '/afs/cern.ch/atlas/project/RTT/Results/rel_0/dev/build/i686-slc3-gcc323-opt/RTTSummary.xml'
-    # path = '/afs/cern.ch/atlas/project/RTT/data/peter/Results/rel_4/dev/build/i686-slc3-gcc323-opt/RTTSummary.xml'
-
-    doit(thePath)
diff --git a/Tools/RunTimeTester/src/validateInputParams.py b/Tools/RunTimeTester/src/validateInputParams.py
deleted file mode 100644
index 62d408085486995a05bc9e46f31c665d6d23bdd3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/validateInputParams.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTSException import RTTInputError
-import os.path
-def validateInputParams(logger, RTTSrcDir, RTTLibDir, topLevelRTTconfFileLocation):
-    #----------------------------------------------
-    # check for the ok-ness of the passed parameter
-    #----------------------------------------------        
-    if not os.path.isabs(topLevelRTTconfFileLocation):
-        m = 'topLevelRTTconfFileLocation %s is not an absolute path; it must be. Exiting...' % topLevelRTTconfFileLocation
-        logger.critical(m)
-        raise RTTInputError(m)
-    if not os.path.exists(topLevelRTTconfFileLocation):
-        m = 'topLevelRTTconfFileLocation %s is not an absolute path; it must be. Exiting...' % topLevelRTTconfFileLocation
-        logger.critical(m)
-        raise RTTInputError(m)
-        
diff --git a/Tools/RunTimeTester/src/validateXMLFile.py b/Tools/RunTimeTester/src/validateXMLFile.py
deleted file mode 100755
index 9689786cb2a8afdac1631245f3bc60c26cd3df03..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/validateXMLFile.py
+++ /dev/null
@@ -1,152 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"run validation against a DTD. The DTD location is written into the xml file."
-
-import os.path, sys
-from xml.parsers.xmlproc      import xmlproc, xmlval, xmldtd
-from xml.dom.minidom          import parse
-from Evaluate                 import Evaluate
-from RTTSException            import RTTInputError
-from RTTSException            import RTTCodingError
-from exc2string2              import exc2string2
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-
-# ====================================================================================================
-# ====================================================================================================
-# ====================================================================================================
-
-class Validator:
-    __alternativeDTDs = {'RTT.dtd': ['http://cern.ch/atlasrtt/alternateDTD/RTT.dtd'],
-                         'unifiedTestConfiguration.dtd': ['http://cern.ch/atlasrtt/alternateDTD/unifiedTestConfiguration.dtd'],
-                         'vetoFile.dtd': ['http://cern.ch/atlasrtt/alternateDTD/vetoFile.dtd']}
-
-    def __init__(self, pathToXMLfile):
-        self.file = pathToXMLfile
-
-        try:
-            self._dom = parse(self.file)
-            self._dom.normalize()
-        except Exception, e:
-            m = 'Exception raised while parsing user input file:\n%s' % str(e)
-            raise RTTInputError(m)
-
-        self.parseOK  = True
-        
-        
-    def validate(self):
-        if not self.parseOK:
-            raise RTTInputError('User xml file failed parse')
-
-        try:
-            self._validate()
-        except Exception, e:
-            msg = 'Exception of type %s raised while validating xml file\nException:\n%s\nTraceback:\n%s' % (e.__class__.__name__, str(e), exc2string2())
-            raise RTTInputError(msg)
-
-        try:
-            self.ensureUniqueTrendID()
-        except Exception, e:
-            msg = 'Exception of type %s raised while emsuring unique trend id\nException:\n%s\nTraceback:\n%s' % (e.__class__.__name__, str(e), exc2string2())
-            raise RTTInputError(msg)
-        
-        logger.debug('%s is valid against the DTD.' % self.file)
-    
-        
-    def grabDTD(self, dtdPaths):
-        # Test if we can get DTD, if so read it
-        import socket
-        import urllib
-        def grab(dtdPath):
-            socket.setdefaulttimeout(15) # prevents unreachable URLs from hanging the request
-            conts = ''
-            try:
-                a = urllib.urlopen(dtdPath)
-                conts = a.read()
-                a.close()
-                socket.setdefaulttimeout(None)
-                return conts
-            except IOError, e:
-                socket.setdefaulttimeout(None)        
-                m = 'Inaccessible DTD! [%s]' % dtdPath
-                if not 'timed out' in str(e):
-                    m  = 'Inaccessible DTD [==> %s]: %s\n\n' % (str(e), dtdPath)
-                logger.error(m)
-                return conts
-            
-        for d in dtdPaths:
-            c = grab(d)
-            if c: return c
-        return ''
-
-    def ensureUniqueTrendID(self):
-        # new step: no XML file can have non-unique CoolDozer trend Id values.
-        docEl  = self._dom.documentElement
-        ath = [thing for thing in [el.getAttribute('trendId') for el in Evaluate('rtt/jobList/athena', docEl)] if thing]
-        trf = [thing for thing in [el.getAttribute('trendId') for el in Evaluate('rtt/jobList/jobTransform', docEl)] if thing]
-        dup = {}
-        [dup.setdefault(t, []).append(t) for t in ath]
-        [dup.setdefault(t, []).append(t) for t in trf]
-        dups = [(k,v) for k,v in dup.items() if len(v)>1]
-        if dups:
-            m  = 'InavlidXMLFile: \n%s is __NOT__ valid XML!\n' % os.path.basename(self.file)
-            m += 'It declares non-unique trendId attributes.\n'
-            for k,v in dups:
-                m += '%s: declared by %d jobs\n' % (k, len(v))
-            raise RTTInputError(m)
-
-    def _validate(self):
-        
-        xmlFileDTD = self._dom.doctype.systemId # get system DTD path from XML file (if it exists)
-        if not xmlFileDTD:
-            logger.info('No external DTD detected in XML file. Considering %s as valid.' % os.path.basename(self.file))
-            return True
-        
-        dtds = [xmlFileDTD]
-        dtds.extend(Validator.__alternativeDTDs.get(os.path.basename(xmlFileDTD), []))
-
-        dtdContent = self.grabDTD(dtds)
-        if not dtdContent:
-            m = 'Invalid XML File: Unable to access any version of the DTD %s. Considering XML file [%s] as invalid.' % (os.path.basename(xmlFileDTD),
-                                                                                                                         os.path.basename(self.file))
-            raise RTTInputError(m)
-
-        # now load this string dtd
-        dtd = xmldtd.load_dtd_string(dtdContent)
-        parser = xmlproc.XMLProcessor()
-        parser.set_application(xmlval.ValidatingApp(dtd, parser))
-        parser.dtd = dtd
-        parser.ent = dtd
-        
-        try:
-            parser.parse_resource(self.file)
-        except Exception, e:
-            m  = 'Invalid XML File: %s is __NOT__ valid against the DTD.' % self.file
-            m += 'Cause: Either this is not a valid DTD, or the XML file is genuinely invalid against the DTD'
-            raise RTTInputError(m)
-
-
-# ====================================================================================================
-# ====================================================================================================
-# ====================================================================================================
-
-if __name__ == '__main__':
-    import getopt, os, sys
-
-    try:
-        opts,args = getopt.getopt(sys.argv[1:],'')
-    except getopt.GetoptError:
-        print '1 usage validateXMLFile fn'
-        sys.exit(0)
-
-    if len(args) != 1:
-        print 'Usage: python validateXMLFile.py /path/to/xml_file'
-        sys.exit(0)
-        
-    file = args[0]
-    if not os.path.exists(file):
-        print 'unknown file', file
-        sys.exit(0)
diff --git a/Tools/RunTimeTester/src/viol2xml.py.obsolete b/Tools/RunTimeTester/src/viol2xml.py.obsolete
deleted file mode 100755
index 972c1909acd508072c6e771333f33d60557f2f72..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/viol2xml.py.obsolete
+++ /dev/null
@@ -1,632 +0,0 @@
-#! /usr/local/bin/python
-"""
-PS 9/04
-
-viol2XML.py
-scripts to relook for viol files in a directory,
-find which rules have been violated, and write out an xml file with the
-information. The xml file is converted to an html file for display.
-
-The scripts assume that all the viol files are in the same directory,
-and that the file name is xxxx_packagename_sourefilename.viol.
-However some care has been taking to isolate these details to the begiing of
-the script to ease adaption if the circumstances change.
-"""
-
-import sys, re, os, fnmatch
-
-from xml.sax.handler import ContentHandler
-import xml.sax
-
-from string import *
-
-# -------------------------------------------------------------------------
-import logging
-logger = logging.getLogger('rtt')
-# -------------------------------------------------------------------------
-
-def packageName(fullPath):
-	(head,tail) = os.path.split(fullPath)
-	return tail
-
-
-def getFileName(fullPath):
-    m = re.search("[^/]*viol", fullPath)
-    m = re.search("[^.]*", m.group(0))
-    fn = m.group(0)
-    return fn
-
-def getPackageName(fullPath):
-    for item in split(fullPath,'/'):
-        pn="Unknown"
-        pn = re.search("^.*-..-..-..", item)
-        if pn:
-            pn = pn.group(0)
-            break
-    return pn
-
-
-                
-def getViolations(vRules):
-
-    violations = {}
-
-    for rule in vRules:
-        
-        if violations.has_key(rule):
-            violations[rule] = violations[rule]+1
-        else:
-            violations[rule]  = 1
-
-    return violations
-            
-def writeViolatedRules(ofile, violations):
-    
-    for rule in violations.keys():
-        ofile.write('        <violation>\n')
-        ofile.write('          <rule>'+rule+'</rule>\n')
-        ofile.write('          <count>')
-        ofile.write("%s" % violations[rule])
-        ofile.write('</count>\n')
-        ofile.write('        </violation>\n')
-        
-true = 1
-false = 0
-
-ruleNames = [
-    'NF1', 'NF2',
-    'NM1',
-    'NI1', 'NI2',
-    'NC1', 'NC2', 'NC3', 'NC4', 'NC5','NC6', 'NC7', 'NC8', 'NC9',
-    'CO1', 'CO2', 'CO3', 'CO4', 'CO5',
-    'CF1', 'CF2', 'CF3', 'CF4', 'CF5',
-    'CL1', 'CL2', 'CL3', 'CL4', 'CL5', 'CL6', 'CL7', 'CL8', 'CL9', 'CL10', 'CL11', 
-    'CC1', 'CC2', 'CC3', 'CC4',
-    'CI1', 'CI2', 'CI3', 'CI4', 'CI5', 'CI6', 'CI7', 'CI8', 'CI9', 'CI10',
-    'CN1', 'CN2', 'CN3', 'CN4',
-    'CS1', 'CS2',
-    'CB1', 'CB2', 'CB3', 'CB4', 'CB5',
-    'CE1', 'CE2',
-    'CH1', 'CH2', 'CH3', 'CH4', 'CH5',
-    'CA1', 'CA2', 'CA3', 'CA4', 'CA5', 'CA6', 'CA7', 'CA8', 'CA9', 'CA10',
-    'CA11', 'CA12', 'CA13', 'CA14',
-    'CR1', 'CR2', 'CR3', 'CR4',
-    'CP1', 'CP2', 'CP3', 'CP4', 'CP5','CP6', 'CP7', 'CP8', 'CP9', 'CP10','CP11',
-    'CP12', 
-    'CT1', 'CT2', 'CT3', 'CT4',
-    'SG1', 'SG2', 'SG3', 'SG4', 'SG5', 'SG6', 'SG7', 'SG8',
-    'SC1', 'SC2', 'SC3', 'SC4'
-    ]
-
-def makeViolDict():
-    vd = {}
-    for rule in ruleNames:
-        vd[rule]=0
-    return vd
-        
-class PackageViolation:
-	"A class built around a dictionaries cotaining violation counts."
-	
-	#def __init__(self, package, vFiles, vDir):
-	def __init__(self, package, vFiles):
-		""" initialise with:
-		- package name
-		- a list of viol files for this package.
-		"""
-		self.vFiles      = vFiles
-		self.package     = package
-		self.vDicts      = self.makeDicts()
-		self.numberOfViols = 0
-		# make a list of rules that are violated (one entry
-		# per violation)
-		self.makeDicts()
-
-	def makeDicts(self):
-		"""
-		Returns a dictionary of dictionaries.
-		Outer dictionary keyed on the viol file names.
-		entry is a dictionary with key rule, entry violation count
-		"""		
-		dicts = {}
-		for file in self.vFiles:
-			dict=makeViolDict()
-			
-			rules = self.getViolatedRules(file)
-				
-		    
-			for rule in rules:
-				dict[rule] = 1+dict[rule]
-				    
-
-			dicts[file]=dict
-		
-		return dicts
-
-	def getViolatedRules(self, vFile):
-		"""
-		Return a list of violations for source file processed by
-		RuleChecker.
-		"""
-		
-		vRules = []
-
-		# fullPath = os.path.join(self.vDir, vFile)
-		
-		file=open(vFile, 'r')
-
-		for line in file.readlines():
-			m = re.search("RULE ", line)
-			
-			if m:
-				words = splitfields(line, ' ')
-				vRules.append(words[1])
-		return vRules
-
-	def getViolationDictionaries(self):
-		return self.vDicts
-                
-        
-	def dump(self):
-		print "dump"
-		print "   package:  ",self.package
-
-		for file in self.vDicts:
-			print "      file: ", file
-			for rule in self.vDicts[file]:
-				print "      rule: ", rule,'   counts: ',self.vDicts[file][rule]
-	
-
-class ZeroSupressedViolations(PackageViolation):
-	def violatedRules(self):
-		rules = {}
-		
-		for rule in self.violations:
-			if self.violations[rule] > 0:
-				rules[rule] = self.violations[rule]
-                
-		return rules
-        
-class AllViolations(PackageViolation):
-	def violatedRules(self):
-		return self.violations
-    
-class XMLwriter:
-
-	def __init__(self, rDir):
-		# path offset to the results directory
-		self.rDirectory = rDir
-		
-		self.fileName = os.path.join(self.rDirectory,
-					     'RuleChecker.xml')
-		ofile = ofile=open(self.fileName,'w')
-		ofile.write('<?xml version="1.0" ?>\n')
-		ofile.write(
-			'<?xml-stylesheet href="../big.xsl" type="application/xml"?>\n')
-		ofile.write('<rulechecker>\n')
-		ofile.close()
-
-
-	def process(self, package, vDicts, srcInfo, fFiles):
-		self.processPackage(package, srcInfo, fFiles)
-		
-		for file in vDicts.keys():
-			self.processFile(file)
-			self.processDict(vDicts[file])
-			self.processFileEnd()
-
-		self.processPackageEnd()
-
-	def processPackage(self, pn, srcInfo, failedList):
-		ofile=open(self.fileName, 'a')
-		ofile.write('  <package>\n')
-		ofile.write('    <packagename>'+pn+'</packagename>\n')
-		ofile.write('    <nsrcfiles>'+srcInfo[0]+'</nsrcfiles>\n')
-		ofile.write('    <srcFileList>'+srcInfo[1]+'</srcFileList>\n')
-		for file in failedList:
-			ofile.write('    <failed>'+file+'</failed>\n')
-			
-		ofile.close()
-
-	def processPackageEnd(self):
-		ofile=open(self.fileName, 'a')
-		ofile.write('  </package>\n')
-		ofile.close()
-
-	def processFile(self, file):
-		ofile=open(self.fileName, 'a')
-		ofile.write('    <file>\n')
-		ofile.write('      <filename>'+file+'</filename>\n')
-		ofile.close()
-
-	def processFileEnd(self):
-		ofile=open(self.fileName, 'a')
-		ofile.write('    </file>\n')
-		ofile.close()
-
-	def processDict(self, dict):
-		ofile=open(self.fileName, 'a')
-		for rule in dict:
-			ofile.write('      <rule>\n')
-			ofile.write('         <label>'+rule+'</label>\n')
-			ofile.write('         <count>'+str(dict[rule])+'</count>\n')
-			ofile.write('      </rule>\n')
-
-		ofile.close()
-		
-		
-	def close(self):
-		ofile=open(self.fileName, 'a')
-		
-		ofile.write('</rulechecker>\n')
-		ofile.close()
-
-
-
-def find(pattern, startdir=os.curdir):
-	matches = []
-	os.path.walk(startdir, findvisitor, (matches, pattern))
-	matches.sort
-	return matches
-
-def findvisitor( (matches, pattern), thisdir, nameshere):
-	for name in nameshere:
-		if fnmatch.fnmatch(name, pattern):
-			fullpath = os.path.join(thisdir, name)
-			matches.append(fullpath)
-        
-
-def myArgs():
-	""" Returns directories containing viol files and results"""
-
-	vDir = '/unix/www/html/atlas/atlfast/NightlyTestsDev2/opt/N.X.0/release/10.2.0'
-	rDir ='/unix/www/html/atlas/atlfast/NightlyTestsDev2/opt/N.X.0/release/10.2.0'
-
-	if len(sys.argv) <= 1 or len(sys.argv) > 3:
-		mess =  'Assuming defaults:\n'
-		mess += 'Viol files in '+vDir+'\n'
-		mess += 'Results in ' +rDir+'\n'
-	elif len(sys.argv) > 2:
-		vDir = sys.argv[2]
-		
-	if len(sys.argv) > 1 :
-		rDir = sys.argv[1]
-
-
-	if not os.path.exists(vDir):
-		logger.error('No Viol directory:stopping ')
-		sys.exit()
-	if not os.path.exists(rDir):
-		logger.error('No Results directory: stopping ')
-		sys.exit()
-
-
-	return (vDir, rDir)
-
-
-class HTMLwriter(ContentHandler):
-
-	def __init__(self, htmlFile):
-
-
-		self.process={
-			'packagename':   self.handlePackageName,
-			# 'releasepath':   self.handleReleasePath,
-			'srcFileList':   self.handleSrcFileList,
-			'nsrcfiles':     self.handleNSrcFiles,
-			'failed':        self.handleFailed,
-			'filename':      self.handleFileName,
-			'label':         self.handleLabel,
-			'count':         self.handleCount
-			}
-
-		self.nViolations = {}
-
-		self.nPackages = 0
-		self.totalViolations = {}
-
-		self.fileName = htmlFile
-		ofile = open(self.fileName, 'w')
-
-		
-		ofile.write('<html>\n')
-		ofile.write(' <head>\n')
-		ofile.write('   <title>RuleChecker Results</title>\n')
-		ofile.write('  </head>\n')
-		ofile.write('  <body>\n')
-		ofile.write('    <h2>RuleChecker Results</h2>\n')
-		ofile.close()
-
-
-
-	def startElement(self, name, attributes):
-
-		if name in self.process.keys():
-			# print 'will process tag ', name
-			self.charsToProcess = True
-			self.what = name
-		elif name == 'file':
-			# print 'will process tag ', name
-			self.charsToProcess = False
-			self.startFileHandler()
-		elif name == 'package':
-			self.nPackages += 1
-			# print 'will process tag ', name
-		else:
-			self.charsToProcess = False
-			# print 'will not process tag ', name
-
-
-		
-		
-	def endElement(self, name):
-		
-		self.charsToProcess = False
-		if name == 'file': self.endFileHandler()
-
-
-
-	def handlePackageName(self, text):
-		
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h3>Package: '+text+'</h3>\n')
-		ofile.close()
-
-	def handleNSrcFiles(self, text):
-		
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h3>No of cxx files: '+text+'</h3>\n')
-		ofile.close()
-		
-	
-	def handleReleasePath(self, text):
-		
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h3>release path: '+text+'</h3>\n')
-		ofile.close()
-		
-	def handleSrcFileList(self, text):
-		
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h3>source files: '+text+'</h3>\n')
-		ofile.close()
-		
-	
-	def handleFailed(self, text):
-		
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h4>failed file: '+text+'</h4>\n')
-		ofile.close()
-		
-	
-	def handleFileName(self, text):
-
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <h4>File: '+text+'</h4>\n')
-		ofile.close()
-	
-		
-		
-	def startFileHandler(self):
-		self.nViolations.clear()
-
-	def endFileHandler(self):
-
-		self.dictToTableData(self.nViolations)
-
-
-			
-			
-	def dictToTableData(self, dict):
-		ofile = open(self.fileName, 'a')
-		ofile.write('    <table border=2>\n')	
-		ofile.write('      <tr>\n')
-
-		for key in dict.keys():
-			if dict[key] == 0:
-				del dict[key]
-
-		s = ''
-		for label in dict.keys():
-			s += '        <td>'+label+'</td>\n'
-		s+='      </tr>\n'
-			
-		for label in dict.keys():
-
-			s +=  '        <td>'
-			s += str(dict[label])
-			s += '</td>\n'
-
-	
-		ofile.write(s)
-		
-		ofile.write('      </tr>\n')		
-		ofile.write('    </table>\n')	
-	
-		
-		ofile.close()
-	def handleLabel(self, text):
-		self.label = text
-
-	def handleCount(self, text):
-		count  = int(text)
-		k = self.label
-		self.nViolations[k] = count+self.nViolations.get(k,0)
-		self.totalViolations[k] = count+self.totalViolations.get(k,0)
-		
-	def characters(self, text):
-
-		if self.charsToProcess: self.process[self.what](text)
-
-	def close(self):
-		ofile = open(self.fileName, 'a')
-
-		s = '   <h3>Summary</h3>'
-		s +=  '  Number of packages processed = '
-		s += str(self.nPackages)
-		s += '<br>\n'
-
-		
-		ofile.write(s)
-		ofile.write('  Total Number of Violations:<br>\n')
-		ofile.close()
-		
-		self.dictToTableData(self.totalViolations)
-
-		ofile = open(self.fileName, 'a')
-		ofile.write('  </body>\n')
-		ofile.write('</html>\n')
-		ofile.close()
-	
-
-def getNSrcFiles(vDir):
-	"""Constructs and returns a disctionary of viol files
-	keyed on the package name
-	
-	Assumes all viol files are in the same directory.
-	The package name is extracted  from the files name
-
-	The viol files are called RuleChecker_<package>_<file>.viol
-	"""
-
-	packageNSrcFiles = {}
-	logger.critical(vDir)
-	for file in os.listdir(vDir):
-		logger.critical('files in viol dir')
-		logger.critical(file)
-		if file.find('srcFileInfo') != -1:
-			sp=file.split('_')
-			lines = open(os.path.join(vDir,file), 'r').readlines()
-			nFiles = lines[0].strip()
-			lines = [line.strip() for line in lines[1:]]
-			srcFiles = reduce(lambda: x+y+' ', lines)
-			packageNSrcFiles[sp[1]] = (srcFiles, nFiles)
-
-	return packageNSrcFiles
-
-#writer = BigTableWriter()
-
-
-#path = "/afs/cern.ch/atlas/software/dist/rulechecker"
-
-def getViolFileDictionary(vDir):
-	"""Constructs and returns a disctionary of viol files
-	keyed on the package name
-	
-	Assumes all viol files are in the same directory.
-	The package name is extracted  from the files name
-
-	The viol files are called RuleChecker_<package>_<file>.viol
-	"""
-
-	packageVFiles = {}
-	for file in os.listdir(vDir):
-	
-		if file.find('.viol') != -1:
-			sp=file.split('_')
-			fullpath=os.path.join(vDir,file)
-		
-			packageVFiles.setdefault(sp[1],[]).append(fullpath)
-
-	return packageVFiles
-
-#writer = BigTableWriter()
-
-
-#path = "/afs/cern.ch/atlas/software/dist/rulechecker"
-
-
-def normalize_whitespace(text):
-	return ' '.join(text.split())
-		
-def isFailedViolFile(fn):
-	text = open(fn, 'r').read()
-	if text.find('Was expecting') == -1: return False
-	return True
-
-def getFailedVFiles(packageVFiles):
-	"Find the viol files with text indicating a crash "
-	failedViolFiles = {}
-	for package in packageVFiles:
-		vfiles = packageVFiles[package]
-		failedViolFiles[package]=[file for file in vfiles if isFailedViolFile(file)]
-	return failedViolFiles
-
-def removeFailedVFiles(allFiles, failedFiles):
-	for package in failedFiles:
-		for file in failedFiles[package]: allFiles[package].remove(file)
-
-def viol2xml(vDir, rDir):
-
-	logger.critical('vdir = '+vDir)
-	logger.critical('rdir = '+rDir)
-
-	doXML  = True
-	doHTML = True
-	
-	# find the viol file directory and results directories
-
-	# all the viol files for all the packages are in the directory vDir.
-	# construct a dictionay of viol files (full paths) for each package
-	packageVFiles      = getViolFileDictionary(vDir)
-	failedVFiles       = getFailedVFiles(packageVFiles)
-	removeFailedVFiles(packageVFiles, failedVFiles)
-	
-		
-	packageNSrcFiles = getNSrcFiles(vDir)
-
-	# For each leaf package:
-	#   Make a Writer object:
-	#      Write package tags
-	#   Process all the viol files:
-	#       Write source file tags
-	#       Write error counts
-
-	if doXML:
-		writer = XMLwriter(rDir)
-		
-		for package in packageVFiles:
-			# make a PackageViolation object for each package.
-			# this objects keeps track of the error counts.
-			all = AllViolations(package, packageVFiles[package])
-			
-			# all.dump()
-			
-			# Write the error counts to an xml file
-			# with the necesaary packagename and filename tags
-
-			logger.critical(str(packageNSrcFiles))
-			srcInfo = packageNSrcFiles[package]
-			fFiles  = failedVFiles[package] 
-			writer.process(package,
-				       all.getViolationDictionaries(),
-				       srcInfo,
-				       fFiles)
-
-		# write the closing tags to the xml file
-		writer.close()
-
-	htmlFile = ''
-	if doHTML:
-		# convert the xml file to html file.
-
-		htmlFile = 'RuleChecker.html'
-		writer = HTMLwriter(os.path.join(rDir,htmlFile))
-		parser = xml.sax.make_parser()
-		
-		parser.setContentHandler(writer)
-		
-		xmlFile = os.path.join(rDir, 'RuleChecker.xml')
-		parser.parse(xmlFile)
-		
-		writer.close()
-		return htmlFile
-				
-if __name__ == '__main__':
-	"""
-	usage:
-	python viol2XML rDir, vDir: assumes results in rDir, viol files in vDir
-	
-	if arg(s) are missing, defaults are printed and used.
-	"""
-	vDir, rDir = myArgs()
-	viol2xml(vDir, rDir)
diff --git a/Tools/RunTimeTester/src/xmlFileTester.py.obsolete b/Tools/RunTimeTester/src/xmlFileTester.py.obsolete
deleted file mode 100755
index d9fd09f13c61fdd776169f0eaa9c8ad2db227a96..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/xmlFileTester.py.obsolete
+++ /dev/null
@@ -1,36 +0,0 @@
-import sys
-from xml.sax.handler import ContentHandler
-import xml.sax, getopt, os.path
-
-
-class counterHandler(ContentHandler):
-    def __init__(self):
-        self.tags = {}
-    def startElement(self, name, attr):
-        if not self.tags.has_key(name):
-            self.tags[name]=0
-        self.tags[name] += 1
-
-
-if __name__ == '__main__':
-    opts, args = getopt.getopt(sys.argv[1:], '')
-    
-    if len(args) != 1:
-        print 'must supply the full path of the xml file'
-        sys.exit(0)
-    else:
-        print 'looking for ',args[0]
-
-    fn = args[0]
-    if not os.path.exists(fn):
-        print 'non existant file', fn
-        sys.exit(0)
-    
-    parser = xml.sax.make_parser()
-    handler = counterHandler()
-    parser.setContentHandler(handler)
-    parser.parse(fn)
-    tags = handler.tags.keys()
-    tags.sort()
-    for tag in tags:
-        print tag, handler.tags[tag]
diff --git a/Tools/RunTimeTester/src/xmlFormatter.py.obsolete b/Tools/RunTimeTester/src/xmlFormatter.py.obsolete
deleted file mode 100755
index 24da82ea6b39801a892fd8b8488e42abfcaad31c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/src/xmlFormatter.py.obsolete
+++ /dev/null
@@ -1,46 +0,0 @@
-import string
-import sys
-
-def process(fn,of):
-    lines = open(fn, 'r').readlines()
-    ofile = open(of, 'w')
-    nblank = 0
-    delta  = 3
-    for line in lines:
-
-        if line.find('<?') != -1:
-            print 'skipping xml version line'
-            ofile.write(string.strip(line)+'\n')
-            continue
-        
-        if line.find('DOCTYPE') != -1:
-            print 'skipping DTD line'
-            ofile.write(string.strip(line)+'\n')
-            continue
-        
-        nblankprev=nblank
-        
-        nstart  = string.count(line, '<')
-        nstop1  = string.count(line, '</')
-        nstop2 = string.count(line, '/>')
-        ncomnt = string.count(line, '<!')
-        nblank += (nstart-2*nstop1-nstop2-ncomnt)*delta
-
-        # print nstart, nstop1, nstop2, nblank
-        nline = string.strip(line)+'\n'
-        # print len(nline)
-        if nblank < nblankprev:
-            nnline = string.rjust(nline, len(nline)+nblank)
-        else:
-            nnline = string.rjust(nline, len(nline)+nblankprev)
-        ofile.write(nnline)
-        # print nnline
-
-if __name__ == '__main__':
-    fn = sys.argv[1]
-    of = 'new_'+fn
-    print 'formatting ',fn
-    print 'formatted file is',of
-    process(fn, of)
-    print 'done'
-    
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeApp.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeApp.py
deleted file mode 100644
index 2660fabc9062cd8637cb0c1bc7bbaff60702207f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeApp.py
+++ /dev/null
@@ -1,522 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubeApp.py
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief implementation of DozerApp and test_DozerApp classes
-
-import sys,os
-from DCubeUtils import *
-from DCubeOptParser import DCubeOptParser
-from DCubeValid import DCubeValid
-from DCubeConfig import DCubeConfig
-from DCubeTester import DCubeTester
-from DCubePHPWriter import DCubePHPWriter
-#import ROOT
-import unittest
-
-##
-# @class DCubeApp 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief DCube client application
-class DCubeApp( DCubeObject ):
-
-    ## parser options (optparse.Values)
-    opts = None
-
-    ## left over positional arguments as list
-    args = None
-
-    ## path to monitored root file 
-    monitoredURI = None
-
-    ## path to reference root file
-    referenceURI = None
-
-    ## monitored file handle
-    monTFile = None
-
-    ## reference file handle
-    refTFile = None
-
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    def __init__( self ):
-
-        parser = DCubeOptParser()
-        self.parsed = self.opts, self.args = parser.parse( sys.argv )
-
-        import ROOT
-        global ROOT
-        
-        monitoredURI = None
-        if ( len(self.args) == 2 ):
-            self.monitoredURI = self.args[-1] 
-                
-        DCubeLogger( self.opts.log, logName="", toConsole=True )
-
-        super( DCubeApp, self ).__init__( self )
-      
-        ## DCubeVersion instance
-        theVersion = DCubeVersion()
-
-        self.info( theVersion )
-        self.info( theVersion.python() )
-        self.info( theVersion.root() )
-
-       
-        ## DCubeValid instance
-        self.valid = DCubeValid()
-        
-        for line in str(self.valid).split("\n"):    
-            self.info( line )
-
-        self.__dumpOptions()
-        self.__dumpArgs()
-
-        ##
-        ## run dispatcher
-        ##
-
-        try:
-            ## generation of config XML file
-            if ( self.opts.generate and self.opts.reference ) :
-
-                self.info("RUN: generation of DCube configuration")
-
-                if ( DCubeConfig( self.parsed ).generate( ) ):
-                    self.info("*** SUCCESS ***")
-                    sys.exit(0)
-
-            elif ( self.opts.config and self.monitoredURI ):            
-
-                self.info("RUN: test suite execution")
-
-                runConfig = DCubeConfig( self.parsed ).getConfig( strict = False )
-
-                if ( self.monitoredURI ):
-                    if ( self.exists( self.monitoredURI ) ):
-                        self.__update( "monitored", self.monitoredURI )
-                    else:
-                        self.error("monitored root file %s not found" % self.monitoredURI )
-                else:
-                    raise DCubeException(  "wrong arguments, monitored file not specified!" )
-                
-            
-                if ( runConfig ):
-                    
-                    if ( self.runConfig ( runConfig ) ):
-
-                        if ( self.tester ):
-
-                            if ( self.__producePHPs() ):
-                                self.info("creation of PHP files completed")
-                                self.info("*** SUCCESS ***")
-                            else:
-                                self.warn("some problems occured when producing PHP files")
-                                self.info("*** WARNING ***")
-                            sys.exit(0)
-                        else:
-                            self.error("no DCubeTester instance, no summary information found")
-                            sys.exit(-1)
-                else:
-                    raise DCubeException( "bad configuration -  wrong configuration requested?" )
-            else:
-                self.error( "wrong set of CLI options, neither generation of config (-g) nor test suite excecution (-c CONFIG.XML MONITORED.root) specified" )
-                sys.exit(-1)
-                
-        except DCubeException, value:
-            self.epanic( value )
-            self.epanic("*** FAILURE ***")
-            sys.exit(-1)
-
-    ## testing and plotting using \<dcube\> node  
-    # @param self "Me, myself and Irene"
-    # @param xmlConfig dcube node from configuration file
-    def runConfig( self, xmlConfig ):
-        self.info("will run test suite using xml configuration..." )
-
-        ## output XML DOM document
-        self.xmldoc = xml.dom.minidom.Document() 
-
-        self.xmldoc.appendChild( xmlConfig )
-
-        ## overwritting xml config to match run config
-        self.debug("will overwrite xml config to match run config...")
-        xmlConfig.setAttribute( "branch", self.opts.branch ) 
-        xmlConfig.setAttribute( "install", self.opts.install )
-        xmlConfig.setAttribute( "cmtconfig", self.opts.cmtconfig )
-        xmlConfig.setAttribute( "project", self.opts.project )
-        xmlConfig.setAttribute( "jobId", self.opts.jobId )
-        self.debug("done!")
-
-        config = "/".join( [ xmlConfig.getAttribute( "branch" ),
-                             xmlConfig.getAttribute( "install" ),
-                             xmlConfig.getAttribute( "cmtconfig" ),
-                             xmlConfig.getAttribute( "project" ),
-                             xmlConfig.getAttribute( "jobId" ) ] )
-
-        self.debug( "the run config is %s" % config )
-
-        ## get pvalue limits
-        pvalues = xmlConfig.getElementsByTagName( "plimit" )
-        if ( len(pvalues) == 0 ):
-            self.warn("tag <plimit> not found, will use CLI pvalue limits FAIL=%f WARN=%f" % ( self.opts.pfail,
-                                                                                               self.opts.pwarn ) )
-        elif ( len(pvalues) > 1 ):
-            self.warn("too many <plimit> tags, will use CLI pvalue limits FAIL=%f WARN=%f" % ( self.opts.pfail,
-                                                                                               self.opts.pwarn ) )
-        else:
-            self.debug("tag <plimit> found, will user p-value limits from its attributes")
-            try:
-                pfail = float ( pvalues[0].getAttribute("fail") )
-                self.__update( "pfail", pfail )
-            except TypeError, value:
-                raise DCubeException( "pvalue limit for FAIL is NAN - %s" % str(value) )
-
-            try:
-                pwarn = float( pvalues[0].getAttribute( "warn" ) )
-                self.__update( "pwarn", pwarn )
-            except TypeError, value:
-                raise DCubeException( "pvalue limit for WARN is NAN - %s" % str(value) )
-
-
-        referenceNode = None
-        self.referenceURI = None
-        ## reference root file    
-        if ( self.opts.reference != None ):
-
-            self.referenceURI = os.path.abspath( self.opts.reference )
-            
-            self.debug("will use reference root file from CLI = %s" % self.referenceURI )
-            referenceNode = self.xmldoc.createElement( "reference" )
-            referenceNode.setAttribute( "file", self.referenceURI )
-
-            if ( xmlConfig.getElementsByTagName("reference") ):
-                xmlConfig.replaceChild( referenceNode, xmlConfig.getElementsByTagName("reference")[0] )
-            else:
-                xmlConfig.appendChild( referenceNode )
-
-        else:
-
-            reference = xmlConfig.getElementsByTagName( "reference" )
-            if ( len(reference) == 0 ): raise DCubeException("tag <reference> not found!")
-            elif ( len(reference) > 1 ): raise DCubeException("too many (>1) <reference> tags")
-            else:
-                referenceNode = reference[0]
-                self.referenceURI = os.path.abspath( referenceNode.getAttribute("file") )
-                self.__update( "reference", self.referenceURI )
-                    
-            self.debug("will use reference root file from XML = %s" % self.referenceURI )
-
-
-        # monitored root file
-        monitoredNode = self.xmldoc.createElement( "monitored" )
-        monitoredNode.setAttribute( "file", self.monitoredURI )
-
-        xmlConfig.insertBefore( monitoredNode, referenceNode  )
-
-        # date node
-        dateNode = self.xmldoc.createElement( "date" )
-        dateCDataNode = self.xmldoc.createTextNode( self.opts.isodate ) 
-
-        dateNode.appendChild( dateCDataNode )
-
-        xmlConfig.insertBefore( dateNode, monitoredNode )
-
-
-        # update xml output file
-        if ( self.opts.output == "" ):
-            self.__update("output", self.opts.monitored + ".dcube.xml" )
-
-
-        # get <tdirectory> tags and run tester
-        tdirs = xmlConfig.getElementsByTagName( "TDirectory" )    
-        if ( len(tdirs) == 0 ): raise DCubeException("no <TDirectory> tags found!")
-        else:
-            ## DCubeTester instance
-            self.tester = DCubeTester( self.xmldoc, self.parsed )
-            
-            self.refTFile = None
-            if ( self.exists( self.referenceURI ) ):
-                try:
-                    self.refTFile = ROOT.TFile.Open( self.referenceURI, "READ" )
-                except Exception, value:
-                    self.epanic( "cannot get handle for referece root file, all tests will be skipped, reason: %s" % str(value) )
-            else:
-                self.warn("reference root file not exists, all tests will be skipped!")
-
-            self.monTFile = None
-            if ( self.exists( self.monitoredURI ) ):
-                try:
-                    self.monTFile = ROOT.TFile.Open( self.monitoredURI , "READ")
-                except Exception, value:
-                    msg = "cannot get handle for monitored root file, reason: %s" % str(value) 
-                    raise DCubeException( msg )
-            else:
-                raise DCubeException( "monitored root file %s not exists!" % self.monitoredURI )
-                
-            # run DCubeTester recursively
-            self.test( tdirs[0], tdirs[0].getAttribute("name") ) 
-
-
-            if ( self.refTFile ): self.refTFile.Close()
-            if ( self.monTFile ): self.monTFile.Close()
-
-            # print and append summary node 
-            xmlConfig.appendChild (  self.tester.summary() )
-            
-        
-        # save output XML to file
-        return self.__save( )
-    
-    
-    ## get CDATA section
-    # @param self "Me, myself and Irene"
-    # @param nodelist list of XML DOM Node objects
-    def __cdata( self, nodelist ):
-        out = ""
-        for node in nodelist:
-            if node.nodeType == node.TEXT_NODE:
-                out = out + node.data
-        return out 
-        
-
-    ## main worker here
-    # @param self "Me, myself and Irene"
-    # @param node xml <tdirectory> node
-    # @param path full path inside both root files
-    def test( self, node, path ):
-
-        self.debug("now testing objects in path = %s" % path )
-
-        status = [ ]
-        for child in node.childNodes:
-
-            if ( child.nodeType == node.ELEMENT_NODE ):
-                self.debug("child tagName %s" % child.tagName )
-                tagName = child.tagName
-                if ( tagName == "TDirectory" ):
-                    status.append ( self.test( child, os.path.join( path , child.getAttribute("name")  ) ) )
-                    
-                    for test in [ "KS", "chi2", "bbb" ]:
-                        for stat in [ "OK", "WARN", "FAIL" ]:
-                            attr = test + stat
-                                
-                            parentCounter = node.getAttribute( attr )
-                            if ( not parentCounter ): parentCounter = "0"
-                            
-                            nodeCounter = child.getAttribute( attr )
-                            if ( not nodeCounter ): nodeCounter = "0"
-                        
-                            node.setAttribute( attr, "%d" % ( int(parentCounter) + int(nodeCounter) )  )
-                    
-                else:
-
-                    if ( child.hasAttribute("name") ):
-                        objName = child.getAttribute( "name" )
-                        objTests = child.getAttribute( "tests" ).split(",")
-                        self.debug( "found type=%s name=%s tests=%s in path %s" % ( tagName, objName, str(objTests), path ) )
-                        
-                        if ( path == "/" ):
-                            objPath = objName
-                        else:
-                            objPath = os.path.join( path, objName )
-                        
-                        refObj = self.refTFile.Get( objPath )
-                        monObj = self.monTFile.Get( objPath )
-                    
-                        status.append( self.tester.test( child, monObj, refObj ) )
-                    else:
-                        self.warn( "empty attribute 'name' found for tag <%s> in path %s, skipping!" % ( tagName, path ) )
-                    
-        statusAttr = "OK"    
-        if ( ( "FAIL" in status ) or
-             ( None in status ) ):
-            statusAttr = "FAIL"    
-        elif ( "WARN" in status ):
-            statsuAttr = "WARN"
-
-        self.debug("TDirectory %s status %s" % ( path, statusAttr ) )
-        node.setAttribute( "status", statusAttr )
-
-        return statusAttr 
-        
-        
-    ## update parsed options from DCubeOptParser
-    # @param self "Me, myself and Irene"
-    # @param key option name
-    # @param what new value
-    def __update(  self, key, what ):
-        self.debug("updating option %s to %s" % ( str(key), str(what) ) )
-        self.opts._update_loose( { key : what } )
-        
-
-    ## dump left arguments to logger
-    # @param self "Me, myself and Irene"
-    def __dumpArgs( self ):
-        if ( len(self.args) ):
-            self.info("dumping not parsed arguments...")
-            i = 1
-            for item in self.args:
-                self.info("[%02d] %s" % ( i, item ) )
-                i += 1
-
-    ## dump parsed options to logger
-    # @param self "Me, myself and Irene"
-    def __dumpOptions( self ):
-        self.info("dumping parsed options...")
-        i = 1
-        for k, v in self.opts.__dict__.iteritems():
-            if ( v == "" ): v = "not set"
-            self.info("[%02d] %s %s" % ( i, k, v))
-            i += 1
-        
-    ## save new output XML to file
-    # @param self "Me, myself and Irene"
-    def __save( self ):
-        
-        outputURI = os.path.abspath( self.opts.output )
-        self.__backup( outputURI )
-        try:
-          fileXML = open( outputURI, "w+")
-          self.xmldoc.normalize()
-      
-          xmlstring =  self.xmldoc.toprettyxml(" ", "\n").split("\n")
-          for line in xmlstring:
-              if line.strip("\n").strip() != "":
-                  fileXML.write( line+"\n" )
-          fileXML.close()
-          self.info("DCube output has been saved to file %s" % outputURI )
-        except IOError, value:
-            msg = "creation of XML output file %s failed - %s" % ( outputURI, str(value) )
-            self.epanic( msg )
-            return False
-
-        return True
-
-    ## create backup for old XML output file file
-    # @param self "Me, myself and Irene"
-    # @param file path to file
-    def __backup( self, file ):
-
-        fileURI = os.path.abspath(file)
-        self.debug( "will create backup file for %s" % file)
-
-        if ( os.path.exists( fileURI ) and os.path.isfile(fileURI) ):
-            self.warn("file %s exists and will be overwritten!" % fileURI )
-            backupURI = fileURI + ".bak"
-            self.debug("will create backup file %s" % backupURI )
-            if ( os.path.isfile( backupURI ) ):
-                self.warn("backup file %s exists and will be overwritten!" % backupURI )
-            try:
-                os.rename( fileURI, backupURI )
-                self.info("backup file %s from file %s has been created" % ( backupURI, fileURI ) )
-            except OSError, value:
-                msg = "creation of backup file %s failed - %s" % ( backupURI, str(value) )
-                self.epanic( msg )
-                    
-    ## check existence of file 
-    # @param self "Me, myself and Irene"
-    # @param filename a file name to check
-    def exists( self, filename ):
-        if ( "://" not in filename ):
-            return os.path.exists( os.path.abspath( str(filename) ) )
-        self.warn("can't check exsitence of file at URI %s" % filename )
-        return True
-
-    ## produce PHP files
-    # @param self "Me, myself and Irene"
-    # @return boolean status 
-    def __producePHPs( self ):
-        self.info("will produce PHP files...")
-
-        if ( not self.opts.server ):
-            self.error("path to DCubeServer not set, PHP files won't be produced")
-            return False
-        
-        if ( None not in ( self.opts.log, self.opts.config, self.opts.output ) ):
-            phpWriter = DCubePHPWriter( self.parsed )
-
-            phpURI = os.path.abspath( self.opts.output + ".php" )
-            logURI = os.path.abspath( os.path.join( os.path.dirname(self.opts.output), "dcubelog.php" ) )
-            
-            self.debug( "out php URI %s" % phpURI )
-            self.debug( "log php URI %s" % logURI )
-            
-            phpOK = False
-            try:
-                filePHP = open( phpURI, "w+" )
-                filePHP.writelines( phpWriter.dcubePHP() )
-                filePHP.close()
-                self.debug("out php file has been created")
-                phpOK = True
-            except Exception, value:
-                self.error("can't create output PHP file %s, reason - %s" % ( phpURI, str(value) ) )
-
-            logOK = False
-            try:
-                filePHP = open( logURI, "w+" )
-                filePHP.writelines( phpWriter.dcubeLOG() )
-                filePHP.close()
-                self.debug("log php file has been created")
-                logOK = True
-            except Exception, value:
-                self.error("can't create log PHP file %s, reason - %s" % ( logURI, str(value) ) )
-
-            return ( phpOK and logOK )
-
-
-    ## main comment in output XML file 
-    # @param self "Me, myself and Irene"
-    # @return a comment string
-    def __mainXMLComment( self ):
-        pass
-
-    
-## 
-# @class test_DCubeApp 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubeApp
-# @todo cover more!
-class test_DCubeApp( unittest.TestCase ):
-
-    ## setup test suite
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-        self.sysArgv = [ "--branch", "dev",
-                         "--install", "build",
-                         "--cmtconfig", "i686-slc4-gcc34-opt",
-                         "--project", "AtlasProduction",
-                         "--jobId", "job_1",
-                         "--server", "/afs/cern.ch/user/c/cibak/scratch1/dcube/dcube.php",
-                         "monitored.root" ]
-        
-        self.sysExitHolder = sys.exit
-        def exitCode( value ):
-            print "sys.exit called with value %s" % str(value)    
-        sys.exit = exitCode
-
-    ## test contructor
-    # @param self "Me, myself and Irene"
-    def test_01_ctor( self ):
-        sys.argv = self.sysArgv
-        theApp = DCubeApp( )
-        del theApp
-
-    ## test public API
-    # @param self "Me, myself and Irene"
-    def test_02_api( self ):
-        sys.argv = self.sysArgv
-        theApp = DCubeApp( )
-        del theApp
-
-    
-## test suite execution
-if __name__ == "__main__":
-    
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubeApp)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-    
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeConfig.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeConfig.py
deleted file mode 100644
index 1fe012177a2a3eb318deb23e790cb22a559174ea..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeConfig.py
+++ /dev/null
@@ -1,634 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubeConfig.py
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief implementation of DCubeConfig and test_DCubeConfig classes
-
-import os
-import re
-import DCubeUtils
-from DCubeOptParser import DCubeOptParser
-import unittest
-import ROOT
-import xml.parsers.expat
-import xml.dom.minidom
-from time import strftime, localtime, strptime
-
-
-## 
-# @class DCubeConfig
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief DCube configuration holder
-class DCubeConfig( DCubeUtils.DCubeObject ):
-
-    ## forcing to be a singleton
-    __metaclass__ = DCubeUtils.DCubeSingleton
-
-    ## configuration dictionary
-    __config = { }
-
-
-    ## flag to trigger transformation from old format 
-    needTransform = False
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param parsed arguments and options from DCubeOptParser 
-    def __init__(self, parsed):
-
-        super(DCubeConfig, self).__init__( self )
-        self.info("constructing DCubeConfig...")
-        
-        self.opts, self.args = parsed
-        
-        if ( self.opts.config != "" ):
-            self.xmldoc = self.__readXML() 
-        
-
-    ## branch name getter
-    # @param self "Me, myself and Irene"
-    def branch( self ):
-        return self.opts.branch
-
-    ## install name getter
-    # @param self "Me, myslf and Irene"
-    def install( self ):
-        return self.opts.install
-
-    ## cmtconfig name getter
-    # @param self "Me, myself and Irene"
-    def cmtconfig( self ):
-        return self.opts.cmtconfig
-
-    ## project name getter
-    # @param self "Me, myself and Irene"
-    def project( self ):
-        return self.opts.project
-
-    ## job id getter
-    # @param self "Me, myself and Irene"
-    def jobId( self ):
-        return self.opts.jobId
-
-    ## remove white spaces nodes from XML DOM tree
-    # @param self "Me, myself and Irene"
-    # @param node xml.dom.Node object
-    def __removeWS( self, node ):
-
-        toRemove = []
-        for child in node.childNodes :
-            
-            if ( ( child.nodeType is xml.dom.Node.TEXT_NODE ) and 
-                 not child.data.strip() ):
-                toRemove.append( child )
-            elif ( child.hasChildNodes() ):
-                self.__removeWS( child )
-        for node in toRemove:
-            self.debug( "removing node %s" % str(node) )
-            node.unlink()
-                     
-
-    ## save new config to file
-    # @param self "Me, myself and Irene"
-    def __save( self ):
-        
-        configURI = self.opts.config
-        self.__backup( configURI )
-        try:
-          fileXML = open( configURI, "w+")
-          self.xmldoc.normalize()
-      
-          xmlstring =  self.xmldoc.toprettyxml(" ", "\n").split("\n")
-          for line in xmlstring:
-              if line.strip("\n").strip() != "":
-                  fileXML.write( line+"\n" )
-          fileXML.close()
-          self.info("DCubeConfig has been saved to file %s" % configURI)
-        except IOError, value:
-            msg = "creation of XML file %s failed - %s" % ( configURI, str(value) )
-            self.epanic( msg )
-
-        return True
-
-    ## create backup for old config file
-    # @param self "Me, myself and Irene"
-    # @param file path to file
-    def __backup( self, file ):
-
-        fileURI = os.path.abspath(file)
-        self.debug( "will create backup file for %s" % file)
-
-        if ( os.path.exists( fileURI ) and os.path.isfile(fileURI) ):
-            self.warn("file %s exists and will be overwritten!" % fileURI )
-            backupURI = fileURI + ".bak"
-            self.debug("will create backup file %s" % backupURI )
-            if ( os.path.isfile( backupURI ) ):
-                self.warn("backup file %s exists and will be overwritten!" % backupURI )
-            try:
-                os.rename( fileURI, backupURI )
-                self.info("backup file %s from file %s has been created" % ( backupURI, fileURI ) )
-            except OSError, value:
-                msg = "creation of backup file %s failed - %s" % ( backupURI, str(value) )
-                self.epanic( msg )
-        
-                 
-    ## (re)genaration of DCube xml configuration file
-    # @param self "Me, myself and Irene"
-    def generate( self ):
-
-        self.error( self.opts.config )
-        configURI = os.path.abspath( self.opts.config )
-        
-        fromScratch = False
-
-        msg = "configuration file %s" % configURI
-        if ( os.path.exists( configURI ) and 
-             os.path.isfile( configURI ) ):
-            msg += " exists"
-            if ( os.path.getsize( configURI ) ):
-                msg += " and will be overwritten"
-            else:
-                msg += ", but is empty, will create configuration from scratch"
-                fromScratch = True
-        else:
-            msg += " doesn't exist, will create it from scratch..."
-            fromScratch = True
-
-        self.info( msg )
-
-        if ( fromScratch ):
-            self.debug( "will create XML DOM document and store dcube config in file %s" % configURI )
-            self.xmldoc = self.__createXMLDOC()
-        else:
-            self.debug( "will ready dcube config from file %s" % configURI )
-            self.xmldoc = self.__readXML() 
-
-        self.refFileHandle = None
-        self.refFileURI = os.path.abspath( self.opts.reference )
-        self.debug("will try to open refernce root file=%s" % self.refFileURI )
-        try:
-            self.refFileHandle = ROOT.TFile.Open( self.refFileURI, "READ" )
-            self.debug("done")
-        except:
-            pass
-
-        if ( not self.refFileHandle ):
-            raise DCubeUtils.DCubeException("cannot open %s reference root file" % self.refFileURI )
-        
-        cliConfig = "/".join( [ self.opts.branch, self.opts.install,
-                      self.opts.cmtconfig, self.opts.project,
-                      self.opts.jobId ] )
-
-        self.info("will try to match CLI config %s to XML config" % cliConfig)
-        runConfigNode = self.getConfig( strict=True )
-        self.info("done!")
-        if ( not runConfigNode ):
-            self.warn("non matching config found in XML, will create it from scratch")
-            runConfigNode = self.__emptyConfig()
-        else:
-            self.warn("XML config has been found and will be overwritten" )
-
-        commentFound = False
-        for child in  runConfigNode.childNodes:
-            if ( child.nodeType == child.COMMENT_NODE ):
-                runConfigNode.replaceChild( self.__configComment(), child )
-                commentFound = True
-                break
-            
-        if ( not commentFound ):
-            runConfigNode.insertBefore( self.__configComment(), runConfigNode.firstChild  )
-
-        ## overwrite reference node
-        refNode = runConfigNode.getElementsByTagName("reference")
-        if ( len(refNode) == 0 ):
-            refNode = self.xmldoc.createElement("reference")
-            runConfigNode.appendChild( refNode )
-        else:
-            refNode = refNode[0]
-        refNode.setAttribute("file", os.path.abspath( self.opts.reference ) )
-
-        ## overwrite pvalue node
-        pValueNode = runConfigNode.getElementsByTagName("plimit")
-        if ( len(pValueNode) == 0 ):
-            pValueNode = self.xmldoc.createElement("plimit")
-            runConfigNode.appendChild( pValueNode )
-        else:
-            pValueNode = pValueNode[0]
-
-        pValueNode.setAttribute("fail", str( self.opts.pfail ) )
-        pValueNode.setAttribute("warn", str( self.opts.pwarn ) )
-        
-        ## overwrite top level tdirectory tag
-        rootDirNodes = runConfigNode.getElementsByTagName("TDirectory")
-
-        if ( len(rootDirNodes) != 0 ):
-            for rootDirNode in rootDirNodes:
-                if ( rootDirNode.parentNode == runConfigNode ): 
-                    runConfigNode.removeChild( rootDirNode )
-                
-        rootDirNode = self.xmldoc.createElement("TDirectory")
-        rootDirNode.setAttribute("name", "/")
-        runConfigNode.appendChild( rootDirNode ) 
-
-        self.info("will scan reference file for objects...")
-        self.__scan( self.refFileHandle, rootDirNode )       
-        self.info("done!")
-
-        msg = "generation of dcube config file='%s' " % os.path.abspath( self.opts.config )
-        if ( self.__save() ):
-            self.info( msg +  "completed" )
-            return True
-        
-        self.error(msg + "failed" % self.opts.config )
-        return False
-
-
-    ## match runtime config (from CLI) to offline XML config
-    # @param self "Me, myself and Irene"
-    # @param fromRun tuple from CLI
-    # @param fromXML tuple from XML
-    # @param strict flag to forve only str to str matching
-    def __match( self, fromRun, fromXML, strict=True ):
-
-        mode = "strict" if strict else "regexp" 
-        self.info("maching config in %s mode" % mode)
-
-        pattern = "[\S]+"
-        whats = ["branch", "install", "cmtconfig", "project", "jobId"]
-        cPairs = dict( zip (whats, zip( fromRun, fromXML ) ) )
-        cMatch = dict( zip(whats, [False*5]))
-
-        for k,v in cPairs.iteritems():
-            cliValue, xmlValue = v
-            msg = "what=%-10s fromCLI=%-23s fromXML=%-23s" % ( k , cliValue, xmlValue )
-            if ( cliValue == xmlValue ): 
-                msg += "match (str to str)" 
-                self.debug( msg )
-                cMatch[k] = True
-            else:
-                if ( not strict ):
-                    if ( "*" in cliValue and "*" not in xmlValue ):
-                        reCLI = re.compile( cliValue.replace("*", pattern, cliValue.count("*") ) )
-                        if ( reCLI.match( xmlValue ) ):
-                            msg += "match (reqexp to str)"
-                            self.debug( msg ) 
-                            cMatch[k] = True
-                            continue
-                    elif ("*" in xmlValue and "*" not in cliValue ):
-                        reXML = re.compile( xmlValue.replace("*", pattern, xmlValue.count("*") ) )
-                        if ( reXML.match( cliValue ) ):
-                            msg += "match (str to regexp)"
-                            self.debug( msg )
-                            cMatch[k] = True
-                    else:
-                        msg += "not match"
-                        self.debug( msg )
-                else:
-                    msg += "not match (strict mode on)"
-                    self.debug( msg )
-                    cMatch[k] = False
-                    
-        return all(cMatch.values())
-
-
-    ## get config XML node based on runtime configuration
-    # @param self "Me, myself and Irene"
-    # @param strict match config strings w/o regexp
-    def getConfig( self, strict=True ):
-        
-        runConfig = ( self.opts.branch,
-                      self.opts.install,
-                      self.opts.cmtconfig,
-                      self.opts.project,
-                      self.opts.jobId )
-        
-        self.info("will try to find CLI config='%s' in DCubeClient configuration file %s" % ( "/".join(runConfig), 
-                                                                                               os.path.abspath( self.opts.config) ) )
-
-       
-        configs = self.__config.keys()
-        configs.sort( reverse=True )
-        for config in configs:
-            if self.__match( runConfig, config, strict ):
-                self.info("runtime config='%s' match to XML config='%s'" % ( "/".join(runConfig), "/".join(config) ) )
-                return self.__config[ config ]
-        
-        self.warn("<dcube> DOM XML node for CLI config %s not found" % "/".join( runConfig ) )
-        return None
-
-    
-    def __oldConfig( self ):
-        pass
-
-
-    def __configComment( self ):
-        comNode = "\n"
-        comNode += "  " + "#"*51 + "\n"
-        comNode += "  DCubeClient for:\n"
-        comNode += "  [1] branch: %s\n" % self.opts.branch
-        comNode += "  [2] install: %s\n" % self.opts.install
-        comNode += "  [3] cmtconfig: %s\n" % self.opts.cmtconfig 
-        comNode += "  [4] project: %s\n" % self.opts.project
-        comNode += "  [5] jobId: %s\n" % self.opts.jobId
-        comNode += "  [6] reference root file: %s\n" % os.path.abspath( self.opts.reference )
-        comNode += "  [7] p-values limits FAIL = %f WARN = %f\n" % ( self.opts.pfail, self.opts.pwarn )
-        comNode += "  " + "#"*51 + "\n"
-        comNode = self.xmldoc.createComment( comNode )
-        return comNode
-     
-    ## prepare empty config XML DOM element with attributes from CLI passed thru DCubeOptParser
-    # @param self "Me, myself and Irene"
-    def __emptyConfig( self ):
-
-        self.info("will create empty <dcube> DOM XML node...")
-
-        runConfigNode = self.xmldoc.createElement("dcube")
-        runConfigNode.setAttribute( "branch", self.opts.branch )
-        runConfigNode.setAttribute( "install", self.opts.install )
-        runConfigNode.setAttribute( "cmtconfig", self.opts.cmtconfig )
-        runConfigNode.setAttribute( "project", self.opts.project )
-        runConfigNode.setAttribute( "jobId", self.opts.jobId )
-
-
-        referenceNode = self.xmldoc.createElement("reference")
-        referenceNode.setAttribute("file", os.path.abspath( self.opts.reference ) )
-
-        referenceDescNode = self.xmldoc.createElement("ref_desc")
-        refDescCData = self.xmldoc.createTextNode("PUT YOUR REFERENCE DESCRIPTION HERE")
-   
-        referenceDescNode.appendChild( refDescCData )
-
-        testDescNode = self.xmldoc.createElement("test_desc")
-        testDescCData = self.xmldoc.createTextNode("PUT YOUR TEST DESCRIPTION HERE")
-
-        testDescNode.appendChild( testDescCData )
-
-        pValueNode = self.xmldoc.createElement("plimit")
-        pValueNode.setAttribute("fail", str( self.opts.pfail ) )
-        pValueNode.setAttribute("warn", str( self.opts.pwarn ) )
-
-        ## append all nodes
-        runConfigNode.appendChild( referenceNode )
-        runConfigNode.appendChild( referenceDescNode )
-        runConfigNode.appendChild( testDescNode )
-
-        runConfigNode.appendChild( pValueNode )
-        
-        rootDirNode = self.xmldoc.createElement("TDirectory")
-        rootDirNode.setAttribute("name", "/")
-        
-        runConfigNode.appendChild( rootDirNode )
-        
-        self.xmldoc.getElementsByTagName("dcube_client_config")[0].appendChild( runConfigNode )
-
-        self.info("creation of empty <dcube> node completed")
-        return runConfigNode
-
-    
-    ## 
-    # @param self "Me, myself and Irene"
-    def __handleConfigNodes( self, parent ):
-
-        
-        runConfig = ( self.opts.branch,
-                      self.opts.install,
-                      self.opts.cmtconfig,
-                      self.opts.project,
-                      self.opts.jobId )
-        
-        configNodes = parent.getElementsByTagName("dcube")
-
-        self.debug("found %d <dcube> tags" % len( configNodes) )
-        if ( len(configNodes) ):
-            for configNode in configNodes:
-
-                nodeConfig = ( configNode.getAttribute("branch") if configNode.getAttribute("branch") != "" else "*",
-                               configNode.getAttribute("install") if configNode.getAttribute("install") != "" else "*",
-                               configNode.getAttribute("cmtconfig") if configNode.getAttribute("cmtconfig") != "" else "*",
-                               configNode.getAttribute("project") if configNode.getAttribute("project") != "" else "*",
-                               configNode.getAttribute("jobId") if configNode.getAttribute("jobId") != "" else "*" ) 
-               
-                self.info("found <dcube> tag for configuration %s/%s/%s/%s/%s" % nodeConfig )
-                self.__config[nodeConfig] = configNode
-        else:
-            self.warn("DCube config file is empty? no <config> tags found!")
-            
-
-
-
-    ## generate and return empty XML DOM document 
-    # @param self "Me, myself and Irene"
-    def __createXMLDOC( self ):
-        xmldoc = xml.dom.minidom.Document()
-        
-        commentNode = xmldoc.createComment( self.__mainXMLComment() )
-        xmldoc.appendChild( commentNode )
-
-        dcubeNode = xmldoc.createElement("dcube_client_config")      
-        xmldoc.appendChild( dcubeNode )
-   
-        return xmldoc
-    
-
-    ## produce scratch config comment string
-    # @param self "Me, myself and Irene"
-    def __mainXMLComment( self ):
-        
-        out  = "\n DCube configuration XML file = %s\n" % self.opts.config
-        out += " auto-generated using %s\n" %  DCubeUtils.DCubeVersion()
-        out += " on %s\n" %  strftime("%a, %d %b %Y %H:%M:%S %Z" , localtime())
-        out += " from Root reference file: %s\n" % str( self.opts.reference )
-        out += " Statistics tests:\n"
-        out += " %r\n" % repr( self.opts.tests )
-        out += " Legend:\n"
-        out += " * KS   - Kolmogorov-Smirnov test (default)\n"
-        out += " * chi2 - chi2 test\n"
-        out += " * bbb  - bin-by-bin comparision\n"
-        out += " * all  - all above tests\n\n"
-        out += " Tweaks:\n"
-        out += " [1] remove or comment out (using HTML comments tags) histograms not relevant\n"
-        out += "     to your analysis\n"
-        out += " [2] provide a new content to \"ref_desc\" and \"mon_desc\" tags to something\n"
-        out += "     more meaningful\n"
-        out += " [3] and of course feel free to change attribute \"tests\" in each \"hist\" tag\n"
-        out += " [4] limits of p-value warning and failure can be set by changing attributes\n"
-        out += "     in <plimit> tag\n\n"
-        out += " [5] if you want to make plots FOR EVERY histogram in your monitored file\n"
-        out += "     change attribute \"inclusive\" in <mode> tag to \"true\"\n\n"
-        return out
-
-
-
-    ## read XML configuration from 
-    # @param self "Me, myself and Irene"
-    def __readXML( self ):
-
-        configURI = os.path.abspath( self.opts.config )
-        self.debug("will try to read and parse configuration file %s" % configURI )
-
-        if ( os.path.exists( configURI ) and 
-             os.path.isfile( configURI ) and 
-             os.path.getsize( configURI ) ):
-            try:
-                xmldoc = xml.dom.minidom.parse( configURI )    
-            except xml.parsers.expat.ExpatError, value:
-                self.panic("configuration file parsing failed, %s" % str( value ) )
-                return False
-            except xml.dom.DOMException, value:
-                self.panic("configuration file parsing failed, %s" % str( value ) )
-                return False
-
-
-            if ( xmldoc ):
-                self.debug("will try to extract <config> tags...")
-                self.__handleConfigNodes( xmldoc )
-                self.debug("done!")
-                return xmldoc
-            
-
-    ## scan reference root file and dump its contents to dom xml nodes  
-    # @param self "Me, myself and Irene"
-    # @param target ROOT.TDirectory to scan
-    # @param configNode XML DOM Element 
-    # @param level level of recursion to pretty print
-    def __scan( self, target, configNode, level=1 ):
-            
-        nbsp = " " * level 
-        target.cd()
-        path = target.GetPath().split(":")[1]
-        curdir = ROOT.gDirectory
-        nextkey = ROOT.TIter( curdir.GetListOfKeys() ) 
-
-        for ik in range( target.GetNkeys() ):
-            
-            key = nextkey.Next()
-
-            obj = key.ReadObj()
-            name = obj.GetName()
-            cl  = obj.Class().GetName()
-            isa = obj.IsA()
-            fp = None
-
-            fp = os.path.join( path, name)
-    
-            self.debug( nbsp + "--> found " + name + " of type " + cl ) 
-
-            if ( isa.InheritsFrom("TH1") ):
-
-                objDim = obj.GetDimension()
-                self.debug( nbsp + " --> found object '%s' at location '%s' of name '%s'" % ( cl, path, name) )
-
-                if ( objDim <= 2 ):
-                    objNode = self.xmldoc.createElement( "hist%dD" % objDim  )
-                    objNode.setAttribute( "name", name )
-                    objNode.setAttribute( "type", cl )
-                    objNode.setAttribute( "tests", ",".join(self.opts.tests)  )          
-                    configNode.appendChild( objNode )
-                else:
-                    self.warn("dimension > 2 for histogram '%s', skipping" % fp)
-
-            elif ( isa.InheritsFrom("TGraph") ):
-
-                self.debug( nbsp + " --> found object '%s' at location '%s' of name '%s'" % ( cl, path, name) )
-
-                objNode = self.xmldoc.createElement( "graph" )
-                objNode.setAttribute( "name", name )
-                objNode.setAttribute( "type", cl )
-                configNode.appendChild( objNode )
-                
-            #elif ( isa.InheritsFrom("TGraph2D") ):
-            #    self.debug( nbsp + " --> found object %s at location %s of name %s" % ( cl, path, name) )
-
-            #    objNode = self.xmldoc.createElement( cl )
-            #    objNode.setAttribute( "name", name )
-            #    configNode.appendChild( objNode )
-               
-            elif ( isa.InheritsFrom("TDirectory") ):
-                self.debug( nbsp +" --> found TDirectory=" + fp )
-                
-                dirNode = self.xmldoc.createElement("TDirectory")
-                dirNode.setAttribute( "name", name )
-                
-                configNode.appendChild( dirNode )
-                
-                self.__scan( obj, dirNode, level+1 )
-           
-            else:
-                self.warn( nbsp +" --> unsupported object of type '%s' found at '%s'" % ( cl, fp ) )  
-                pass
-   
-        return configNode       
-
-            
-    ## 
-    # @param self "Me, myself and Irene"
-    def __DOMXML( self ):
-        return self.xmldoc
-  
-##
-# @class test_DCubeConfig
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubeConfig class
-class test_DCubeConfig( unittest.TestCase ):
-    
-
-    ## test suite setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-
-        import sys
-        def sysExit( value ):
-            pass
-        sys.exit = sysExit
-        
-        DCubeUtils.DCubeLogger("test_DCubeConfig.log", "test_DCubeConfig", True )
-        self.log =  DCubeUtils.DCubeLogger.getLogger( self )
-
-        self.info = self.log.info 
-        
-        self.args = [ __file__,
-                      "-c",  "dcube_config.xml",
-                      "-r" , "reference.root",
-                      "-t", "bbb", "KS",
-                      "--branch", "dev",
-                      "--project", "AtlasProduction" ]
-        self.parser = DCubeOptParser()
-        self.parsed =  self.parser.parse( self.args )
-
-
-
-        self.config = DCubeConfig( self.parsed )
-        
-
-    ## test constructor
-    # @param self "Me, myself and Irene"
-    def test_01_constructor( self ):
-        try:
-            self.config = DCubeConfig( self.parsed )
-        except:
-            pass
-        self.assertEqual( isinstance( self.config, DCubeConfig), True )
-
-    ## test generate()
-    # @param self "Me, myself nad Irene"    
-    def test_02_generate( self ):       
-        self.config.generate()
-
-    ## test get()
-    # @param self "Me, myself and Irene"
-    def test_03_get( self ):
-        configNode = self.config.getConfig( strict = False )
-        print configNode
-        
-
-
-
-if __name__ == "__main__":
-        
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubeConfig)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-
-
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml
deleted file mode 100644
index d6fa6abf357a573344636e6038df2cb9c3cedf24..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml
+++ /dev/null
@@ -1,28 +0,0 @@
-<!--
-   file: DCubeForRTTCfg.xml 
-   brief: DCubeClient for AtlasRTT ini file 
-   author: Krzysztof Daniel Ciba
-   date: Sep 4, 2007
-
-   * Oct 23, 2007 - cibak - switch to root 5.17/04
-   * Apr 10, 2008 - cibak - switch to root 5.19/02, python 2.5, DCubeServer-00-00-10
-                            
--->
-<init version="3.1.4.1.5">
-
-<!-- environment presettings -->
-<!-- version of binaries to use, by default slc4_amd64_gcc34 --> 
-<bin>slc4_amd64_gcc34</bin>
-<!-- ROOT -->
-<root version="5.19.02">/afs/cern.ch/sw/lcg/external/root</root>
-<!-- PYTHON -->
-<python version="2.5">/afs/cern.ch/sw/lcg/external/Python</python>
-
-<!-- runtime parameters -->
-<!-- base directory with reference files -->
-<ref_dir>/afs/cern.ch/atlas/project/RTT/Results/reference/</ref_dir>
-<!-- base directory with dcube server installation --> 
-<server>/afs/cern.ch/atlas/project/RTT/Results/dcube/DCubeServer-00-00-13/</server>
-
-</init>
-<!-- end of dcube_for_rtt.xml file -->
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeOptParser.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeOptParser.py
deleted file mode 100644
index 7ef0b604cdd85cee93ee0e618c3af99cdba1d6de..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeOptParser.py
+++ /dev/null
@@ -1,622 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubeOptParser.py
-# @author Krzysztof Danile Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief implementation of DCubeOptParser
-
-import sys
-import optparse
-from DCubeUtils import DCubeObject, DCubeLogger, DCubeSingleton, DCubeVersion
-from DCubeValid import DCubeValid
-import unittest
-
-
-##
-# @class DCubeOptParser
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief sys.argv parser for DCube
-class DCubeOptParser( object ):
-
-    ## forcing to be a singleton
-    __metaclass__ = DCubeSingleton
- 
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param opts options to parse
-    # @param args args to parse 
-    def __init__( self  ):
-
-
-        self.valid = DCubeValid()
-
-        today = "today"
-        try:
-            from datetime import date
-            today = date.today().isoformat()
-        except: 
-            pass
-
-        usage = "dcube.py [options] [args] [MONITORED.root]"
-
-        self.__version = DCubeVersion()
-        self.__par = optparse.OptionParser( usage=usage, version=self.__version.version() )
-
-        self.__par.add_option( "-r",
-                               "--reference", 
-                               action="store", 
-                               type="string", 
-                               dest="reference",
-                               metavar="REFERENCE",
-                               help="reference root file" )
-
-        self.__par.add_option( "-x", 
-                               "--xml",
-                               action="store",
-                               dest="output",
-                               metavar="OUTPUTXML",
-                               help="XML output file" )
-
-        self.__par.add_option( "-c", 
-                               "--config",
-                               type="string",
-                               dest="config",
-                               metavar="CONFIGXML",
-                               help="XML configuration file" )
-
-        self.__par.add_option( "-l", 
-                               "--logname",
-                               type="string",
-                               dest="log",
-                               metavar="LOGNAME",
-                               help="log file [default %default]" )
-
-        self.__par.add_option( "-t",
-                               "--tests",
-                               action="callback",
-                               callback=self.tests_callback,
-                               dest="tests",
-                               help="statistics test to perform [default %default]")
-
-        self.__par.add_option( "-f", 
-                               "--pfail",
-                               action="callback",
-                               type="float",
-                               callback=self.check_pfail_callback,
-                               dest="pfail",
-                               help="p-value limit for test failure [default %default]" )
-
-        self.__par.add_option( "-w", 
-                               "--pwarn",
-                               type="float", 
-                               action="callback",
-                               callback=self.check_pwarn_callback,
-                               dest="pwarn",
-                               help="p-value limit for test warning [default %default]" )
-
-        self.__par.add_option( "-d", 
-                               "--date", 
-                               action="callback",
-                               type="string",
-                               callback=self.isodate_callback,
-                               dest="isodate",
-                               metavar="YYYY-MM-DD",
-                               help="date of run in ISO format [default %default]")
-
-        self.__par.add_option( "-s", 
-                               "--server",
-                               type="string",
-                               dest="server",
-                               metavar="DCUBESERVER",
-                               help="path to DCubeServer")
-
-        self.__par.add_option( "--branch",
-                               type="string",
-                               action="callback",
-                               callback=self.branch_callback,
-                               dest="branch",
-                               help="branch name [default %default]" )
-
-        self.__par.add_option( "--cmtconfig",
-                               type="string",
-                               action="callback",
-                               callback=self.cmtconfig_callback,
-                               dest="cmtconfig",
-                               help="CMTCONFIG name [default %default]" )
-        
-        self.__par.add_option( "--install",
-                               type="string",
-                               action="callback",
-                               callback=self.install_callback,
-                               dest="install",
-                               help="install name [default %default]" )
-
-        self.__par.add_option( "--project",
-                               type="string",
-                               action="callback",
-                               callback=self.project_callback,
-                               dest="project",
-                               help="project name [default %default]" )
-
-        self.__par.add_option( "--jobId",
-                               type="string",
-                               dest="jobId",
-                               help="job identification name [default %default]")
-
-        self.__par.add_option( "-g",
-                               "--generate",
-                               action="store_true",
-                               dest="generate",
-                               help="trigger for configuration file generation [default off]")
-        
-        self.__par.add_option( "-p", 
-                               "--plots",
-                               action="store_true",
-                               dest="makeplots",
-                               help="trigger for histograms plots generation [default off]")
-
-
-        self.__par.add_option( "-b",
-                               action="store_true",
-                               dest="batch",
-                               help="batch mode for PyROOT [always on]")
-
-    
-        self.__par.set_defaults( pwarn=0.95,
-                                 pfail=0.75, 
-                                 config="dcube_config.xml",
-                                 output="",
-                                 log="./dcube.log",
-                                 tests=["KS"],
-                                 isodate=today,
-                                 server="",
-                                 branch="*",
-                                 cmtconfig="*",
-                                 install="*",
-                                 project="*",
-                                 jobId="*",
-                                 generate=False,
-                                 makeplots=False,
-                                 root2null=False)
-
-
-
-    ## prints out help
-    # @param self "Me, myself and Irene"
-    def help( self ):
-        self.__par.print_help()
-        print self.valid
-
-    ## prints out usage
-    # @param self "Me, myself and Irene"
-    def usage( self ):
-        self.__par.print_usage()
-
-    ## prints out version
-    # @param self "Me, myself and Irene"
-    def version( self ):
-        self.__par.print_version()
-
-    ## parses arguments
-    # @param self "Me, myself and Irene"
-    # @param args list of arguments to parse (default sys.argv[1:])
-    # @return tuple ( argumenst, optparse.Values )
-    def parse( self, args = sys.argv[1:] ):
-
-        args, opts = self.__par.parse_args( args )
-        
-        #if ( args.generate and not args.reference ):
-        #    print "can't generate DCube config, reference root file not set (use -r, --reference)"
-        #    self.usage()
-        #    sys.exit(0)
-        #if ( ( args.config and not args.generate ) and len(opts) < 2 ):
-        #    self.usage()
-        #    sys.exit(0)
-        
-        return ( args, opts )
-        
-
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @param option option name
-    # @param opt_str option long name
-    # @param value returned value
-    # @param parser OptParser instance
-    def check_generation_callback( self, option , opt_str, value, parser ):
-
-        print parser.values.reference
-        if ( not parser.values.reference ):
-            pass
-
-    ## check pwarn value
-    # @param self "Me, myself and Irene"
-    # @param option option name
-    # @param opt_str option long name
-    # @param value returned value
-    # @param parser OptParser instance
-    def check_pwarn_callback( self, option, opt_str, value, parser ):
-
-        if ( value > 1.0 or value < 0.0 ):
-            raise optparse.OptionValueError("%s value %f out of range (0, 1)" % (opt_str, value ) )
-        
-        if ( ( "-f", "--pfail" ) not in parser.rargs ):
-            if ( value <= parser.values.pfail ):
-                raise optparse.OptionValueError("value for %s <= p-value failure (%f <= %f)" % ( opt_str, 
-                                                                                                 value,
-                                                                                                 parser.values.pfail ) )
-
-        setattr( parser.values, option.dest, value )
- 
-    ## callback for -f and -pfail options
-    # @param self "Me, myself and Irene"
-    # @param option option name
-    # @param opt_str option long name
-    # @param value returned value
-    # @param parser OptParser instance
-    def check_pfail_callback( self, option, opt_str, value, parser ):
-      
-        if ( value > 1.0 or value < 0.0 ):
-            raise optparse.OptionValueError("%s value %f out of range (0, 1)" % (opt_str, value ) )
-
-        if ( ( "-w", "--pwarn" ) not in parser.rargs ):
-            if ( value >= parser.values.pwarn ):
-                raise optparse.OptionValueError("value for %s >= p-value warning (%f >= %f)" % ( opt_str,
-                                                                                                 value, 
-                                                                                                 parser.values.pwarn ) )
-
-        setattr( parser.values, option.dest, value )
-
-    ## callback for "-d" and "--date" options
-    # @param self "Me, myself and Irene"
-    # @param option option name
-    # @param opt_str option long name
-    # @param value returned value
-    # @param parser OptParser instance 
-    def isodate_callback( self, option, opt_str, value, parser ):
-        info = ( opt_str, value )
-        
-        if ( ( "-" not in value ) or 
-             ( value.count("-") != 2 ) ): 
-            raise optparse.OptionValueError("%s option value (=%s) error, not in ISO date format (YYYY-MM-DD)!" % info ) 
-        
-        year, month, day = value.split("-")
-
-        try:
-            year = int(year)
-        except ValueError, value:
-            raise optparse.OptionValueError("%s option value (=%s) error, year isn't an integer!" % info )
-            
-        try:
-            month = int(month)
-        except ValueError, value:
-            raise optparse.OptionValueError("%s option value (=%s) error, month isn't an integer!" % info )
-    
-        try:
-            day = int(day)
-        except ValueError, value:
-            raise optparse.OptionValueError("%s option value (=%s) error, day isn't an integer!" % info )
-        
-        try:
-            from datetime import datetime
-            datetime( year, month, day, 0, 0, 0  )
-        except ValueError, val:
-            raise optparse.OptionValueError("%s option value (=%s) error, %s" % ( opt_str, value, val ) )
-            
-        setattr( parser.values, option.dest, value )
-
-    ## callback for parsing "-t" and "--tests" options
-    # @param self "Me, myself and Irene"
-    # @param option option name
-    # @param opt_str option long name
-    # @param value option value
-    # @param parser OptParser instance
-    def tests_callback( self, option, opt_str, value, parser ):
-        assert value is None
-        done = 0
-        value = []
-        rargs = parser.rargs
-        while rargs:
-            arg = rargs[0]
-            if ( (arg[:2] == "--" and len(arg) > 2) or
-                 (arg[:1] == "-" and len(arg) > 1 and arg[1] != "-") ):
-                break
-            else:
-                if ( arg in ("KS", "chi2", "bbb", "all" ) ):
-                    value.append(arg)
-                    del rargs[0]
-                else:
-                    raise optparse.OptionValueError("unknown statistics test value '%s'" % arg ) 
-
-        if ( "all" in value ): value = [ "KS", "bbb", "chi2" ]
-        assert len(value) != 0
-        setattr( parser.values, option.dest, value )
-
-
-    ## callback for parsing --branch option
-    # @param self "Me, myself and Irene"
-    # @param option option string
-    # @param opt_str long option name
-    # @param value option value
-    # @param parser OptParser instance
-    def branch_callback(  self, option, opt_str, value, parser ):
-        if ( "*" not in value ):
-            found = False
-            for k, v in self.valid.branches().iteritems():
-                if ( v.match( value ) ): 
-                    found = True
-            if ( not found ):
-                print self.valid.listBranches()
-                raise optparse.OptionValueError("invalid %s option value '%s'" % (option, value ) ) 
-        setattr( parser.values, option.dest, value )
-        
-
-    ## callback for parsing --install option
-    # @param self "Me, myself and Irene"
-    # @param option option string
-    # @param opt_str long option name
-    # @param value option value
-    # @param parser OptParser instance
-    def install_callback(  self, option, opt_str, value, parser ):    
-        if ( "*" not in value ):
-            if ( value not in self.valid.installs() ):
-                print self.valid.listInstalls()
-                raise optparse.OptionValueError("invalid %s option value '%s'" % (option, value) ) 
-        setattr( parser.values, option.dest, value )
-        
-
-    ## callback for parsing --cmtconfig option
-    # @param self "Me, myself and Irene"
-    # @param option option string
-    # @param opt_str long option name
-    # @param value option value
-    # @param parser OptParser instance
-    def cmtconfig_callback(  self, option, opt_str, value, parser ):    
-        if ( "*" not in value ):
-            if ( value not in self.valid.cmtconfigs() ):
-                print self.valid.listCmtconfigs()
-                raise optparse.OptionValueError("invalid %s option value '%s'" % (option, value)  ) 
-        setattr( parser.values, option.dest, value )
-        
-    ## callback for parsing --project option
-    # @param self "Me, myself and Irene"
-    # @param option option string
-    # @param opt_str long option name
-    # @param value option value
-    # @param parser OptParser instance
-    def project_callback(  self, option, opt_str, value, parser ):    
-        if ( "*" not in value ):
-            if ( value not in self.valid.projects() ):
-                print self.valid.listProjects()
-                raise optparse.OptionValueError("invalid %s option value '%s'" % (option, value)  ) 
-        setattr( parser.values, option.dest, value )
-   
-        
-
-##
-# @class test_DCubeOptParse
-# @author Krzysztof Daniel Ciba (Krzsyztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubeOptParser
-class test_DCubeOptParser( unittest.TestCase ):
-
-    ## test suite setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-
-        self.logger = DCubeLogger( "test_DCubeOptParser.log", logName="test_DCubeParser", toConsole=True )
-
-        opts = [ ]
-        args = [ __file__, "-h" ]
-
-        self.opt = DCubeOptParser( )
-
-        self.log = self.logger.getLogger(self)
-        
-        self.info = self.logger.info
-       
-     
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    def test_01_constructor( self ):
-        try:
-            self.opt = DCubeOptParser() 
-        except:
-            pass
-
-        self.assertEqual( isinstance( self.opt, DCubeOptParser), True )
-
-    ## usage(), version(), help()
-    # @param self "Me, myself and Irene"
-    def test_02_printouts( self ):
-        self.opt.version()
-        self.opt.usage()
-        self.opt.help()
-
-    ## parse()
-    # @param self "Me, myself and Irene"
-    def test_03_parse( self ):
-     
-
-        self.info( "sys.argv[1:] = %s" % str( sys.argv[1:] ) ) 
-        args, opts = self.opt.parse()
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-     
-        sysExitHolder = sys.exit
-        def sysExit( value ):
-            print "sys.exit called with value = %s" % str(value)
-        sys.exit = sysExit
-        
-        args = [ __file__, "-h"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-
-        args = [ __file__, "--version"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-
-
-        args = [ __file__, "-g"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.generate = %s " % args.generate )
-
-
-        args = [ __file__, "-c", "configuration.xml"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.config = %s " % args.config )
-
-        args = [ __file__, "-x", "output.xml"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.output = %s " % args.output )
-
-        args = [ __file__, "-r", "reference.xml"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.reference = %s " % args.reference )
-
-        args = [ __file__, "-d", "2008-12-24"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.isodate = %s " % args.isodate )
-
-        args = [ __file__, "-t", "KS", "bbb", "chi2"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.tests = %s " % str(args.tests) )
-
-        args = [ __file__, "-t", "all"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.tests = %s " % str(args.tests) )
-
-        args = [ __file__, "-t", "KS", "bbb"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.tests = %s " % str(args.tests) )
-
-        args = [ __file__, "--date", "2008-02-31"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-
-        args = [ __file__, "-w", "0.85"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pwarn = %s " % str(args.pwarn) )
-
-        args = [ __file__, "-f", "0.65"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pfail = %s " % str(args.pfail) )
-
-        args = [ __file__, "-f", "-1"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pfail = %s " % str(args.pfail) )
-
-        args = [ __file__, "-f", "2"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pfail = %s " % str(args.pfail) )
-
-        args = [ __file__, "-w", "-1"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pwarn = %s " % str(args.pwarn) )
-
-        args = [ __file__, "-w", "2"]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pwarn = %s " % str(args.pwarn) )
-
-        args = [ __file__, "-f", "0.65" , "-w", "0.85" ]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-        self.info( "args (after) =%s" % str(args) )
-        self.info( "opts (after) =%s" % str(opts) )
-        self.info( "args.pfail = %s " % str(args.pfail) )
-        self.info( "args.pwarn = %s " % str(args.pwarn) )
-
-        args = [ __file__, "-w", "0.65" , "-f", "0.85" ]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-     
-        args = [ __file__, "-f", "0.75" , "-w", "0.65" ]
-        self.info( "args (before) =%s" % str(args) )
-        args, opts = self.opt.parse( args )
-
-
-    def test_04_callback( self ):
-        args = [ __file__, "--branch", "blah" ]
-        args, opts = self.opt.parse( args )
-        
-        args = [ __file__, "--branch", "14.2X.Y" ]
-        args, opts = self.opt.parse( args )
-        self.info( "args.branch = %s" % str(args.branch) )
-
-
-        args = [ __file__, "--install", "blah" ]
-        args, opts = self.opt.parse( args )
-        
-        args = [ __file__, "--install", "build" ]
-        args, opts = self.opt.parse( args )
-        self.info( "args.install = %s" % str(args.install) )
-        
-        args = [ __file__, "--cmtconfig", "blah" ]
-        args, opts = self.opt.parse( args )
-        
-        args = [ __file__, "--cmtconfig", "i686-slc4-gcc34-opt" ]
-        args, opts = self.opt.parse( args )
-        self.info( "args.cmtconfig = %s" % str(args.cmtconfig) )
-
-
-        args = [ __file__, "--project", "blah" ]
-        args, opts = self.opt.parse( args )
-        
-        args = [ __file__, "--project", "AtlasProduction" ]
-        args, opts = self.opt.parse( args )
-        self.info( "args.cmtconfig = %s" % str(args.cmtconfig) )
-
-
-## test suite execution
-if __name__ == "__main__":
-
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubeOptParser)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-    
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePHPWriter.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePHPWriter.py
deleted file mode 100644
index cc6a9d87309dda14eb6b05a9e8d2d3a5fd130f2d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePHPWriter.py
+++ /dev/null
@@ -1,193 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubePHPWriter.py
-# @author Krzyszotf Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief implementation of DCubePHPWriter and test_DCubePHPWriter classes
-
-import sys
-import os
-from time import strftime, localtime, strptime
-from DCubeUtils import DCubeObject, DCubeException, DCubeVersion
-from DCubeOptParser import DCubeOptParser 
-
-
-import unittest
-
-
-##
-# @class DCubePHPWriter
-# @author Krzysztof Daniel Ciba (Krzytof.Ciba@NOSPAMagh.edu.pl) 
-# @brief PHP files producer
-class DCubePHPWriter( DCubeObject ):
-    
-    ## c'tor
-    # @param self "My, myself and Irene"
-    # @param parsed opts and args from DCubeOptParser
-    def __init__( self, parsed ):
-        super( DCubePHPWriter, self ).__init__( self )
-        self.opts, self.args = parsed
-
-        self.server = self.opts.server
-        if ( not os.path.isdir( self.server ) ):
-            self.server = os.path.dirname( self.server )
-
-        self.debug( "server %s" % self.server )
-        self.debug( "output %s" % os.path.dirname( os.path.abspath( self.opts.output ) ) )
-
-        self.server = self.__relpath( self.server, os.path.dirname(self.opts.output ) ) 
-
-        if ( self.server == "..") : self.server = "../"
-        self.server = "./" + self.server
-        self.info( "server relative path %s" % self.server )
-
-        
-
-    ## give relative path between target and base
-    # @param self "Me, myself and Irene"
-    # @param target target directory name
-    # @param base base directory name
-    def __relpath( self, target, base="." ):
-   
-        if ( not os.path.exists(target) ):
-            self.error( "target directory %s does not exist" % target )
-
-        if ( not os.path.isdir(base) ):
-            self.warn( "base %s is not a directory or does not exist" % base )
-        
- 
-        base_list = (os.path.abspath(base)).split(os.sep)
-        target_list = (os.path.abspath(target)).split(os.sep)
-
-        for i in range(min(len(base_list), len(target_list))):
-            if base_list[i] <> target_list[i]: break
-            else:
-                i+=1
-
-        rel_list = [os.pardir] * (len(base_list)-i) + target_list[i:]
-        return os.path.join(*rel_list)
-    
-    ## php heading string
-    # @param self "Me, myself and Irene"
-    def __head( self ):
-        return "<?php\n"
-
-    ## php tail string
-    # @param self "Me, myself and Irene"
-    def __tail( self ):
-        return "?>\n"
-
-    ## php comment string
-    # @param self "Me, myself and Irene"
-    def __comment( self, what="output" ):
-        
-        out  = "/**\n"
-        out += "  * DCubeClient PHP %s file\n" % what
-        out += "  * autogenerated using %s\n" % DCubeVersion().version()
-        out += "  * on %s\n" % strftime("%a, %d %b %Y %H:%M:%S %Z" , localtime())
-        out += "  * Files:\n"
-        out += "  * [1] monitored file  = %s\n" % str( self.opts.monitored )
-        out += "  * [2] reference file  = %s\n" % str( self.opts.reference )
-        out += "  * [3] config XML file = %s\n" % str( self.opts.config )
-        out += "  * [4] output XML file = %s\n" % str( self.opts.output )
-        out += "  * [5] log file        = %s\n" % str( self.opts.log )
-        out += "  *\n"
-        out += "  *  ***************************************************\n"
-        out += "  *  *                   !!! WARNINIG !!!              *\n"
-        out += "  *  * make sure that dcube.php is in PHP include path *\n"
-        out += "  *  ***************************************************\n"
-        out += "  *\n"
-        out += "  *\n"
-        out += "  *  ***************************************************\n"
-        out += "  *  *                   !!! WARNINIG !!!              *\n"
-        out += "  *  * make sure to put relative path from your result *\n"
-        out += "  *  * directory to the the directory with dcube.php   *\n"
-        out += "  *  ***************************************************\n"
-        out += "  *\n" 
-        
-        return out
-
-    ## body of output PHP file
-    # @param self "Me, myself and Irene"
-    def __bodyPHP( self ):
-        out  = "/* ADD TO include_path TO LOCAL INSTALLATION OF DCUBE PHP PART */\n"
-        out += "$where = \"%s\";\n" % self.server
-        out += "set_include_path($where);\n";
-        out += "require \"dcube.php\";\n\n";
- 
-        out += "$xml_file = \"%s\";\n" % self.opts.output
-        out += "$log_file = \"%s\";\n" % self.opts.log 
-        out += "$dcube = new dcube( $xml_file, $log_file, $where );\n"
-        return out
-
-    ## body for log php file
-    # @param self "Me, myself and Irene"
-    def __bodyLOG( self ):
-        out  = "/* ADD TO include_path TO LOCAL INSTALLATION OF DCUBE PHP PART */\n";
-        out += "$where = \"%s\";\n" % self.server
-        out += "set_include_path($where);\n"
-        out += "require \"rw.php\";\n\n"
-        out += "$log_file = \"%s\";\n" % self.opts.log
-        out += "$page = new rainbow( $log_file );\n"
-        return out
-
-    ## dcube output PHP file contents 
-    # @param self "Me, myself and Irene"
-    def dcubePHP( self ):
-        out  = self.__head()
-        out += self.__comment()
-        out += self.__bodyPHP()
-        out += self.__tail()
-        return out
-        
-
-    ## dcube log PHP file contents
-    # @param self "Me, myself and Irene"
-    def dcubeLOG( self ):
-        out  = self.__head()
-        out += self.__comment( "log" )
-        out += self.__bodyLOG()
-        out += self.__tail()
-        return out
- 
-
-##
-# @class test_DCubePHPWriter
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubePHPWriter class
-class test_DCubePHPWriter( unittest.TestCase ):
-
-    ## test case setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-        sys.argv = [__file__, "-r", "testRef.root", "-s", "/path/to/server/", "-x", "output.xml", "montitored.root" ]
-        self.parsed  = DCubeOptParser().parse( sys.argv )
-        self.parsed[0]._update_loose( { "monitored" : "monitored.root" } )
-
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    def test_01_ctor( self ):
-        try:
-            self.phpWriter = DCubePHPWriter( self.parsed )
-        except:
-            pass
-        self.assertEqual( isinstance(self.phpWriter, DCubePHPWriter), True)
-
-
-    ## dcubePHP  nad dcubeLOG
-    # @param self "Me, myself and Irene"
-    def test_02_writer( self ):
-        phpWriter = DCubePHPWriter( self.parsed )
-        print phpWriter.dcubePHP( )
-        print phpWriter.dcubeLOG( )
-    
-
-## test suite execution
-if __name__ == "__main__":
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubePHPWriter)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-
- 
-
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePlotter.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePlotter.py
deleted file mode 100644
index 195dbe8cbf0bf1e4fb6edb6b8093bb98c579ffda..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubePlotter.py
+++ /dev/null
@@ -1,1011 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubePlotter.py
-# @author Krzysztof Daniel Ciba (Krzyszotf.Ciba@NOSPAagh.edu.pl)
-# @brief implementation of DCubePlotter, DCubeStyle and test_DCubePlotter classes
-
-import os
-import uuid
-import ROOT
-from DCubeUtils import DCubeObject, DCubeException
-import unittest
-
-##
-# @class DCubePlotter 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief root plotter for DCubeClient package
-class DCubePlotter( DCubeObject ):
-
-    ## DCubePlotter TCanvas
-    canvas = None
-
-    ## TPaveText with configuration
-    configPave = None
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param xmldoc DOM XML Document instance
-    # @param parsed tuple with args and opts from option parser
-    def __init__( self, xmldoc, parsed):
-        super( DCubePlotter, self).__init__( self )
-        self.debug( "constructing DCubePlotter...")
-
-        self.xmldoc = xmldoc
-        self.opts, self.args = parsed
-
-        self.outputDir = os.path.join( os.path.dirname( self.opts.output ), "plots" )
-        self.info("will store all plots in directory %s" % self.outputDir )
-
-        try:
-            if (not os.path.exists( self.outputDir ) ):
-                os.mkdir( self.outputDir )
-             
-        except:
-            self.error("cannot create output plot directory, plot creation will be disabled")
-            self.opts._update_careful( { "makeplots":  False } )
-
-
-        self.__applyTStyle()
-
-    ## prepare/return ROOT.TCanvas instance
-    # @param self "Me, myself and Irene" 
-    def __getCanvas( self ):
-        if ( not self.canvas ):
-            self.canvas = ROOT.TCanvas( "dcube", "dcube" , 600, 600 )
-        return self.canvas
-
-    ## set DCube plot dtaw style 
-    # @param self "Me, myself and Irene"
-    def __applyTStyle( self, name="DCube" ):
-
-        self.debug("setting up DCube plot style...")
-
-        ROOT.gStyle.SetOptTitle(0)
-        ROOT.gStyle.SetOptStat(0)
-        ROOT.gStyle.SetOptFit(0)
-  
-        ROOT.gStyle.SetFillStyle(1001)
-
-        ROOT.gStyle.SetMarkerSize( 0.9 )
-        ROOT.gStyle.SetMarkerStyle( ROOT.kFullCircle )
-        
-        ROOT.gStyle.SetLabelFont(42, "xyz")  
-        ROOT.gStyle.SetLabelSize(0.03, "xyz")
-        
-        ROOT.gStyle.SetTitleFont(42, "xyz")
-        ROOT.gStyle.SetTitleFontSize(0.04)
-        
-        ROOT.gStyle.SetTitleFont(42, "a")   
-
-        ROOT.gStyle.SetFuncStyle(1)
-        ROOT.gStyle.SetFuncWidth(2)
-        
-        ROOT.gStyle.SetHistLineColor( ROOT.kBlack )
-        ROOT.gStyle.SetHistFillColor( ROOT.kRed )
-        ROOT.gStyle.SetHistLineWidth(1)
-
-        ROOT.gStyle.SetPadBorderSize(1)
-        ROOT.gStyle.SetPadBorderMode(0)
-                
-        ROOT.gStyle.SetPadLeftMargin( 0.1 )
-        ROOT.gStyle.SetPadBottomMargin( 0.1 )
-        ROOT.gStyle.SetPadRightMargin( 0.1 )
-        ROOT.gStyle.SetPadTopMargin( 0.1 )
-            
-        ROOT.gStyle.SetCanvasBorderSize(1)
-        ROOT.gStyle.SetCanvasBorderMode(0)
-  
-        ROOT.gStyle.SetGridStyle(3)
-        ROOT.gStyle.SetGridWidth(1)
-        
-        ROOT.gStyle.SetOptDate(21)
-        ROOT.gStyle.GetAttDate().SetTextFont(42)
-        ROOT.gStyle.GetAttDate().SetTextSize(0.025)
-           
-        ROOT.gStyle.SetLegendBorderSize(1)
- 
-        ROOT.gStyle.SetTextColor( ROOT.kBlack )
-        ROOT.gStyle.GetAttDate().SetTextColor( ROOT.kBlack )
-        ROOT.gStyle.SetLabelColor(ROOT.kBlack, "xyz")
-        ROOT.gStyle.SetTitleColor(ROOT.kBlack, "xyz")
-        
-        ROOT.gStyle.SetFillColor( ROOT.kWhite )
-        ROOT.gStyle.SetMarkerColor( ROOT.kBlue )
-        ROOT.gStyle.SetCanvasColor( ROOT.kWhite )    
-
-        ROOT.gStyle.SetFrameFillColor(ROOT.kWhite)
-        ROOT.gStyle.SetPadColor( ROOT.kGray )
-        ROOT.gStyle.SetTitleColor(390, "a")
-        ROOT.gStyle.SetFuncColor(ROOT.kOrange)
-        
-        ROOT.gStyle.SetGridColor(1)
-
-        ROOT.gStyle.SetCanvasDefH( 800 )
-        ROOT.gStyle.SetCanvasDefW( 800 )
-
-        ROOT.gROOT.ForceStyle(1)
-        self.debug("done!")
-        
-
-    ## get unique plot name
-    # @param self "Me, myself and Irene"
-    def __plotName( self, what="" ):
-        return  what + "_" + str(uuid.uuid4()) + ".png"
-
-
-    ## plot dispatcher
-    # @param self "Me, myself and Irene"
-    # @param node DOM XML element
-    # @param mon monitored object
-    # @param ref reference object
-    def plot( self, node, mon=None, ref=None ):
-
-        self.mon = self.ref = self.node = None
-        self.debug("plot called for object of name " + node.getAttribute("name") )
-
-        self.node = node
-        self.mon = mon
-        self.ref = ref
-
-        if ( not self.mon ):
-            self.error("monitored object not found, creation of plots skipped!")
-            return "FAIL"
-
-        self.name = self.mon.GetName()
-        self.title = self.mon.GetTitle()
-
-        if ( not self.ref ):
-            self.warn("reference object not found, not all plots will be produced")
-
-        self.className = mon.Class().GetName()
-
-        status = "OK" 
-        if ( "TGraph" in self.className ):
-            status = self.__plotGraph()
-        elif ( "TH1" in self.className ):
-            status = self.__plotTH1()
-        elif ( "TH2" in self.className ):
-            status = self.__plotTH2()
-        elif ( "TProfile" == self.className ):
-            status = self.__plotProf1D()
-        elif ( "TProfile2D" == self.className ):
-            status = self.__plotProf2D()
-        else:
-            self.error( "unsupported object name=%s class=%s" % ( self.name, self.className ) )
-            status = "FAIL"
-
-        return status
-
-    ## save TCanvas to png file
-    # @param self "Me, myself and Irene"
-    # @param type string with type attribute
-    # @param plot plot file naem prefix
-    # @return status string 
-    def __saveAs( self, canvas, type, plot ):
-        try:
-            plotName = self.__plotName( plot )
-            absPath = os.path.join( self.outputDir, plotName )
-            canvas.Update()
-            #canvas.SaveAs( absPath )
-            canvas.SaveAs( absPath )
-            imgNode = self.xmldoc.createElement( "img" )
-            imgNode.setAttribute( "src" , os.path.join( "plots", plotName ) )
-            imgNode.setAttribute( "type", type )
-            if ( os.path.exists(absPath) ):
-                self.debug("plot '%s' has been created" % absPath ) 
-                self.node.appendChild( imgNode )
-                return "OK"
-            else:
-                self.warn("problem when saving plot '%s' to file" % absPath )
-                return "WARN"
-        except:
-            self.warn("unknown error when plotting %s for %s" % ( type, self.name )  )
-            
-        return "WARN"
-        
-
-    ## plotter for TGraph, TGraphError, TGraphAsymmError etc.
-    # @param self "Me, myself and Irene"
-    def __plotGraph( self ):
-
-        status = [ "OK" ]
-        self.debug("plotting %s of type %s" % ( self.name, self.className ) )
-
-        graph = ROOT.TMultiGraph()
-
-        canvas = self.__getCanvas()
-        canvas.cd()
-
-        xmin, xmax, ymin, ymax = self.__getMinMaxTGraph( self.mon, self.ref )
-
-        dx = abs(xmax - xmin) * 0.15
-        dy = abs(ymax - ymin) 
-        
-        ROOT.gPad.DrawFrame( xmin - dx, ymin - dy * 0.4, xmax + dx, ymax + dy*1.4)
-
-        ROOT.gStyle.SetPadBottomMargin(0.2)
-        
-        #self.mon.SetMarkerStyle( ROOT.kFullCircle )
-        self.mon.SetMarkerColor( ROOT.kRed )
-        self.mon.SetMarkerSize( 0.9 )        
-        self.mon.SetFillColor( ROOT.kRed )
-        self.mon.SetFillStyle( 3005 )
-        self.mon.SetLineColor( ROOT.kRed )
-            
-        graph.Add( self.mon, "P") 
-                            
-        if ( self.ref ):
-            #self.ref.SetMarkerStyle( ROOT.kFullCircle )
-            self.ref.SetMarkerColor( ROOT.kBlue )
-            self.ref.SetMarkerSize( 0.9 )
-            self.ref.SetFillColor( ROOT.kBlue )
-            self.ref.SetFillStyle( 3004 )
-            self.ref.SetLineColor( ROOT.kBlue )
-            
-            graph.Add( self.ref, "P") 
-
-        graph.Draw("P")        
-
-        legend = self.__legend()
-        legend.Draw()
-
-        titlePave = self.__titlePave()
-        titlePave.Draw()
-        
-        configPave = self.__configPave()
-        configPave.Draw()
-
-        status.append( self.__saveAs( canvas, "reg", "gr" ) )
-        
-            
-        if ( None not in ( self.ref, self.mon ) ):
-
-            canvas.Clear()
-            
-            diffGraph = ROOT.TGraph()
-
-            diffGraph.SetMarkerColor( ROOT.kRed )
-            
-            monN = self.mon.GetN()
-            refN = self.ref.GetN()
-
-            monXraw = self.mon.GetX()
-            monYraw = self.mon.GetY()
-
-            refXraw = self.ref.GetX()
-            refYraw = self.ref.GetY()
-            
-            points = {}
-            for i in range( monN ):
-                if ( monXraw[i] not in points ):
-                    points [ monXraw[i] ] = monYraw[i]        
-            for i in range( refN ):
-                if ( refXraw[i] not in points ):
-                    points[ refXraw[i] ] = refYraw[i]
-                else:
-                    points[ refXraw[i] ] = points[ refXraw[i] ] - refYraw[i]
-            i = 0
-            for x in sorted( points.keys() ):
-                diffGraph.SetPoint(i, x, points[x] )
-                i += 1
-
-            xl = xmin - dx
-            xr = xmax + dx
-
-            xmin, xmax, ymin, ymax = self.__getMinMaxTGraph( diffGraph )
-            
-            dy = abs(ymax - ymin) 
-
-            if ( ymin > ymax ):
-                a = ymin
-                ymin = ymax
-                ymax = ymin
-            
-            ROOT.gPad.DrawFrame( xl, ymin - (dy * 0.4), xr, ymax + (dy*1.4) )
-            
-            diffGraph.SetTitle( self.mon.GetTitle() )
-            diffGraph.SetName( self.mon.GetName() )
-
-            diffGraph.Draw("P")
-                       
-            titlePave = self.__titlePave("diff (mon - ref)")
-            titlePave.Draw()
-            
-            configPave = self.__configPave()
-            configPave.Draw()
-
-            status.append( self.__saveAs( canvas, "dif", "gd" ) )
-            
-        return self.__getStatus( status )
-
-    ## get xmin, xmax, ymin, ymax for monitored and reference  TGraphs  
-    # @param self "Me, myself and Irene"
-    def __getMinMaxTGraph( self, mon=None, ref=None ):
-
-        xmin = xmax = ymin=  ymax = None
-        if ( mon ):
-            x = ROOT.Double(0)
-            y = ROOT.Double(0)
-
-            exl = ROOT.Double(0)
-            exr = ROOT.Double(0)
-            eyl = ROOT.Double(0)
-            eyh = ROOT.Double(0) 
-
-            if ( self.className == "TGraph" ):
-                
-                for i in range( mon.GetN()  ):
-                    
-                    mon.GetPoint(i, x, y)
-                    
-                    if ( xmin == None ): xmin = ROOT.Double(x)
-                    if ( xmax == None ): xmax = ROOT.Double(x)
-                    if ( ymin == None ): ymin = ROOT.Double(y)
-                    if ( ymax == None ): ymax = ROOT.Double(y)
-                    
-                    if ( x < xmin ): xmin = x
-                    if ( x > xmax ): xmax = x
-                    if ( y < ymin ): ymin = y
-                    if ( y > ymax ): ymax = y
-                    
-                if ( ref ):
-                    for i in range( ref.GetN() ):
-
-                        ref.GetPoint(i, x, y)
-                        
-                        if ( xmin == None ): xmin = ROOT.Double(x)
-                        if ( xmax == None ): xmax = ROOT.Double(x)
-                        if ( ymin == None ): ymin = ROOT.Double(y)
-                        if ( ymax == None ): ymax = ROOT.Double(y)
-                        
-                        if ( x < xmin ): xmin = x
-                        if ( x > xmax ): xmax = x
-                        if ( y < ymin ): ymin = y
-                        if ( y > ymax ): ymax = y
-            else:
-
-                for i in range( mon.GetN() ):
-
-                    mon.GetPoint(i, x, y)
-
-                    exl = mon.GetErrorXhigh( i )
-                    exr = mon.GetErrorXlow( i )
-                    eyh = mon.GetErrorYhigh( i )
-                    eyl = mon.GetErrorYlow( i )
-
-                    xl = x - exr
-                    xr = x + exl
-                    yl = y - eyl
-                    yh = y + eyh
-
-                    if ( xmin == None ): xmin = ROOT.Double(xl)
-                    if ( xmax == None ): xmax = ROOT.Double(xr)
-                    if ( ymin == None ): ymin = ROOT.Double(yl)
-                    if ( ymax == None ): ymax = ROOT.Double(yh)
-
-                    if ( xl < xmin ): xmin = xl
-                    if ( xr > xmax ): xmax = xr
-                    if ( yl < ymin ): ymin = yl
-                    if ( yh > ymax ): ymax = yh
-                    
-                if ( ref ):
-
-                    for i in range( ref.GetN() ):
-
-                        ref.GetPoint(i, x, y)
-                        
-                        exl = ref.GetErrorXhigh( i )
-                        exr = ref.GetErrorXlow( i )
-                        eyh = ref.GetErrorYhigh( i )
-                        eyl = ref.GetErrorYlow( i )
-                        
-                        xl = x - exr
-                        xr = x + exl
-                        yl = y - eyl
-                        yh = y + eyh
-
-                        if ( xmin == None ): xmin = ROOT.Double(xl)
-                        if ( xmax == None ): xmax = ROOT.Double(xr)
-                        if ( ymin == None ): ymin = ROOT.Double(yl)
-                        if ( ymax == None ): ymax = ROOT.Double(yh)
-
-                        if ( xl < xmin ): xmin = xl
-                        if ( xr > xmax ): xmax = xr
-                        if ( yl < ymin ): ymin = yl
-                        if ( yh > ymax ): ymax = yh
-
-
-            return ( xmin, xmax, ymin, ymax)
-            
-
-    ## plotter for TH1 histograms
-    # @param self "Me, myself and Irene"
-    def __plotTH1( self ):
-
-        status = [ "OK" ]
-        self.debug("plotting %s of type %s" % ( self.name, self.className ) )
-
-        canvas = self.__getCanvas()
-        titlePave = self.__titlePave( )
-        title = "%s;%s;%s" % ( self.mon.GetTitle(),
-                               self.mon.GetXaxis().GetTitle(),
-                               self.mon.GetYaxis().GetTitle() )
-        stack = ROOT.THStack( "dcube stack", title )
-
-        self.mon.SetFillStyle( 3004 )
-        self.mon.SetLineColor( ROOT.kRed )
-        stack.Add( self.mon )
-
-        ymax = self.mon.GetMaximum()
-        
-        if ( self.ref ):
-            self.ref.SetFillColor( ROOT.kBlue )
-            self.ref.SetLineColor( ROOT.kBlue )
-            self.ref.SetFillStyle( 3005 )
-            stack.Add( self.ref )
-
-            if ( ymax < self.ref.GetMaximum() ): ymax = self.ref.GetMaximum()
-
-        stack.SetMaximum( 1.4 * ymax )
-
-        config = self.__configPave()
-        legend = self.__legend()
-
-        canvas.cd()
-        stack.Draw( "NOSTACK" )
-        titlePave.Draw()
-        config.Draw()
-        legend.Draw()
-        pvaluePave = self.__pvaluePave()
-        if ( pvaluePave ): pvaluePave.Draw()
-
-        status.append( self.__saveAs( canvas, "reg", "h1r" ) )
-        
-        diffHist = self.mon.Clone()
-        diffHist.Reset()
-        diffHist.SetTitle( "diff %s" % title )
-        diffHist.SetLineColor( ROOT.kRed )
-
-       
-
-        if ( self.ref ):
-            diffHist.Add( self.ref, self.mon, 1.0, -1.0 )
-        else:
-            diffHist.Add( self.mon, self.mon, 1.0, -1.0 )
-
-        canvas.Clear()
-        canvas.cd()
-        
-        titlePave = self.__titlePave( "diff" )
-        configPave = self.__configPave()
-        
-        ymax = diffHist.GetMaximum()
-        if ( ymax ) : diffHist.SetMaximum( 1.4 * ymax )
-
-        diffHist.Draw()
-        titlePave.Draw()
-        configPave.Draw()
-        
-        status.append( self.__saveAs( canvas, "dif", "h1d" ) )
-
-        return self.__getStatus( status )
-
-    ## plotter for TH2 histograms
-    # @param self "Me, myself and Irene"
-    def __plotTH2( self ):
-
-        status = [ "OK" ]
-        self.debug( "plotting %s of type %s" % ( self.name, self.className) )
-
-        canvas = self.__getCanvas()
-        title = "%s;%s;%s;%s" % ( self.mon.GetTitle(),
-                                  self.mon.GetXaxis().GetTitle(),
-                                  self.mon.GetYaxis().GetTitle(),
-                                  self.mon.GetZaxis().GetTitle() )
-
-        configPave = self.__configPave( )
-        titlePave = self.__titlePave() 
-        pvaluePave = self.__pvaluePave()
-
-        if ( self.ref ):
-
-            bottom = self.mon.Clone()
-            top = self.mon.Clone()
-            same = self.mon.Clone()
-
-            bottom.Reset()
-            bottom.SetLineColor( ROOT.kBlack )
-            bottom.SetFillColor( ROOT.kBlue )
-            bottom.SetLineWidth( 2 )
-            top.Reset()
-            top.SetLineColor( ROOT.kBlack )
-            top.SetFillColor( ROOT.kRed )
-            top.SetLineWidth(2)
-            same.Reset()
-            same.SetLineColor( ROOT.kBlack )
-            same.SetFillColor( ROOT.kGreen )
-            same.SetLineWidth( 2)
-
-            for i in range(self.ref.GetNbinsX()+1): 
-                for j in range(self.ref.GetNbinsY()+1 ) :             
-
-                    refBin = self.ref.GetBinContent(i, j)
-                    monBin = self.mon.GetBinContent(i, j)
-
-                    if ( refBin < monBin ):
-                        top.SetBinContent(i, j, monBin - refBin)
-                        bottom.SetBinContent( i, j, refBin )
-                    elif ( refBin > monBin ):
-                        bottom.SetBinContent( i, j, refBin - monBin)
-                        top.SetBinContent(i, j, monBin)
-                    else:
-                        same.SetBinContent(i, j, refBin )
-              
-            stack = ROOT.THStack( "dcube stack", title )
-            stack.Add( same )
-            stack.Add( bottom )
-            stack.Add( top )
-            #stack.UseCurrentStyle()
-
-            legend = ROOT.TLegend(0.1, 0.80, 0.45, 0.72, "", "NDC")
-            legend.SetTextFont(102)
-            legend.SetTextSize(0.02)
-            legend.SetTextColor(1)
-            legend.SetBorderSize(1)
-
-            legend.AddEntry( top, "monitored", "F" )
-            legend.AddEntry( bottom, "reference", "F" )
-            legend.AddEntry( same, "same", "F")
-
-            canvas.cd()
-            stack.Draw( "lego1 nostack" )
-            titlePave.Draw()
-            configPave.Draw()
-            legend.Draw()
-
-            if ( pvaluePave ): pvaluePave.Draw()
-
-            canvas.Draw()
-
-        else:
-            canvas.cd()
-            self.mon.Draw()
-            
-            configPave.Draw()
-            titlePave.Draw()
-            if ( pvaluePave ): pvaluePave.Draw()
-            canvas.Draw()
-            
-        status.append( self.__saveAs( canvas, "reg", "h2r" ) )
-
-        # make diff plot
-        diffHist = self.mon.Clone()
-        diffHist.Reset()
-        diffHist.SetFillColor( ROOT.kRed )
-        diffHist.SetLineColor( ROOT.kBlack )
-
-        if ( self.ref ):
-            diffHist.Add( self.ref, self.mon, 1.0, -1.0 )
-        else:
-            diffHist.Add( self.mon, self.mon, 1.0, -1.0 )
-
-        canvas.Clear()   
-        diffHist.Draw("LEGO1 0")
-        titlePave = self.__titlePave( "diff" )
-        titlePave.Draw()
-        configPave.Draw()
-        if ( pvaluePave ): pvaluePave.Draw()
-        status.append( self.__saveAs( canvas, "dif", "h2d" ) )
-
-
-        # make projection x plot
-        canvas.Clear()
-        
-    
-        projX = ROOT.THStack("projectionX", "projectionX")
-        monPX = self.mon.ProjectionX( "mon px" )
-        monPX.SetFillStyle( 3004 )
-        monPX.SetLineColor( ROOT.kRed )
-        projX.Add( monPX )
-
-        ymax = monPX.GetMaximum()
-
-        if ( self.ref ):
-            refPX = self.ref.ProjectionX( "ref px")
-            refPX.SetFillStyle( 3005 )
-            refPX.SetFillColor( ROOT.kBlue )
-            refPX.SetLineColor( ROOT.kBlue )
-            if ( ymax < refPX.GetMaximum() ):
-                ymax = self.ref.GetMaximum()
-                
-            projX.Add( refPX )
-
-        projX.SetMaximum( ymax * 1.4 )
-        
-        canvas.cd()    
-        projX.Draw("NOSTACK")
-
-        titlePave = self.__titlePave( "proj X")
-        titlePave.Draw()
-        if ( pvaluePave ): pvaluePave.Draw()
-        configPave.Draw()
-
-        legend = ROOT.TLegend(0.1, 0.80, 0.45, 0.74, "", "NDC")
-        legend.SetTextFont(102)
-        legend.SetTextSize(0.02)
-        legend.SetTextColor(1)
-        legend.SetBorderSize(1)
-
-        legend.AddEntry( monPX, "mon projection X", "F" )
-        legend.AddEntry( refPX, "ref projection X", "F" )
-        legend.Draw()
-
-        status.append( self.__saveAs( canvas, "prx", "h2x" ) )
-        
-        # make projection y plot
-        canvas.Clear()
-
-        projY = ROOT.THStack("projectionY", "projectionY")
-        monPY = self.mon.ProjectionY( "mon py" )
-        monPY.SetFillStyle( 3004 )
-        monPY.SetLineColor( ROOT.kRed )
-        projY.Add( monPX )
-
-        ymax = monPY.GetMaximum()
-
-        if ( self.ref ):
-            refPY = self.ref.ProjectionX( "ref py")
-            refPY.SetFillStyle( 3005 )
-            refPY.SetFillColor( ROOT.kBlue )
-            refPY.SetLineColor( ROOT.kBlue )
-            if ( ymax < refPY.GetMaximum() ):
-                ymax = refPY.GetMaximum()
-            
-            projY.Add( refPY )
-
-        projY.SetMaximum( ymax * 1.4 )
-            
-        canvas.cd()    
-        projY.Draw("NOSTACK")
-
-        titlePave = self.__titlePave( "proj Y")
-        titlePave.Draw()
-        configPave.Draw()
-        if ( pvaluePave ): pvaluePave.Draw()
-
-        legend = ROOT.TLegend(0.1, 0.80, 0.45, 0.74, "", "NDC")
-        legend.SetTextFont(42)
-        legend.SetTextSize(0.02)
-        legend.SetTextColor(1)
-        legend.SetBorderSize(1)
-
-        legend.AddEntry( monPX, "mon projection Y", "F" )
-        legend.AddEntry( refPX, "ref projection Y", "F" )
-        legend.Draw()
-
-        status.append( self.__saveAs( canvas, "pry", "h2y" ) )
-        
-        return self.__getStatus( status )
-
-    ## plotter for 1D TProfile
-    # @param self "Me, myself and Irene"
-    def __plotProf1D( self ):
-        status = [ ]
-        self.debug("plotting %s of type %s" % ( self.name, self.className ) )
-
-        canvas = self.__getCanvas()
-
-        titlePave = self.__titlePave()
-        configPave = self.__configPave()
-        legend = self.__legend()
-
-     
-        self.mon.SetMarkerColor( ROOT.kRed )
-        self.mon.SetMarkerSize( 0.9 )        
-        self.mon.SetFillColor( ROOT.kRed )
-        self.mon.SetFillStyle( 3005 )
-        self.mon.SetLineColor( ROOT.kRed )
-        
-        ymax = self.mon.GetMaximum()
-
-        if ( self.ref ):
-     
-            self.ref.SetMarkerColor( ROOT.kBlue )
-            self.ref.SetMarkerSize( 0.9 )        
-            self.ref.SetFillColor( ROOT.kBlue )
-            self.ref.SetFillStyle( 3004 )
-            self.ref.SetLineColor( ROOT.kBlue )
-
-            if ( ymax < self.ref.GetMaximum() ):
-                ymax = self.ref.GetMaximum()
-    
-        self.mon.SetMaximum( 1.4 * ymax )
-        self.mon.Draw()
-        if ( self.ref ): self.ref.Draw( "same" )
-        titlePave.Draw()
-        configPave.Draw()
-        legend.Draw()
-            
-        status.append( self.__saveAs( canvas, "reg", "p1r" ) )
-
-        canvas.Clear()
-        canvas.cd()
-        
-        diffProfile = self.mon.Clone()
-        diffProfile.Reset()
-        if ( self.ref ):
-            diffProfile.Add( self.ref, self.mon, 1.0, -1.0 )
-        else:
-            diffProfile.Add( self.mon, self.mon, 1.0, -1.0 )
-
-        titlePave = self.__titlePave( "diff" )
-        diffProfile.Draw()
-        titlePave.Draw()
-        configPave.Draw()
-        
-        status.append( self.__saveAs( canvas, "dif", "p1d" ) )
-        
-        return self.__getStatus( status ) 
-
-    ## plotter for TProfile2D 
-    # @param self "Me, myself and Irene"
-    def __plotProf2D( self ):
-        status = [ ]
-        self.debug("plotting %s of type %s" % ( self.name, self.className ) )
-
-        canvas = self.__getCanvas()
-
-        titlePave = self.__titlePave()
-        configPave = self.__configPave()
-        legend = self.__legend()
-
-     
-        self.mon.SetMarkerColor( ROOT.kRed )
-        self.mon.SetMarkerSize( 0.9 )        
-        self.mon.SetFillColor( ROOT.kRed )
-        self.mon.SetFillStyle( 3005 )
-        self.mon.SetLineColor( ROOT.kBlack )
-
-        self.mon = ROOT.TH2D( self.mon )
-        
-        title = "%s;%s;%s;%s" % ( self.mon.GetTitle(),
-                                  self.mon.GetXaxis().GetTitle(),
-                                  self.mon.GetYaxis().GetTitle(),
-                                  self.mon.GetZaxis().GetTitle() )
-        
-        
-        ymax = self.mon.GetMaximum()
-        if ( self.ref ):
-            self.ref = ROOT.TH2D( self.ref )
-            
-            bottom = self.ref.Clone()
-            top = self.ref.Clone()
-            same = self.ref.Clone()
-
-            bottom.Reset()
-            bottom.SetLineColor( ROOT.kBlack )
-            bottom.SetFillColor( ROOT.kBlue )
-            bottom.SetLineWidth( 2 )
-            top.Reset()
-            top.SetLineColor( ROOT.kBlack )
-            top.SetFillColor( ROOT.kRed )
-            top.SetLineWidth(2)
-            same.Reset()
-            same.SetLineColor( ROOT.kBlack )
-            same.SetFillColor( ROOT.kGreen )
-            same.SetLineWidth( 2 )
-
-            for i in range(self.ref.GetNbinsX()+1): 
-                for j in range(self.ref.GetNbinsY()+1 ) :             
-
-                    refBin = self.ref.GetBinContent(i, j)
-                    monBin = self.mon.GetBinContent(i, j)
-            
-                    if ( refBin < monBin ):
-                        top.SetBinContent(i, j, monBin - refBin)
-                        bottom.SetBinContent( i, j, refBin )
-                    elif ( refBin > monBin ):
-                        bottom.SetBinContent( i, j, refBin - monBin)
-                        top.SetBinContent(i, j, monBin)
-                    else:
-                        same.SetBinContent(i, j, refBin )
-              
-            stack = ROOT.THStack( "dcube stack", title )
-            stack.Add( same )
-            stack.Add( bottom )
-            stack.Add( top )
-            
-            legend = ROOT.TLegend(0.1, 0.80, 0.45, 0.72, "", "NDC")
-            legend.SetTextFont(102)
-            legend.SetTextSize(0.02)
-            legend.SetTextColor(1)
-            legend.SetBorderSize(1)
-
-            legend.AddEntry( top, "monitored", "F" )
-            legend.AddEntry( bottom, "reference", "F" )
-            legend.AddEntry( same, "same", "F")
-
-            canvas.cd()
-            stack.Draw( "lego1 0 nostack" )
-            titlePave.Draw()
-            configPave.Draw()
-            legend.Draw()
-
-            pvaluePave = self.__pvaluePave()
-            if ( pvaluePave ): pvaluePave.Draw()
-            canvas.Draw()
-
-        else:            
-            canvas.cd()
-            self.mon.Draw()
-            configPave.Draw()
-            titlePave.Draw()
-            canvas.Draw()
-            
-        status.append( self.__saveAs( canvas, "reg", "p2r" ) )
-
-        canvas.Clear()
-        canvas.cd()
-        
-        diffProfile = self.mon.Clone()
-        diffProfile.Reset()
-        diffProfile.SetFillColor( ROOT.kRed )
-        diffProfile.SetLineColor( ROOT.kBlack )
-
-        if ( self.ref ):
-            diffProfile.Add( self.ref, self.mon, 1.0, -1.0 )
-        else:
-            diffProfile.Add( self.mon, self.mon, 1.0, -1.0 )
-
-        titlePave = self.__titlePave( "diff" )
-        diffProfile.Draw( "lego1 0")
-        titlePave.Draw()
-        configPave.Draw()
-        
-      
-        status.append( self.__saveAs( canvas, "dif", "p2d" ) )
-        
-        return self.__getStatus( status ) 
-        
-       
-
-
-    ## get status string from a list of strings
-    # @param self "Me, myself and Irene"
-    # @param sl list of strings
-    def __getStatus( self, sl ):
-        if ( "FAIL" in sl ): return "FAIL"
-        if ( "WARN" in sl ): return "WARN"
-        return "OK"
-
-    ## runtime configuration pave
-    # @param self "Me, myself and Irene"
-    def __configPave( self ):
-        if ( not self.configPave ):
-            self.configPave = ROOT.TPaveText( 0.1, 0.9, 0.45, 0.8, "NDC" )
-            self.configPave.SetBorderSize( 1 )
-            self.configPave.SetTextColor( 1 )
-            self.configPave.SetTextSize( 0.02 )
-            self.configPave.SetTextFont( 102 )
-            self.configPave.SetTextAlign( 12 ) 
-            self.configPave.AddText( "branch:  " + self.opts.branch  )
-            self.configPave.AddText( "install: " + self.opts.install   )
-            self.configPave.AddText( "cmt:     " + self.opts.cmtconfig )
-            self.configPave.AddText( "project: " + self.opts.project   )
-            self.configPave.AddText( "jobId:   " + self.opts.jobId   )
-        return self.configPave
-
-    ## p-value pave
-    # @param self "Me, myself and Irene"
-    def __pvaluePave( self ):
-        pvaluePave = ROOT.TPaveText( )
-        pvaluePave.SetBorderSize( 1 )
-        pvaluePave.SetTextColor( 1 )
-        pvaluePave.SetTextSize( 0.02 )
-        pvaluePave.SetTextFont( 42 )
-        pvaluePave.SetTextAlign( 22 )
-
-        lines = 0
-        pvalues = self.node.getElementsByTagName( "pvalue" )
-        for pvalue in pvalues:
-            test = pvalue.getAttribute( "test" )
-            status = pvalue.getAttribute( "status" )
-            pval = self.__getCData( pvalue.childNodes )
-            text = "p-value = %s" % pval
-            
-           
-            if test == "chi2": text = "#chi^{2} %s" % text
-            else:
-                text = "%-3s  %s" % (test, text)
-            
-            
-            text = pvaluePave.AddText( text  )
-            if ( status == "FAIL" ):
-                text.SetTextColor( ROOT.kRed )
-            if ( status == "WARN" ):
-                text.SetTextColor( ROOT.kOrange )
-            lines += 1
-
-        if ( lines ):
-            Y2NDC = 0.9 - (lines*0.021)
-            pvaluePave.SetX1NDC( 0.69 )
-            pvaluePave.SetX2NDC( 0.9  )
-            pvaluePave.SetY1NDC( 0.9  )
-            pvaluePave.SetY2NDC( 0.9 - (lines*0.021) )
-            return pvaluePave
-            
-        
-    ## title pave
-    # @param self "Me, myself and Irene"
-    def __titlePave( self, what="" ):
-        titlePave = ROOT.TPaveText( 0.0, 1.0, 1.0, 0.93 , "NDC")
-        titlePave.SetBorderSize( 0 )
-        titlePave.SetTextColor( ROOT.kBlack )
-        titlePave.SetTextSize( 0.028 )
-        titlePave.SetTextFont( 42 )
-        titlePave.SetFillColor( ROOT.kGray )
-        titlePave.SetTextAlign( 12 )
-        titlePave.AddText( "title: %s" % self.title  ) 
-        #titlePave.AddText( "name: %s" % self.name ) 
-        if ( what == "" ): what = "normal"
-        titlePave.AddText( "%s plot" % what  ) 
-        return titlePave
-    
-    ## plot legend
-    # @param self "Me, myself and Irene"
-    def __legend( self ):
-        legend = ROOT.TLegend(0.1, 0.80, 0.45, 0.75, "", "NDC")
-        legend.SetTextFont(42)
-        legend.SetTextSize(0.02)
-        legend.SetTextColor( ROOT.kBlack )
-        legend.SetBorderSize(1)
-
-        if ( "TH" in self.className ):
-            legend.AddEntry( self.mon, "monitored", "f" )
-            if ( self.ref ): legend.AddEntry( self.ref, "reference", "f" )
-        else:
-            legend.AddEntry( self.mon, "monitored", "pe")
-            if ( self.ref ): legend.AddEntry( self.ref, "reference", "pe")
-            
-        return legend
-
-    ## CDATA getter 
-    # @param self "Me, myself and Irene"
-    # @param nodeList list of DOM nodes 
-    def __getCData( self, nodeList ):
-        cdata = ""
-        for node in nodeList:
-            if ( node.nodeType == node.TEXT_NODE ):
-                cdata += node.data
-        return cdata
-##
-# @class test_DCubePlotter
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubePlotter
-class test_DCubePlotter( unittest.TestCase ):
-
-    ## test case setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):        
-        pass
-
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    def test_01_ctor( self ):
-        try:
-            self.plotter =  DCubePlotter()
-        except:
-            pass
-        self.assertEqual( isinstance( self.plotter, DCubePlotter), True )
-            
-    
-
-# test suite execution
-if __name__ == "__main__":
-
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubePlotter)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-    
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeRunner.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeRunner.py
deleted file mode 100644
index 62b9e67a329a916837cd16683822e6edeba702f3..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeRunner.py
+++ /dev/null
@@ -1,377 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-## 
-# @module DCubeRunner
-# @author Krzysztof Daniel Ciba
-# @date 4 Sep 2007
-# @brief Glue module to plugin DCube to Atlas RTT. 
-#
-# * Feb 4, 2007 - switch to DCubeClient-00-00-13
-#                 and DCubeServer-00-00-08
-# * Apr 10, 2008 - switch to DCubeServer-00-00-10
-# * Nov 7, 2008 - switch to new DCubeServer-00-00-11
-# * Nov 11, 2008 - switch to new DCubeClient-00-00-19 and DCubeServer-00-00-12 
-# * Dec 15, 2008 - switch to new DCubeClient-00-00-21 and DCubeServer-00-00-13
-# 
-
-## some imports 
-import xml.dom.minidom as minidom
-import os, sys
-import os.path 
-import re
-
-## @class DCubeRunner
-#  @brief prepares and runs shell script with DCube command 
-class DCubeRunner:
-
-      ## DCubeClient API classes
-      __reAPIClasses = [ re.compile("DCubeApp"), 
-                         re.compile("DCubeConfig"), 
-                         re.compile("DCubeTester"), 
-                         re.compile("DCubePlotter"), 
-                         re.compile("DCubePHPWriter") ]
-      
-      ## base directory for references
-      baseRef = "/afs/cern.ch/atlas/project/RTT/Results/reference/"
-
-      ## DCubeServer directory
-      dcubeServer = "/afs/cern.ch/atlas/project/RTT/Results/dcube/DCubeServer-00-00-13"
-
-      ## DCubeClient directory
-      dcubeClient = "/afs/cern.ch/atlas/project/RTT/DCubeClient/DCubeClient-00-00-21/python"
-
-
-      ## c'tor
-      # @param self "Me, myself and Irene"
-      # @param argDict RTT action runner dictionary 
-      def __init__( self, argDict={} ):
-
-            try:
-                  from Logger import Logger
-                  self.logger = Logger()
-            except ImportError:
-                  self.logger = False
-
-            self.error = None 
-
-            self.debug("DCubeRunner constructor is getting args from dictionary...")
-
-
-            self.cfgFile = None 
-            if ( "DCubeCfg" in argDict ):
-                  self.cfgFile = argDict["DCubeCfg"]
-            else:
-                  self.info( "*** No DCubeCfg in DCubeRunner action dictionary ***" )
-                  self.error = "No DCubeCfg in DCubeRunner argDict"
-
-            self.refFile = None 
-            if ( "DCubeRef" in argDict ): 
-                  self.refFile = argDict["DCubeRef"]
-            else:
-                  self.info( "*** No DCubeRef in DCubeRunner action dictionay, will use value from config XML file ***" )
-
-            self.monFile = None 
-            if ( "DCubeMon" in argDict ):
-                  self.monFile = argDict["DCubeMon"]
-            else:
-                  self.info( "*** No DCubeMon in DCubeRunner action dictionary ***" )
-                  self.error = "No DCubeMon in DCubeRunner argDict"
-
-            if ( "DCubeJobId" in argDict ):
-                  self.jobId = argDict["DCubeJobId"]
-            else:
-                  self.jobId = None
-
-            jobDtor = argDict.get('JobDescriptor', None )
-
-            if ( jobDtor != None ):
-                  self.DCubeForRTTCfg = jobDtor.paths.dCubeCfgFile
-                  self.outputPath = jobDtor.runPath
-                  self.installArea = jobDtor.paths.installArea
-                  self.cvsPath = 'offline/' + jobDtor.paths.containerPackage
-                  self.cmtPath = jobDtor.paths.cmtPath
-                  self.nightly = jobDtor.paths.branch # branch
-                  self.install = jobDtor.paths.runType #  install
-                  self.cmtconfig = jobDtor.paths.targetCMTCONFIG # CMTCONFIG
-                  self.project = jobDtor.paths.topProject # Atlas project name
-            else:
-                  ## dummy values for testing
-                  self.DCubeForRTTCfg = "./DCubeForRTTCfg.xml"
-                  self.outputPath  = "/DUMMY/RUN/PATH/FROM/JOB/DESCRIPTOR"
-                  self.installArea = "/DUMMY/INSTALL/AREA/FROM/JOB/DESCRIPTOR"
-                  self.cvsPath     = "/DUMMY/CONTAINER/PACKAGE/FROM/JOB/DESCRIPTOR"
-                  self.cmtPath     = "/DUMMY/CMT/PATH/FROM/JOB/DESCRIPTOR"
-                  self.nightly     = "dev"
-                  self.install     = "build"
-                  self.cmtconfig   = "i686-slc4-gcc34-opt"
-                  self.project     = "AtlasProduction"
-
-            self.debug("[01] will use '%s' as DCubeClient configuration file for RTT" % self.DCubeForRTTCfg )
-            self.debug("[02] will use '%s' as output path" % self.outputPath )
-            self.debug("[03] will use '%s' as install area path" % self.installArea )
-            self.debug("[04] will use '%s' as package CVS path" % self.cvsPath )
-            self.debug("[05] will use '%s' as cmt path" % self.cmtPath )
-            self.debug("[06] will use DCubeClient monitored file '%s'" % self.monFile )
-            self.debug("[07] will use DCubeClient reference file '%s'" % self.refFile )
-            self.debug("[08] will use DCubeClient configuration file '%s'" % self.cfgFile )
-            self.debug("[09] will use '%s' as DCubeClient branch name" % self.nightly )
-            self.debug("[10] will use '%s' as DCubeClient install name" % self.install )
-            self.debug("[11] will use '%s' as DCubeClient cmtconfig name" % self.cmtconfig )
-            self.debug("[12] will use '%s' as DCubeClient project name" % self.project  )
-            if ( self.jobId ):
-                  self.debug("[13] will use '%s' as DCubeClient jobId" % self.jobId )
-            else:
-                  self.debug("[13] DCubeClient jobId not set!")
-                  
-            self.command = [ ]
-            if ( self.error == None ):
-                  for line in self.__parseDCubeRTTCfg():
-                        self.command.append( line ) 
-            msg = "command to run is:\n"+20*"-"+"\n"
-            for line in self.command:
-                  msg += line
-            msg += "\n"+20*"-"+"\n"
-            self.debug( msg )
-            self.debug( "DCubeRunner is ready to work...")
-            
-      ## run method
-      # @param self "Me, myself and Irene"
-      def run( self ):
-            if ( self.error == None ):
-                  try:
-                        from ShellCommand import ShellCommand
-                        # FIXME - self.logger could be just python False...
-                        sc = ShellCommand( self.command,
-                                           self.logger ) 
-                        replay = sc.getReply()
-                        reLine = re.compile("OVERALL STATISTICS STATUS")
-                        status = 1
-                        for line in replay:
-                              for apiClass in self.__reAPIClasses:
-                                    if ( apiClass.search( line ) ):
-                                          self.debug( line )
-                                          if ( reLine.search( line ) ):
-                                                if ( "OK" in line ): status = 0
-                                    
-                        return status
-                  except ImportError: 
-                        self.info("No ShellCommand in PYTHONPATH! 'No Cin-Cin, no run...'")
-                        return 1
-            else:
-                  self.info(self.error)
-                  return 1
-
-      ## XML parsing of DCubeForRTTCfg.xml 
-      # @param self "Me, myself and Irene"
-      def __parseDCubeRTTCfg( self ):
-            self.debug( "parsing DCubeRunner configuration file %s" % self.DCubeForRTTCfg )
-            try:
-                  xmldoc = minidom.parse( self.DCubeForRTTCfg )   
-
-                  self.bin = self.getText( xmldoc.getElementsByTagName("bin")[0].childNodes )                   
-                  self.debug( "[14] will use binary version %s" % self.bin )
-
-                  self.root = self.getText( xmldoc.getElementsByTagName("root")[0].childNodes)
-                  self.rootVersion = xmldoc.getElementsByTagName("root")[0].getAttribute("version")
-                  rootBasePath = self.root + os.sep + self.rootVersion + os.sep + self.bin + os.sep + "root"
-            
-                  self.debug( "[15] will use root version %s installed in %s" % ( self.rootVersion , rootBasePath ) )
-                  
-                  self.python = self.getText( xmldoc.getElementsByTagName("python")[0].childNodes )
-                  self.pythonVersion = xmldoc.getElementsByTagName("python")[0].getAttribute("version")
-                  
-                  pythonBasePath = self.python + os.sep + self.pythonVersion + os.sep + self.bin  
-                  self.debug( "[16] will use python version %s installed in %s" % ( self.pythonVersion, pythonBasePath ) )
-            
-                  self.refBaseDir = self.getText( xmldoc.getElementsByTagName("ref_dir")[0].childNodes )
-                  self.debug( "[17] will use '%s' as base reference directory" % self.refBaseDir )
-
-                  self.dcubeServer = self.getText( xmldoc.getElementsByTagName("server")[0].childNodes ) 
-                  self.debug( "[18] will use DCubeServer installed in %s" % self.dcubeServer )
-         
-            except:
-                  self.error("error when parsing DCubeRunner configuration file, no run will be preformed")
-                  return [ "" ]
-
-            self.debug("constructing command to run DCube...")
-
-            out  = [ "export ROOTSYS=" + rootBasePath + "\n" ]
-            out.append( "export LD_LIBRARY_PATH=${ROOTSYS}/lib:" + pythonBasePath + "/lib\n" )
-            out.append( "export PATH=" + pythonBasePath + "/bin:${PATH}\n" )
-            out.append( "export PYTHONPATH=" + self.dcubeClient + ":" + rootBasePath + "/lib:" + pythonBasePath +  "/lib/python2.5\n\n" )
-            
-            
-            # changing the outputPath 
-            monPath = self.outputPath 
-
-            self.outputPath = self.outputPath + os.sep +  "DCube-" + self.refFile
-
-            # creating DCube subdirectory
-            out.append("mkdir -p " + self.outputPath + "\n")
-
-            dcubePath = self.dcubeClient
-                
-            dcubeConvert = os.path.join( dcubePath, "./dcubeConvert.py")
-
-            run = os.path.join(dcubePath,"./dcube.py")
-            run += " -p " # generate plot
-            run += " --branch " + self.nightly # set branch
-            run += " --install " + self.install # set install
-            run += " --cmtconfig " + self.cmtconfig # set cmtconfig
-            run += " --project " + self.project # set project
-            # set jobId
-            if ( self.jobId ): 
-                  run += " --jobId " + self.jobId 
-      
-                  
-            run += " -l " + self.outputPath + os.sep + self.monFile + ".dcube.log" # save log to monitored.dcube.log
-            run += " -s " + self.dcubeServer  # path to the DCubeServer installation
-            run += " -x " + self.outputPath + os.sep + self.monFile + ".dcube.xml " # save output to monitored.dcube.xml
-                 
-            # dangerous hacking at the moment... 
-            if ( os.path.isabs( self.cfgFile ) or "DUMMY-TEST" in self.cfgFile ):
-                  run += " -c " + self.cfgFile 
-            else: 
-                  # checking cfg file in reference dir
-                  refCfg = "/afs/cern.ch/atlas/project/RTT/Results/reference/" + self.cvsPath + os.sep + self.cfgFile 
-                  self.debug("will check the timestamp of DCube configuration file (reference volume): " + refCfg )
-                  refCfgStat = None
-
-                  if ( os.path.exists(refCfg) ):
-                        refCfgStat = os.stat(refCfg)
-                        self.debug("File %s modified %d" %( refCfg, refCfgStat[8]) )
-                  else:
-                        self.debug("DCube configuration file %s not found on 'reference' volume" % self.cfgFile )
-                        
-                        # checking cfg file in cvs co dir 
-                        
-                  cvsCfg = self.cmtPath + os.sep + self.cvsPath.lstrip("offline") + "/test/" + self.cfgFile
-                  self.debug("will check the timestamp of DCube configuration file (CVS checkout volume):" + cvsCfg)
-                  cvsCfgStat = None 
-                  if ( os.path.exists(cvsCfg) ) :
-                        cvsCfgStat = os.stat( cvsCfg )
-                        self.debug("File %s modified %d" %( cvsCfg, cvsCfgStat[8]) )
-                  else:
-                        self.debug("DCube configuration file %s not found in CVS checkout directory" % self.cfgFile )
-
-                        
-                  # choose configuration file based on latest timestamp 
-                  config = None
-
-                  # both are present, check time stamps
-                  if ( None not in ( refCfgStat, cvsCfgStat) ):
-                        # install area newer
-                        if ( refCfgStat[8] <= cvsCfgStat[8] ):
-                              config = cvsCfg
-                        else: 
-                              config = refCfg
-
-                  # reference is present
-                  elif ( None != refCfgStat ):
-                        config = refCfg
-                  # install area is present
-                  elif ( None != cvsCfgStat ):
-                        config = cvsCfg
-                  # both are absent  
-                  else:
-                        pass
-
-                        
-                  if ( config != None ):
-                        run += " -c " + config # add configuration file 
-                  else:
-                        self.debug("DCube configuration file %s not found in ASF reference or InstallArea directories" % self.cfgFile )
-                        return [ "" ]
-                  
-            if ( self.refFile != None ):
-                  run += " -r " + os.path.join( os.path.join(self.refBaseDir , self.cvsPath) , self.refFile ) # overwrite reference
-                  run += " " + os.path.join( monPath , self.monFile)  # add monitored file at the end 
-                  
-
-            if ( "DUMMY-TEST" in self.cfgFile ):
-                  dcubeConvert += " %s " % os.path.abspath(self.cfgFile)
-            else:
-                  dcubeConvert += " %s " % os.path.abspath(config)
-
-            dcubeConvert += " %s\n"  % os.path.join( os.path.join(self.refBaseDir , self.cvsPath) , self.refFile )  
-            out.append( dcubeConvert )
-            out.append( run + "\n")
-                  
-            return out
-            
-
-      ## logger level INFO
-      # @param self "Me, myself and Irene"
-      # @param msg message
-      def info( self, msg ):
-            if ( bool(self.logger) ):
-                  self.logger.info( str(msg) )
-            else:
-                  print "DCubeRunner INFO  " + str(msg) 
-
-      ## logger level DEBUG
-      # @param self "Me, myself and Irene"
-      # @param msg message 
-      def debug( self, msg ):
-            if ( bool(self.logger) ):
-                  self.logger.debug( str( msg) )
-            else:
-                  print "DCubeRunner DEBUG " + str(msg)
-
-      
-      ## get CDATA section
-      # @param self "me, myself and Irene"
-      # @param nodelist list of XML DOM Nodes
-      def getText( self, nodelist ):
-        out = ""
-        for node in nodelist:
-            if node.nodeType == node.TEXT_NODE:
-                out = out + node.data
-        return out 
-
-      ## split path
-      # @param self "Me, myself and Irene"
-      # @param path path to split
-      # @param rest splitted tail
-      def pathSplit(self, path, rest=[] ):
-            ( head , tail ) = os.path.split( path )
-            if len(head) < 1: return [tail]+ rest
-            if len(tail) < 1: return [head]+ rest
-            return self.pathSplit( head, [tail] + rest )
-
-      ## find common names in path
-      # @param self "Me, myself and Irene"
-      # @param l1 first list with splitted path
-      # @param l2 second  lust with splitted path
-      # @param common list of common dirnames
-      def commonPath(self,  l1, l2, common=[] ):
-            if len(l1) < 1: return (common, l1, l2)
-            if len(l2) < 1: return (common, l1, l2)
-            if l1[0] != l2[0]: return (common, l1, l2)
-            return self.commonPath( l1[1:], l2[1:], common+[l1[0]])
-
-      ## finds relative path between p1 and p2
-      # @param self "Me, myself and Irene"
-      # @param p1 first path
-      # @param p2 second path
-      def relPath(self, p1, p2):
-            ( common, l1, l2 ) = self.commonPath( self.pathSplit(p1), self.pathSplit(p2))
-            p = []
-            if len(l1) > 0:
-                  p = [ '../' * len(l1) ]
-            p = p + l2
-            return os.path.join( *p )
-
-
-## dummy test running
-if __name__ == "__main__":
-      print "running DCubeRunner..."
-      testCfg = { "DCubeCfg"         : "DUMMY-TEST-DCUBE-CFG.XML",
-                  "DCubeMon"         : "monitored.root",
-                  "DCubeRef"         : "reference.root",
-                  "DCubeJobId"       : "test_01",
-                  "JobDescriptor"    : None }
-      dc = DCubeRunner( testCfg )
-      dc.run()
-
-      
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeTester.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeTester.py
deleted file mode 100644
index 4293f5e3af413623c51658e430c72140cc63e958..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeTester.py
+++ /dev/null
@@ -1,705 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file DCubeTester.py
-# @author Krzysztof Daniel Ciba (Krzyszotf.Ciba@NOSPAMagh.edu.pl)
-# @brief implementation of DCubeTester and test_DCubeTester classes
-
-import sys
-from DCubeUtils import DCubeObject, DCubeException
-from DCubePlotter import DCubePlotter
-import ROOT
-import unittest
-
-##
-# @class DCubeTester
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief perform statictics tests and plotting  
-class DCubeTester( DCubeObject ):
-
-    ## XML DOM Document instance
-    xmldoc = None
-
-    ## XML DOM Element instance 
-    node = None
-
-    ## handle for monitored root object
-    mon = None
-
-    ## handle for reference root object
-    ref = None
-
-    ## errors counters
-    errors = { }
-
-    ## intarnal histogram counter
-    __nbObjs = 0
-    
-    ## statistics summary table
-    sumTable = { "KS"   : { "OK"   : 0,
-                            "WARN" : 0,
-                            "FAIL" : 0 },
-                 "chi2" : { "OK"   : 0,
-                            "WARN" : 0,
-                            "FAIL" : 0 },
-                 "bbb"  : { "OK"   : 0,
-                            "WARN" : 0,
-                            "FAIL" : 0 },
-                 "meanY": { "OK"   : 0,
-                            "WARN" : 0,
-                            "FAIL" : 0 }}
-
-
-    ## chi2 error communicates
-    chi2igood = { 0 : "no problems",
-                  1 : "there is bin in mon hist with low then 1 exp number of event",
-                  2 : "there is bin in ref hist with low then 1 exp number of event" }
-
-    
-    ## summary status
-    __status = ""
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param xmldoc XML DOM Document instance
-    def __init__( self, xmldoc, parsed ):
-        
-        super( DCubeTester, self ).__init__( self )
-        self.xmldoc = xmldoc
-        self.opts, self.args = parsed
-        self.tests = { "KS"   : self.__testKS,
-                       "bbb"  : self.__testBBB,
-                       "chi2" : self.__testChi2,
-                       "meanY": self.__testMeanY }
-       
-
-        if ( self.opts.makeplots ):
-            self.info("will produce plot files")
-            self.__plotter = DCubePlotter( xmldoc, parsed )
-        else:
-            self.warn("making of plots disabled")
-            
-    ## 
-    # @param self "Me, myself and Irene"
-    # @param node DOM XML  node 
-    # @param mon monitored ROOT object
-    # @param ref reference ROOT object
-    # @return modified XML node 
-    def test( self, node, mon=None, ref=None ):
-
-        self.__nbObjs += 1
-
-        self.node = self.mon = self.ref = None
-
-        self.node = node
-        self.parent = self.node.parentNode
-
-        self.mon = mon
-        self.ref = ref
-
-        # mon exists?
-        if ( not self.mon ):
-            status = "FAIL;monitored not found"
-            self.node.setAttribute( "status", status )
-            if status not in self.errors.keys():
-                self.errors[ status ] = 1
-            else:
-                self.errors[ status ] = self.errors[ status ] + 1
-            return "FAIL"
-
-
-        cl = self.mon.Class().GetName() 
-        isHist = False
-        if ( cl[:2] in ( "TH", "TP") ): isHist = True
-       
-        # dimension OK?
-        if ( isHist and self.mon.GetDimension() > 2 ):
-            status = "FAIL;unsupported object, dimension bigger than 2"
-            self.node.setAttribute( "status", status )
-            if ( status not in self.errors.keys() ):
-                self.errors[ status ] = 1
-            else:
-                self.errors[ status ] = self.errors[ status ] + 1
-            return "FAIL"
-
-        # reference exists?
-        if ( not self.ref ):
-            status = "WARN;reference histogram not found"
-            self.node.setAttribute( "status", status )
-            if ( status not in self.errors.keys() ):
-                self.errors[ status ] = 1
-            else:
-                self.errors[ status ] = self.errors[ status ] + 1
-            
-    
-        if ( self.mon and self.ref ):
-
-            # same class?
-            monClassName = self.mon.Class().GetName()
-            refClassName = self.ref.Class().GetName()
-           
-            if ( monClassName != refClassName ):
-                status = "FAIL;different root types"
-                self.node.setAttribute( "status", status + " mon=%s ref=%s" % ( monClassName,
-                                                                                refClassName ) )
-                if ( status not in self.errors.keys() ):
-                    self.errors[ status ] = 1
-                else:
-                    self.errors[ status ] = self.errors[ status ] + 1 
-                return "FAIL"
-
-        if ( isHist and self.mon and self.ref ):
-            # same binnig?
-            self.monBins = [ self.mon.GetNbinsX(), self.mon.GetNbinsY() ]
-            
-            self.refBins = [ self.ref.GetNbinsX(), self.ref.GetNbinsY() ]
-
-            monBins = [ "%s=%d" % ( x, y) for ( x, y) in zip(["x", "y"], self.monBins ) ]
-            refBins = [ "%s=%d" % ( x, y) for ( x, y) in zip(["x", "y"], self.refBins ) ]
-                             
-            diffBins = [ "mon %s ref %s" % (x, y) for (x,y) in zip(monBins,refBins) if x != y ]
-
-            if ( diffBins  ):
-                status = "FAIL;different binnig"
-                self.node.setAttribute( "status", status + " %s" % " ".join(diffBins) )
-                if ( status not in self.errors.keys() ):
-                    self.errors[ status ] = 1
-                else:
-                    self.errors[ status ] = self.errors[ status ] + 1
-                return "FAIL" 
-            
-        status = "OK"
-                    
-        if ( isHist ):
-            status = self.__grabHistogramStat()
-        else:
-            status = self.__grabGraphStat() 
-
-        plotStatus = "OK"
-        if ( self.opts.makeplots ):
-            plotStatus = self.__makePlots() 
-        else:
-            plotStatus = "WARN;plots not reqested"
-        self.node.setAttribute( "plots", plotStatus)
-
-        return status
-        
-
-    ## grab statistic info for TGraphXXX  
-    # @param self "Me, myself and Irene"
-    def __grabGraphStat( self ):
-        statNode = self.xmldoc.createElement( "stat" )
-        self.node.appendChild( statNode )
-
-        nbPointsNode = self.xmldoc.createElement( "points" )
-        nbPointsCData = self.xmldoc.createTextNode( "%d" % self.mon.GetN() )
-        if ( self.ref ):
-            nbPointsNode.setAttribute( "ref" , "%d" % self.ref.GetN() )
-        nbPointsNode.appendChild( nbPointsCData )
-
-        meanNode = self.xmldoc.createElement( "mean" )
-        meanCData = self.xmldoc.createTextNode( "%4.3f" % self.mon.GetMean() )
-        if ( self.ref ):
-            meanNode.setAttribute( "ref" , "%4.3f" % self.ref.GetMean() )
-        meanNode.appendChild( meanCData )
-
-        rmsNode = self.xmldoc.createElement( "rms" )
-        rmsCData = self.xmldoc.createTextNode( "%d" % self.mon.GetRMS() )
-        if ( self.ref ):
-            rmsNode.setAttribute( "ref" , "%d" % self.ref.GetRMS() )
-        rmsNode.appendChild( rmsCData )
-
-        statNode.appendChild( nbPointsNode )
-        statNode.appendChild( meanNode )
-        statNode.appendChild( rmsNode )
-
-        return "OK"
-        
-        
-        
-    ## grab basic statistics - nb of entries, mean and RMS
-    # @param self "Me, myself and Irene"
-    def __grabHistogramStat( self ):
-
-        statNode = self.xmldoc.createElement( "stat" )
-
-        self.node.appendChild( statNode )
-
-        entriesNode = self.xmldoc.createElement("entries")
-        entriesCDataNode = self.xmldoc.createTextNode( str( self.mon.GetEntries() ) )
-        if ( self.ref ):
-            entriesNode.setAttribute( "ref", str( self.ref.GetEntries() ) )
-        entriesNode.appendChild( entriesCDataNode )
-
-        statNode.appendChild( entriesNode )
-
-        dim = self.mon.GetDimension() 
-
-        ### To dump MeanY values
-        if ( "TProfile" in self.mon.Class().GetName() ): dim = 2
-
-        # store under and overflows 
-        monU = []
-        monO = []
-        refU = []
-        refO = []
-
-        if ( dim == 1 ):
-            monU.append( self.mon.GetBinContent(0) )
-            monO.append( self.mon.GetBinContent( self.mon.GetNbinsX()+1 ) )
-            if ( self.ref ):
-                refU.append( self.ref.GetBinContent(0) )
-                refO.append( self.ref.GetBinContent( self.ref.GetNbinsX()+1 ) )
-        elif ( dim == 2 ):
-            monU.append( self.mon.GetBinContent( 0, 0 ) )
-            monO.append( self.mon.GetBinContent( self.mon.GetNbinsX()+1, 0 ) )
-            monU.append( self.mon.GetBinContent( 0,  self.mon.GetNbinsY()+1) )
-            monO.append( self.mon.GetBinContent( self.mon.GetNbinsX()+1, self.mon.GetNbinsY()+1 ) )  
-            if ( self.ref ):
-                refU.append( self.ref.GetBinContent( 0, 0 ) )
-                refO.append( self.ref.GetBinContent( self.ref.GetNbinsX()+1, 0 ) )
-                refU.append( self.ref.GetBinContent( 0,  self.ref.GetNbinsY()+1) )
-                refO.append( self.ref.GetBinContent( self.ref.GetNbinsX()+1, self.ref.GetNbinsY()+1 ) )  
-
-        underflowNode = self.xmldoc.createElement( "underflow" )
-        underflowCData = self.xmldoc.createTextNode( "%d" % sum( monU ) )
-        if ( self.ref ):
-            underflowNode.setAttribute( "ref", "%d" % sum( refU) )
-        underflowNode.appendChild( underflowCData )
-
-        overflowNode = self.xmldoc.createElement( "overflow" )
-        overflowCData = self.xmldoc.createTextNode( "%d" % sum( monO ) )
-        if ( self.ref ):
-            overflowNode.setAttribute( "ref", "%d" % sum( refO ) )
-        overflowNode.appendChild( overflowCData )
-
-        statNode.appendChild( underflowNode )
-        statNode.appendChild( overflowNode )
-        
-
-        dimName = [ "x", "y", "z" ]
-        for i in range( 1, dim+1 ):
-            self.debug( "gathering statistics along %s axis" % dimName[i-1] )
-
-            dimNode = self.xmldoc.createElement( "dim" )
-            dimNode.setAttribute( "name", dimName[i-1] )
-            dimNode.setAttribute( "bins", "%d" % self.monBins[i-1])
-        
-            underflowNode = self.xmldoc.createElement( "underflow" )
-            underflowCData  = self.xmldoc.createTextNode( "%d" % monU[i-1] )
-            if ( self.ref ):
-                underflowNode.setAttribute( "ref", "%d" % refU[i-1] )
-            underflowNode.appendChild( underflowCData )
-
-            dimNode.appendChild( underflowNode )
-        
-            overflowNode = self.xmldoc.createElement( "overflow" )
-            overflowCData  = self.xmldoc.createTextNode( "%d" % monO[i-1] )
-            if ( self.ref ):
-                overflowNode.setAttribute( "ref", "%d" % refO[i-1] )
-            overflowNode.appendChild( overflowCData )
-
-            dimNode.appendChild( overflowNode )
-            
-            meanNode = self.xmldoc.createElement( "mean" )
-            if ( self.ref ):
-                meanNode.setAttribute( "ref", "%3.2e" % self.ref.GetMean(i) )
-
-            meanCData = self.xmldoc.createTextNode( "%3.2e" % self.mon.GetMean( i ) )
-            meanNode.appendChild( meanCData )
-            
-            meanErrorNode = self.xmldoc.createElement( "mean_unc" )
-            if ( self.ref ):
-                meanErrorNode.setAttribute( "ref", "%3.2e" % self.ref.GetMeanError(i) )
-
-            meanErrorCData = self.xmldoc.createTextNode( "%3.2e" % self.mon.GetMeanError( i ) )
-            meanErrorNode.appendChild( meanErrorCData )
-            
-            rmsNode = self.xmldoc.createElement( "rms" )
-            if ( self.ref ):
-                rmsNode.setAttribute( "ref", "%3.2e" % self.ref.GetRMS(i) )
-            rmsCData = self.xmldoc.createTextNode( "%3.2e" % self.mon.GetRMS(i) )
-            rmsNode.appendChild( rmsCData )
-        
-            rmsErrorNode = self.xmldoc.createElement( "rms_unc" )
-            if ( self.ref ):
-                rmsErrorNode.setAttribute( "ref", "%3.2e" % self.ref.GetRMSError(i) )
-            rmsErrorCData = self.xmldoc.createTextNode( "%3.2e" % self.mon.GetRMSError(i))
-            rmsErrorNode.appendChild( rmsErrorCData )
-
-            dimNode.appendChild( meanNode )
-            dimNode.appendChild( meanErrorNode )
-            dimNode.appendChild( rmsNode )
-            dimNode.appendChild( rmsErrorNode )
-
-            statNode.appendChild( dimNode )
-
-        status = [ ]
-        if ( "TProfile" in self.mon.Class().GetName() ):
-            #self.warn( "will skip statistics testing for TProfile objects...")
-            self.info( "will run Mean test for TProfile objects...")
-            status.append( self.__testMeanY.__call__( statNode ) )
-
-        else:
-
-            tests = self.node.getAttribute("tests").split(",")
-            if ( "all" in tests ): tests = [ "KS", "chi2", "bbb" ]
-
-            if ( None not in ( self.mon, self.ref )  ):
-                for test in tests:
-                    status.append( self.tests[ test ].__call__( statNode ) )
-                    
-        statusAttr = "OK"
-        if ( "FAIL" in status ): statusAttr = "FAIL"
-        elif ( "WARN" in status ): statusAttr = "WARN"
-        
-        self.node.setAttribute( "stest", statusAttr )
-
-        return statusAttr
-                    
-    ## perform @f$\chi^2@f$ test
-    # @param self "Me, myself and Irene"
-    # @param statNode <stat> element
-    def __testChi2( self, statNode ):
-
-        chi2 = ROOT.Double( 0.0 )
-        igood = ROOT.Long( 0 )
-        ndf = ROOT.Long( 0 )
-
-        pval = self.mon.Chi2TestX( self.ref, chi2, ndf, igood, "UUDNORM")
-
-        self.debug( "*** Pearson's chi2 *** (UUDNORM) p-value= %4.3f" % pval )
-        ig = "*** Pearson's chi2 *** igood = %d, %s" % ( igood, self.chi2igood[igood] )
-        if ( igood == 0 ):
-            self.debug( ig )
-        else:
-            self.warn( ig )
-
-        status = self.__getTestStatus( pval ) 
-        self.sumTable["chi2"][status] = self.sumTable["chi2"][status] + 1   
-
-        if ( ndf != 0 ):
-            self.info( "*** Pearson's chi2 *** chi2 (/ndf) = %4.3f (/%d = %4.3f) status=%s" % ( chi2,
-                                                                                                ndf,
-                                                                                                chi2/ndf,
-                                                                                                status ) )
-        else:
-            self.info( "*** Pearson's chi2 *** chi2 = %4.3f ndf = %d status=%s" % ( chi2,
-                                                                                    ndf,
-                                                                                    status ) )
-
-        pValueNode = self.xmldoc.createElement( "pvalue" )
-        pValueCData = self.xmldoc.createTextNode( "%4.3f" % pval )
-
-        pValueNode.setAttribute( "test", "chi2" )
-        pValueNode.setAttribute( "status", status )
-       
-        counter = self.parent.getAttribute( "chi2" + status )
-        if ( not counter ): counter = "0"
-        self.parent.setAttribute( "chi2" + status, "%d" % ( int(counter) + 1 )  )
-
-        pValueNode.appendChild( pValueCData )
-
-        statNode.appendChild( pValueNode )
-
-
-        return status
-
-    ## perform Kolmogorov-Smirnoff test
-    # @param self "Me, myself and Irene"
-    # @param statNode <stat> element
-    def __testKS( self, statNode ):
-
-        pval = self.mon.KolmogorovTest( self.ref, "NDX" )
-        
-        status = self.__getTestStatus( pval ) 
-        self.sumTable["KS"][status] = self.sumTable["KS"][status] + 1     
-
-        self.info( "*** Kolmogorov-Smirnoff *** (NDX) p-value=%4.3f status=%s" % ( pval, status ) )
-
-        pValueNode = self.xmldoc.createElement( "pvalue" )
-        pValueCData = self.xmldoc.createTextNode( "%4.3f" % pval )
-
-        pValueNode.setAttribute( "test", "KS" )
-        pValueNode.setAttribute( "status", status )
-
-        counter = self.parent.getAttribute( "KS" + status )
-        if ( not counter ): counter = "0"
-        self.parent.setAttribute( "KS" + status, "%d" % ( int(counter) + 1 )  )
-
-        pValueNode.appendChild( pValueCData )
-
-        statNode.appendChild( pValueNode )
-        return status
-        
-
-    ## perform "bin-by-bin" statistic test
-    # @param self "Me, myself and Irene"
-    # @param statNode <stat> element
-    def __testBBB( self, statNode ):
-       
-        nbBins = 0.0
-        nbBinsSame = 0.0
-
-        binX = self.mon.GetNbinsX()
-        binY = self.mon.GetNbinsY()
-
-        for i in range( binX+1 ):
-            for j in range( binY +1 ):
-                mon = self.mon.GetBinContent( i, j )
-                ref = self.ref.GetBinContent( i, j )
-                if ( mon or ref ):
-                    nbBins += 1
-                    if ( mon == ref ):
-                        nbBinsSame += 1
-
-        self.debug("*** bin-by-bin *** all none-empty bins=%d all equal non-empty bins=%d" % ( nbBins, nbBinsSame ) )
-
-        pval = 0.0
-        if ( nbBins ):
-            pval = nbBinsSame / nbBins  
-            self.debug("*** bin-by-bin *** p-value=%4.3f" % pval )
-        elif ( not nbBinsSame ):
-            pval = 1.0
-            self.warn( "*** bin-by-bin *** both histograms are empty!" )
-            self.debug( "*** bin-by-bin *** p-value=%4.3f" % pval )
-        else:
-            self.warn( "*** bin-by-bin *** reference histogram is empty, while monitored has some entries" )
-            self.debug( "*** bin-by-bin *** test failed" )
-        
-        status = self.__getTestStatus( pval )
-        self.info("*** bin-by-bin *** p-value=%4.3f status=%s" % ( pval, status ) )
-
-        self.sumTable["bbb"][status] = self.sumTable["bbb"][status] + 1     
-
-        pValueNode = self.xmldoc.createElement( "pvalue" )
-        pValueCData = self.xmldoc.createTextNode( "%4.3f" % pval )
-
-        pValueNode.setAttribute( "test", "bbb" )
-        pValueNode.setAttribute( "status", status )
-
-        counter = self.parent.getAttribute( "bbb" + status )
-        if ( not counter ): counter = "0"
-        self.parent.setAttribute( "bbb" + status, "%d" % ( int(counter) + 1 )  )
-
-        pValueNode.appendChild( pValueCData )
-
-        statNode.appendChild( pValueNode )
-
-        return status
-
-
-    ## perform "diff(MeanY)" test for TProfile histos
-    # @param self "Me, myself and Irene"
-    # @param statNode <stat> element
-    def __testMeanY( self, statNode ):
-
-        avgEffmon = self.mon.GetMean( 2 ) * 100
-        avgEffref = self.ref.GetMean( 2 ) * 100
-
-        self.debug("*** MeanY Test *** refMean=%4.3f and monMean=%4.3f" % ( avgEffref, avgEffmon ) )
-        
-        pval = abs(avgEffmon - avgEffref)
-
-        status = self.__getMeanTestStatus( pval )
-        self.info("*** Mean Test *** p-value=%4.3f status=%s" % ( pval, status ) )
-
-        self.sumTable["meanY"][status] = self.sumTable["meanY"][status] + 1     
-
-        pValueNode = self.xmldoc.createElement( "pvalue" )
-        pValueCData = self.xmldoc.createTextNode( "%4.3f" % pval )
-
-        pValueNode.setAttribute( "test", "meanY" )
-        pValueNode.setAttribute( "status", status )
-
-        counter = self.parent.getAttribute( "meanY" + status )
-        if ( not counter ): counter = "0"
-        self.parent.setAttribute( "meanY" + status, "%d" % ( int(counter) + 1 )  )
-
-        pValueNode.appendChild( pValueCData )
-
-        statNode.appendChild( pValueNode )
-
-        return status
-    
-    
-    ## get test status for given pvalue 
-    # @param self "Me, myself and Irene"
-    # @param pval p-value from test
-    def __getTestStatus( self, pval ):
-        if ( ( pval < 0.0 or pval > 1.0 ) or  
-             ( pval <= self.opts.pfail ) ): return "FAIL"
-        if ( pval > self.opts.pwarn ): return "OK"
-        return "WARN"
-
-
-    ## get test status for given pvalue 
-    # @param self "Me, myself and Irene"
-    # @param pval p-value from test
-    def __getMeanTestStatus( self, pval ):
-        if ( pval > 1.0 and pval <= 5.0 ): return "WARN"
-        if ( pval <= 1.0  ): return "OK"
-        return "FAIL"
-
-
-    ## make plots
-    # @param self "Me, myself and Irene"
-    def __makePlots( self ):
-        if ( self.__plotter ):
-            status = self.__plotter.plot( self.node, self.mon, self.ref )
-            return status
-    
-
-    ## string representation of DCubeTester
-    # @param self "Me, myself and Irene" 
-    def __str__( self ):
-        out  = "*"*61 + "\n"
-        out += "* RUN SUMMARY\n"
-        out += "*"*61 + "\n"
-        out += "* objects processed = %d\n" % self.__nbObjs
-        out += "* STATISTICS TESTS TABLE\n"
-        out += "* " + "-"*45 + "\n"
-        out += "* | %-8s | %8s | %8s | %8s |\n" % ( "test", "OK", "WARN", "FAIL") 
-        self.allGOOD = 0
-        self.allWARN = 0
-        self.allFAIL = 0
-        for key, value  in self.sumTable.iteritems():
-            nbGOOD = value["OK"]
-            nbWARN = value["WARN"]
-            nbFAIL = value["FAIL"]
-
-            self.allGOOD += nbGOOD
-            self.allWARN += nbWARN
-            self.allFAIL += nbFAIL
-            out += "* " + "-"*45+"\n"
-            out += "* | %-8s | %8d | %8d | %8d |\n" % ( key, nbGOOD, nbWARN, nbFAIL )
-            
-        out += "* " + "-"*45+"\n"
-
-        
-        out += "* | %-8s | %8d | %8d | %8d |\n" % ( "all", self.allGOOD, self.allWARN, self.allFAIL )
-        out += "* " + "-"*45+"\n"
-
-        self.fracGOOD = 0.0
-        self.fracWARN = 0.0
-        self.fracFAIL = 0.0
-        all = float( self.allGOOD + self.allWARN + self.allFAIL ) 
-        if ( all ):
-            self.fracGOOD = 100 * float( self.allGOOD ) / all 
-            self.fracWARN = 100 * float( self.allWARN ) / all 
-            self.fracFAIL = 100 * float( self.allFAIL ) / all
-        out += "* | %%        |    %04.2f |    %04.2f |    %04.2f |\n" % ( self.fracGOOD, self.fracWARN, self.fracFAIL )
-        out += "* " + "-"*45+"\n"
-
-        out += "*"*61+"\n"
-        self.nbErrors = sum( self.errors.itervalues() )
-        out += "* WARNINGS = %3d\n" % self.nbErrors
-        i = 1
-        for key, value in self.errors.iteritems( ):
-            sev, what = key.split(";")
-            
-            out += "* [%02d] %4s %-30s - occured %d " %  (i, sev, what, value )
-            if ( value == 1 ): out += "time\n"
-            else: out += "times\n"
-            i += 1
-
-        out += "*"*61+"\n"
-        
-        self.__status = ""
-        
-        if ( self.allFAIL != 0 ):
-            self.__status = "FAIL"
-        elif ( self.allWARN != 0 or self.nbErrors != 0 ):
-            self.__status = "WARN"
-        else:
-            self.__status = "OK"
-
-        out += "* OVERALL STATISTICS STATUS: %-4s\n" % self.__status
-                    
-        out += "*"*61
-        
-        return out
-
-    ## put summary to the logger and create summary node 
-    # @param self "Me, myself and Irene"
-    # @return summary XML Element
-    def summary( self ):
-        
-        for line in str(self).split("\n"): 
-            self.info( line )
-
-        summaryNode = self.xmldoc.createElement( "summary" )
-        summaryNode.setAttribute( "status" , self.__status )
-        summaryNode.setAttribute( "objs", "%d" % self.__nbObjs )
-        summaryNode.setAttribute( "errors", "%d" % self.nbErrors  )
-
-        testTable = self.xmldoc.createElement( "table" )
-        testTable.appendChild( self.__sumTableRow( [ "test", "OK", "WARN", "FAIL" ] ) )
-        for test, results in self.sumTable.iteritems(): 
-            testTable.appendChild( self.__sumTableRow( [ test,
-                                                         results["OK"],
-                                                         results["WARN"],
-                                                         results["FAIL"] ] ) )
-            
-        testTable.appendChild( self.__sumTableRow( ["sum", self.allGOOD, self.allWARN, self.allFAIL ] ) )
-        testTable.appendChild( self.__sumTableRow( ["fraction",
-                                                    "%4.2f" % self.fracGOOD,
-                                                    "%4.2f" % self.fracWARN,
-                                                    "%4.2f" % self.fracFAIL ] ) )
-        
-        summaryNode.appendChild( testTable )
-
-        errorsNode = self.xmldoc.createElement( "errors" )
-        errorsNode.setAttribute( "nb", "%d" % self.nbErrors )
-
-        for error,times in self.errors.iteritems():
-            error = error.replace(";", " ")
-            errorNode = self.xmldoc.createElement( "error" )
-            errorNode.setAttribute( "times", "%d" % times )
-            errorNode.setAttribute( "what", error )
-            errorsNode.appendChild( errorNode )
-            
-        summaryNode.appendChild( errorsNode )
-        
-        return summaryNode 
-
-    ## produce summary table row 
-    # @param self "Me, myself and Irene"
-    def __sumTableRow( self, what ):
-        row = self.xmldoc.createElement( "tr" )
-        for item in what:
-            cellNode = self.xmldoc.createElement("td")
-            cellCData = self.xmldoc.createTextNode( str(item) )
-            cellNode.appendChild( cellCData )
-            row.appendChild( cellNode )
-        return row
-
-    ## return overall summary as string
-    # @param self "Me, myself and Irene"
-    def status( self ):
-        return self.__status
-        
-        
-##
-# @class test_DCubeTester
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl) 
-# @brief test case for DCubeTester class
-class test_DCubeTester( unittest.TestCase ):
-
-    ## setup test cases
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-        pass
-
-    ## ctor
-    # @param self "Me, myself and Irene"
-    def test_01_ctor( self ):
-        pass
-
-
-## test case execution
-if __name__ == "__main__":
-    pass
-    
-
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeUtils.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeUtils.py
deleted file mode 100644
index efe38188e97c8053a81e95dd029d4c2d1c42b8ed..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeUtils.py
+++ /dev/null
@@ -1,423 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-## 
-# @file DCubeUtils.py
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief Utility classes for DCubeApp
-#
-
-import xml.dom.minidom
-import logging
-import os
-import sys
-from types import *
-import unittest
-
-##
-# @class Singleton 
-# @author Krzysztof Daniel Ciba (Kerzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief Forcing singleton type for python class instance. 
-class DCubeSingleton( type ):
-
-    ## c'tor
-    # @param cls class instance
-    # @param name class name
-    # @param bases base classes
-    # @param dic parameters dictionary
-    def __init__(cls, name, bases, dic):
-        super( DCubeSingleton, cls ).__init__( name, bases, dic )
-        cls.instance = None
-
-    ## call operator
-    # @param cls class instance
-    # @param *args unnamed arguments
-    # @param **kw keyword arguments
-    def __call__( cls, *args, **kw ):
-        if cls.instance is None:
-            cls.instance = super( DCubeSingleton, cls ).__call__( *args, **kw )
-        return cls.instance
-
-##
-# @class DCubeLogger 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAagh.edu.pl)
-# @brief DCube customized logger
-class DCubeLogger( object ):
-
-    ## metaclass type
-    __metaclass__ = DCubeSingleton
-
-    ## reference to original logger instance
-    __log = None
-
-    ## name of the log file 
-    __logFile = None 
-    
-    ## default name of the log file
-    __defaultLogFile = "./dcube.log"
-
-    ## flag to trigger stdout print outs
-    __console = None
-
-    ## logging levels
-    __levels = { "CRITICAL": logging.CRITICAL,
-                 "ERROR": logging.ERROR, 
-                 "WARNING" : logging.WARNING, 
-                 "INFO" : logging.INFO,
-                 "DEBUG" : logging.DEBUG }
-
-    ## c'tor
-    # @brief c'tor
-    # @param self "Me, myself and Irene"
-    # @param fileName name of log file 
-    # @param logName name of logger
-    # @param toConsole flag to trigger stdout printout
-    def __init__( self, fileName=None, logName="DCube", toConsole=True ):
-                
-        if ( fileName == None ): fileName = self.__defaultLogFile
-
-        self.__logFile = os.path.abspath( fileName )
-      
-        logging.basicConfig( level=logging.DEBUG,
-                             format='%(name)-12s %(asctime)s %(levelname)-8s: %(message)s',
-                             filename=fileName,
-                             filemode='w' )
-          
-        self.toConsole( toConsole )
-        
-        self.__log = logging.getLogger( str(logName) )
-
-        ## save stdout and stderr handle
-        #self.__stdout = sys.stdout
-        #self.__stderr = sys.stderr
-
-         
-    ## 
-    # @brief set logging level
-    # @param cls class reference
-    # @param level string with level name 
-    @classmethod
-    def setLevel(cls, level="DEBUG"):
-        if ( level in cls.__levels.keys() ):
-            cls.__log.setLevel( cls.__levels[level] ) 
-
-
-    ## 
-    # @brief log file name getter
-    # @param cls class reference
-    # @return name of log file 
-    def getFileName( self ):
-        return self.__logFile
-
-    ##
-    # @brief logger getter
-    # @param cls class reference
-    # @param caller who call me?
-    @classmethod
-    def getLogger( cls, caller ):
-        logName = caller.__class__.__name__ 
-        cls.__log = logging.getLogger( logName )
-        return cls
-
-    ## add/remove logging to the stdout and stderr
-    # @param cls class reference
-    # @param toConsole bool flag [default True]
-    @classmethod
-    def toConsole( cls, toConsole=True):
-        if ( toConsole ):
-
-            if ( not isinstance(cls.__console, logging.StreamHandler)  ):
-                cls.__console = logging.StreamHandler()
-                cls.__console.setLevel(logging.DEBUG)
-                formatter = logging.Formatter('%(name)-12s: %(levelname)-8s %(message)s')
-                cls.__console.setFormatter(formatter)
-
-            logging.getLogger('').addHandler( cls.__console )
-
-        else:
-            if ( isinstance(cls.__console, logging.StreamHandler)  ):
-                logging.getLogger('').removeHandler( cls.__console )
-            
-    ## 
-    # @brief debug level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def debug( cls, msg ):
-        cls.__log.debug( msg )
-
-    ## 
-    # @brief info level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def info( cls, msg ):
-        cls.__log.info( msg )
-
-    ## 
-    # @brief warning level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def warn( cls, msg ):
-        cls.__log.warning( msg )
-
-    ## 
-    # @brief error level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def error( cls, msg ):
-        cls.__log.error( msg )
-
-    ## 
-    # @brief critical level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def panic( cls, msg ):
-        cls.__log.critical( msg )
-
-    ## 
-    # @brief critical level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def critical( cls, msg ):
-        cls.__log.critical( msg )
-
-    ## 
-    # @brief exception level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def exception( cls, msg ):
-        cls.__log.exception( msg )
-
-    ## 
-    # @brief exception level logger 
-    # @param cls class reference
-    # @param msg message to log
-    @classmethod
-    def epanic( cls, msg ):
-        cls.__log.exception( msg )
-
-
-##
-# @class DCubeObject
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief base class for all DCube classes
-class DCubeObject( DCubeLogger ):
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param caller reference to inherited object 
-    def __init__( self, caller ):
-        ## setting up logger
-        self.__log = DCubeLogger.getLogger( caller )
-
-    ## debug level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def debug( self, msg ):
-        if ( self.__log ):
-            self.__log.getLogger(self).debug( str(msg) )
-        else:
-            print msg
-
-    ## info level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def info( self, msg  ):
-        if ( self.__log ):
-            self.__log.getLogger(self).info( str(msg) )
-        else:
-            print msg
-        
-    ## warning level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def warn( self, msg ):
-        if ( self.__log ):
-            self.__log.getLogger(self).warn( str(msg) )
-        else:
-            print msg
-    
-    ## error level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def error( self, msg ):
-        if ( self.__log ):
-            self.__log.getLogger(self).error( str(msg) )
-        else:
-            print msg
-
-    ## critical level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def panic( self, msg ):
-        if ( self.__log ):
-            self.__log.getLogger(self).critical( str(msg) )
-        else:
-            print msg
-
-    ## exception level logger
-    # @param self "Me, myself and Irene"
-    # @param msg logging string
-    def epanic( self, msg ):
-        if ( self.__log ):
-            self.__log.getLogger(self).epanic( str(msg) )
-        else:
-            print msg
-
-
-##
-# @class DCubeVersion
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief DCube version holder
-class DCubeVersion( object ):
-
-    __project = "DCube"
-    __version = "4.66920"
-    __author = "Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)"
-          
-    ## str operator
-    # @param self "Me, myself and Irene"
-    def __str__( self ):
-        return "%s %s by %s" % ( self.__project, self.__version, self.__author )
-
-    ## DCube version getter
-    # @param self "Me, myself and Irene"
-    def version( self ):
-        return str(self)
-
-    ## python version getter 
-    # @param self "Me, myself and Irene"
-    def python( self ):
-        return "using Python %s" % sys.version.replace("\n", "").strip("\n") 
-
-    ## root version getter
-    # @param self "Me, myself and Irene"
-    def root( self ):
-        out = "using ROOT "
-        try:
-            if ( "ROOT" not in dir() ):
-                import ROOT
-            out += ROOT.gROOT.GetVersion().strip("\n") 
-        except:
-            out += "UNKNOWN"
-        return out
-
-## 
-# @class DCubeException
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAagh.edu.pl)
-# @brief class for DCube triggered execution errors 
-class DCubeException( Exception ):
-    
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param value whatever, who cares, this is python, dumb!  
-    def __init__( self, value ):
-        self.value = value
-   
-    ## exception printing
-    # @param self "Me, myself and Irene"
-    def __str__( self ):
-        return repr( self.value )
-
-
-##
-# @class test_DCubeUtils
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief test case for DCubeUtils module classes
-class test_DCubeUtils( unittest.TestCase ):
-
-    ## test case setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-
-        self.logger = DCubeLogger( "test_DCubeUtils.log", logName="test_DCubeUtils", toConsole=True )
-       
-        self.exception = DCubeException("test exception, don't panic")
-        self.version = DCubeVersion() 
-        self.base = DCubeObject( self )
-        self.log = DCubeLogger.getLogger( self )
-        pass
-
-    ## c'tors
-    # @param self "Me, myself and Irene"
-    def test_01_constructors( self ):
-
-        try:
-            self.exception = DCubeException("test exception, don't panic!")
-            self.version = DCubeVersion() 
-            self.base = DCubeObject( self )
-            self.log = DCubeLogger.getLogger( self )
-        except:
-            pass
- 
-        self.assertEqual( isinstance( self.exception, DCubeException), True )
-        self.assertEqual( isinstance( self.version, DCubeVersion), True )
-        self.assertEqual( isinstance( self.base, DCubeObject), True )
-
-    ## DCubeLogger interface
-    # @param self "me, myself and Irene"
-    def test_02_logger( self ):
-        self.log.debug("debug level text")
-        self.log.info("info level text")
-        self.log.warn("warning level text")
-        self.log.error("error level text")
-        self.log.panic("fatal level text")
-        try:
-            raise NameError, "intentional NameError, don't panic!"
-        except NameError, value:
-            self.log.epanic("exception level text")
-
-        self.log.toConsole(False)
-        self.log.info("this won't be printed to console")
-        self.log.toConsole(True)
-        self.log.info("this will be printed to console")
-
-        self.logger.setLevel("INFO")
-        self.log.debug("this should not be printed")
-        
-        self.logger.setLevel("DEBUG")
-        self.log.debug("while this should be printed")
-
-    ## DCubeException interface
-    # @param self "Me, myslf and Irene"
-    def test_03_exception( self ):
-        try:
-            raise self.exception
-        except DCubeException, value:
-            self.log.epanic(value)
-        
-    ## DCubeObject interface
-    # @param self "Me, myself and Irene"
-    def test_04_DCubeObject( self ):
-        self.base.debug("debug level text")
-        self.base.info("info level text")
-        self.base.warn("warn level text")
-        self.base.error("error level text")
-        self.base.panic("panic level text")
-        try:
-            raise NameError, "intentional NameError, don't panic!"
-        except NameError, value:
-            self.base.epanic("exception level text")
-            
-    ## DCubeVersion interface
-    # @param self "Me, myself and Irene"
-    def test_05_DCubeVersion( self ):
-        self.log.info( self.version.version() )
-        self.log.info( self.version.root() )
-        self.log.info( self.version.python() )
-        self.log.info( self.version )
-
-
-## test case execution
-if __name__ == "__main__":
-
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubeUtils)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-    
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.py
deleted file mode 100644
index 7494c5e6cce0c2b9f8996cfbd5fa05f821de93a6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.py
+++ /dev/null
@@ -1,300 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-##
-# @file DCubeValid.py
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @date Apr 21st, 2008
-# @brief implementation of DCubeValid class
-
-
-import xml.dom.minidom
-import re
-import sys
-import os
-import unittest
-
-from DCubeUtils import DCubeLogger, DCubeObject, DCubeException, DCubeSingleton
-
-##
-# @class DCubeValid 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief Holder for valid configurations and presettings.
-#
-# <h3>Usage</h3>
-# 
-# DCubeValid object holds legal values for:
-# <ul> 
-# <li> nightly names
-# <li> installation names
-# <li> CMTCONFIG values
-# <li> Atlas project names
-# <li> test frameworks names
-# </ul>
-#
-# The above values are taken from the CoolDozer XML configuration file picked up during DozerValid object construction.
-# If user does not pass the path to the CoolDozer configuration XML file, the default one will be chosen.
-# 
-class DCubeValid( object ):
-
-    __default_dcubeValid_xmlfile = "DCubeValid.xml"
-
-    __metaclass__ = DCubeSingleton
-
-    __server = ""
-
-    __testFrameworks = { }
-
-    __nightlies = { }
-
-    __installs = { }
-
-    __cmtconfigs = { }
-
-    __projects = { }
-    
-
-    ## c'tor
-    # @param self "Me, myself and Irene"
-    # @param xmlfile path to file with XMl configuration
-    # @param log DCubeLogger instance 
-    def __init__( self, xmlfile=None, log=None ):
-
-
-        inInstallArea = None
-        if ( not xmlfile ):
-            inInstallArea = os.path.abspath( os.path.join( os.path.dirname( globals()["__file__"] ), "../share/DCubeValid.xml" ) )
-            if ( not xmlfile and os.path.exists( inInstallArea) ): xmlfile = inInstallArea
-
-        inDCubePath = None
-        ## in current dir
-        if ( not xmlfile ):
-            path = os.path.abspath( globals()["__file__"] ).split("/")[0:-1]
-            path.append( self.__default_dcubeValid_xmlfile )
-            path = "/".join(path)
-            inDCubePath = path
-            if ( os.path.exists( inDCubePath ) ): xmlfile = path
-
-        if ( xmlfile ):        
-            self.__xmlfile = xmlfile
-            self.__parse( xmlfile )
-        else:
-            print "DCubeValid.xml not found in InstallArea nor DCubeClient install directories, wrong installation?"
-            sys.exit( -1 )
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @brief valid nightlies name getter
-    def branches( self ):
-        return self.__nightlies
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @brief valid installs names getter
-    def installs( self ):
-        return self.__installs
-    ##
-    # @param self "Me, myself and Irene"
-    # @brief valid CMTCONFIGs values getter
-    def cmtconfigs( self ):
-        return self.__cmtconfigs
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @brief valid project names getter
-    def projects( self ):
-        return self.__projects
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @param nodeList children of XML node element
-    # @brief CDATA section getter 
-    def __getCData( self, nodeList ):
-        cdata = ""
-        for node in nodeList:
-            if node.nodeType == node.TEXT_NODE:
-                cdata += node.data
-        return cdata
-
-    ##
-    # @param self "Me, myself and Irene"
-    # @param where path to XML file 
-    def __parse( self, where ):
-
-             xmldoc = xml.dom.minidom.parse( where )
-             rootNode = xmldoc.documentElement
-                
-             # get server location
-             server = xmldoc.getElementsByTagName("server")
-             if ( len(server) == 1 ):
-                 server = self.__getCData( server[0].childNodes )
-                 if ( server != "" ):
-                     self.__server = server
-                 else:
-                     self.panic( msg )
-                     raise DCubeException( msg )
-
-
-             # get known nightlies 
-             nightlies = xmldoc.getElementsByTagName("branch")
-            
-             if ( len(nightlies) != 0 ):
-                 for nightly in nightlies:
-                     ndata = self.__getCData( nightly.childNodes )
-                     if ( ndata != "" ):
-                         nre = ndata
-                         if ( nightly.hasAttribute("re") ):
-                             nre = nightly.getAttribute("re")
-                         if ( nre == ndata ): nre = "^" + nre + "$"
-                         self.__nightlies[ndata] = re.compile(nre)                
-           
- 
-             # get known install list
-             install = xmldoc.getElementsByTagName("install")
-             if ( len(install) != 0):
-                 for ins in install:
-                     idata = self.__getCData( ins.childNodes )
-                     if ( idata != "" ):
-                         self.__installs[idata] = idata
-           
-           
-             # get known CMTCONFIGs list
-             cmts = xmldoc.getElementsByTagName("cmt")
-             if ( len(cmts) != 0):
-                 for  cmt in cmts:
-                     cdata = self.__getCData( cmt.childNodes )
-                     if ( cdata != "" ):
-                         self.__cmtconfigs[cdata] = cdata
-           
-             # projects 
-             projects = xmldoc.getElementsByTagName("project")
-             if ( len(projects) != 0):
-                 for bundle in projects:
-                     cdata = self.__getCData( bundle.childNodes )
-                     if ( cdata != "" ):
-           
-                         self.__projects[cdata] = cdata
-
-    ## list valid branch names
-    # @param self "Me, myself and Irene"
-    def listBranches( self ):
-        out = "- valid branches values and thier regexp patterns:\n"
-        i = 1
-        for k  in sorted(  self.branches().keys() ):
-            v = self.branches()[k]
-            out += "  [%02d] %-14s regexp pattern=%s\n" % ( i, k, v.pattern )
-            i += 1
-        return out
-                   
-    ## list valid install names
-    # @param self "Me, myself and Irene"
-    def listInstalls( self ):
-        out = "- valid installs values:\n"
-        i = 1
-        for k in sorted( self.installs() ):
-            out += "  [%02d] %s\n" % ( i, k )
-            i += 1
-        return out
- 
-    ## list valid cmtconfig names
-    # @param self "Me, myself and Irene"
-    def listCmtconfigs( self ):
-        out = "- valid cmtconfigs names:\n"
-        i = 1
-        for k in sorted( self.cmtconfigs() ):
-            out += "  [%02d] %s\n" % ( i, k )
-            i += 1
-        return out
-    
-    ## list valid project names
-    # @param self "Me, myself and Irene"
-    def listProjects( self ):
-        out = "- valid projects names:\n"
-        i = 1
-        for k in sorted( self.projects() ):
-            out += "  [%02d] %s\n" % ( i, k )
-            i += 1    
-        return out
-
-    ## str operator
-    # @param self "Me, myself and Irene"
-    def __str__( self ):
-        out  = "DCube valid names based on DCube XML configuarionm file %s\n" % self.__xmlfile
-        out += self.listBranches()
-        out += self.listInstalls()
-        out += self.listCmtconfigs()
-        out += self.listProjects()
-        return out.strip()
-
-##
-# @class test_DCubeValid
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @date May 22, 2008
-# @brief test case for DCubeValid class
-class test_DCubeValid( unittest.TestCase ):
-
-    ## test case setup
-    # @param self "Me, myself and Irene"
-    def setUp( self ):
-        self.__log = DCubeLogger("test_DCubeValid.log", "test_DCubeValid",toConsole=True)
-        self.cv = DCubeValid()
-    
-    ##__init__
-    # @param self "Me, myself and Irene"
-    def test_01_constructor( self ):
-        try:
-            self.cv = DCubeValid()
-        except:
-            pass
-        self.assertEqual( isinstance( self.cv, DCubeValid ), True)
-
-    ## nightlies()
-    # @param self "Me, myself and Irene"
-    def test_02_nightlies( self ):
-        nightlies = self.cv.branches()
-        self.assertEqual( "dev" in nightlies, True )
-        self.assertEqual( "devval" in nightlies, True )
-        self.assertEqual( "bugfix" in nightlies, True )
-        self.assertEqual( "*" in nightlies, True )
-        self.assertEqual( "val" in nightlies, True )
-        self.assertEqual( "numbered" in nightlies, True )
-        self.assertEqual( "pcache" in nightlies, True )
-        self.assertEqual( "NN.N.NX" in nightlies, True )
-        print nightlies
-
-    ## installs()
-    # @param self "Me, myself and Irene"
-    def test_03_installs( self ):
-        installs = self.cv.installs()
-        self.assertEqual( "build" in installs, True )
-        self.assertEqual( "kit" in installs, True )
-        self.assertEqual( "*" in installs, True )
-
-    ## cmtconfigs()
-    # @param self "Me, mylsef and Irene"
-    def test_04_cmtconfigs( self ):
-        cmtconfigs = self.cv.cmtconfigs()
-        self.assertEqual( len(cmtconfigs.keys()), 7)
-        
-    ## projects()
-    # @param self "Me, myslf and Irene"
-    def test_05_projects( self ):
-        projects = self.cv.projects()
-        self.assertEqual( len( projects.keys()), 12 )
-
-
-    ## str()
-    # @param self "Me, myself and Irene"
-    def test_06_str( self ):
-        print self.cv
-   
-        
-## test suite execution
-if __name__ == "__main__":
-    
-    testLoader = unittest.TestLoader()
-    suite = testLoader.loadTestsFromTestCase(test_DCubeValid)      
-    unittest.TextTestRunner(verbosity=3).run(suite)
-    
-
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.xml b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.xml
deleted file mode 100644
index 1f558cb4bb3ec0773b1dafd3fdcb893360e570ce..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/DCubeValid.xml
+++ /dev/null
@@ -1,67 +0,0 @@
-<?xml version="1.0" ?>
-<!-- 
- DCube API configuration file
- author: Krzysztof Daniel Ciba
- date: November 14th, 2007
- purpose: holds configuration for DCubeClient 
--->
-<dcube version="0.0">
-
-<!-- path to CoolDozer server directory -->
-<server>/afs/cern.ch/user/c/cibak/scratch2</server>
-
-<!-- known branches -->
-<branch>dev</branch>
-<branch>devval</branch>
-<branch>bugfix</branch>
-<branch>point1</branch>
-<branch re="\d{2}\.\d{1}\.\d{1}X$">NN.N.NX</branch>
-<branch re="\d{2}\.\d{1}\.\d{1}X\.Y$">NN.N.NX.Y</branch>
-<branch re="\d{2}\.\d{1}\.\d{1}X\.Y\-VAL$">NN.N.NX.Y-VAL</branch>
-<branch re="\d{2}\.\d{1}\.OLD\.Y$">NN.N.OLD.Y</branch>
-<branch>val</branch>
-<branch>mig0</branch>
-<branch>mig1</branch>
-<branch>mig2</branch>
-<branch>mig3</branch>
-<branch>mig4</branch>
-<branch>mig5</branch>
-<branch>mig6</branch>
-<branch re="(^(?:\d{1,2}\.){3}\d{1,2}$)|(^(?:\d{1,2}\.){2}\d{1,2}$)">numbered</branch>
-<branch re="(^(?:\d{1,2}\.){3}Y$)|(^pcache$)">pcache</branch>
-<!-- this one matches to ALL -->
-<branch re='.+'>*</branch>
-
-<!-- AFS or KIT installation?-->
-<install>build</install>
-<install>kit</install>
-<!-- this one matches to ALL -->
-<install>*</install>
-
-<!-- known CMTCONFIG values -->
-<cmt>i686-slc3-gcc323-dbg</cmt>
-<cmt>i686-slc3-gcc323-opt</cmt>
-<cmt>i686-slc4-gcc34-dbg</cmt>
-<cmt>i686-slc4-gcc34-opt</cmt>
-<cmt>x86_64-slc4-gcc34-dbg</cmt>
-<cmt>x86_64-slc4-gcc34-opt</cmt>
-<!-- this one matches to ALL -->
-<cmt>*</cmt>
-
-
-<!-- known projects -->
-<project>AtlasProduction</project>
-<project>AtlasOffline</project>
-<project>AtlasAnalysis</project> 
-<project>AtlasTrigger</project>
-<project>AtlasReconstruction</project>      
-<project>AtlasSimulation</project>
-<project>AtlasEvent</project>
-<project>AtlasConditions</project>
-<project>AtlasCore</project>
-<project>AtlasTier0</project>
-<project>AtlasP1HLT</project>
-<!-- this one matches to ALL -->
-<project>*</project>
-                    
-</dcube>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/__init__.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/__init__.py
deleted file mode 100644
index 74583d364ec2ca794156596c7254d9b234a940c6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/__init__.py
+++ /dev/null
@@ -1,2 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcube.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcube.py
deleted file mode 100755
index 35bb513140f1ce0317236405eecbe8ecedf1e05f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcube.py
+++ /dev/null
@@ -1,12 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys, os
-from DCubeApp import DCubeApp
-
-if __name__ == "__main__":
-    
-    ## to run ROOT in batch mode
-    sys.argv.append("-b")
-    DCubeApp( )
diff --git a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcubeConvert.py b/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcubeConvert.py
deleted file mode 100755
index f2b3a1ab12295d06408071f832a1261959e5ed63..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DCubeClient-00-00-21/python/dcubeConvert.py
+++ /dev/null
@@ -1,302 +0,0 @@
-#!/bin/env python
-
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-##
-# @file dcubeConvert.py 
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# 
-import os, sys
-from time import strftime, localtime, strptime
-import xml.dom.minidom
-import ROOT
-
-##
-# @class DCubeConvert
-# @author Krzysztof Daniel Ciba (Krzysztof.Ciba@NOSPAMagh.edu.pl)
-# @brief converter for XML DCubeClient configuration files
-class DCubeConvert( object ):
-
-    oldURI = None
-    newURI = None
-    referenceURI = None
-
-    ## c'tor
-    # @param self "Me, myself nad Irene"
-    # @param old old configuration file
-    # @param new new configuration file
-    def __init__( self, old, ref=None, new=None):
-
-        print "dcubeConvert arguments:"
-        whatOLD = old
-        whatREF = ref
-        whatNEW = new
-        if ( not whatREF ):
-            whatREF = "not specified, will be taken from old configuration file"
-        if ( not whatNEW ):
-            whatNEW = "not specified, will be set to old configuration file"
-
-        print "[01] old dcube configuration file: %s" % str(whatOLD)
-        print "[02]          reference root file: %s" % str(whatREF)
-        print "[03] new dcube configuration file: %s" % str(whatNEW)
-
-        if ( not new ):
-            new = old
-            
-        print "will convert old DCubeClient configuration file '%s' and save new configuration to file '%s'" % ( old, new )
-        
-        self.oldURI = os.path.abspath(old)
-        self.newURI = os.path.abspath(new)
-
-        if ( not os.path.isfile( self.oldURI) ):
-            sys.exit("cannot find old DCubeClient config in path %s - exiting!" % self.oldURI )
-
-        if ( self.oldURI == self.newURI ):
-            print "DCubeClient configuration file %s exists and will be overwritten" % self.oldURI
-        
-        if ( ref ):
-            self.referenceURI = os.path.abspath( ref )
-            print "will use '%s' as reference root file" % self.referenceURI
-            if ( not os.path.exists( self.referenceURI ) ):
-                sys.exit("reference file '%s' does not exist!" % self.referenceURI )
-        sys.exit( self.convert() )
-
-    ## get CDATA from node list
-    # @param self "Me, myself and Irene"
-    # @param nodeList list of XML DOM Nodes
-    def __cdata( self, nodeList ):
-        out = ""
-        for node in nodeList:
-            if node.nodeType == node.TEXT_NODE:
-                out += node.data
-        return out
-        
-        pass
-
-    ## parse old configuration file
-    # @param self "Me, myself and Irene"
-    def __parseOLD( self ):
-
-        print "parsing old configuration file %s" % self.oldURI
-        
-        self.xmlOLD =  xml.dom.minidom.parse( self.oldURI )
-
-        if ( self.xmlOLD.documentElement.tagName != "cfg" ):
-            sys.exit("%s isn't an old DCubeClient configuration file, root node name is different form <cfg>" % os.path.basename(self.oldURI ))
-        else:
-            
-            if ( not self.referenceURI ):
-                self.referenceURI = self.__cdata( self.xmlOLD.getElementsByTagName("reference")[0].childNodes )
-
-            print "OLD: reference URI=%s" % self.referenceURI
-
-            if ( not os.path.exists(self.referenceURI ) ):
-                sys.exit( "reference root file not exists!")
-                
-            
-            self.refDesc = self.__cdata( self.xmlOLD.getElementsByTagName("ref_desc")[0].childNodes )
-
-            print "OLD: reference description=%s" % self.refDesc
-
-            self.monDesc = self.__cdata( self.xmlOLD.getElementsByTagName("mon_desc")[0].childNodes )
-
-            print "OLD: monitored description=%s" % self.monDesc
-            
-            self.plimit = self.xmlOLD.getElementsByTagName("plimit")[0]
-
-            print "OLD: plimit attributes fail=%s warn=%s" % ( self.plimit.getAttribute("fail"), self.plimit.getAttribute("warn"))
-
-            self.hists = self.xmlOLD.getElementsByTagName("hist")
-
-            i = 1
-            for hist in self.hists:
-                print "OLD: [%02d] hist path=%s tests=%s" % ( i, hist.getAttribute("path"), hist.getAttribute("tests") )
-                i += 1
-                
-    ## create new XML DOM Document instance 
-    # @param self "Me, myself and Irene"
-    def __createNEW( self ):
-        self.xmlNEW = xml.dom.minidom.Document()
-        self.dcubeClientConfig = self.xmlNEW.createElement("dcube_client_config")
-        self.xmlNEW.appendChild( self.dcubeClientConfig )
-        self.dcubeClientConfig.appendChild( self.__mainComment() )
-
-        self.dcube = self.xmlNEW.createElement("dcube")
-        self.dcube.setAttribute("branch", "*")
-        self.dcube.setAttribute("install", "*")
-        self.dcube.setAttribute("cmtconfig", "*")
-        self.dcube.setAttribute("project", "*")
-        self.dcube.setAttribute("jobId", "*")
-
-        reference = self.xmlNEW.createElement( "reference" )
-        reference.setAttribute("path", self.referenceURI )
-
-        self.dcube.appendChild( reference )
-        
-        refDesc = self.xmlNEW.createElement("ref_desc")
-        refDescCDATA = self.xmlNEW.createTextNode( self.refDesc )
-        refDesc.appendChild( refDescCDATA )
-
-        self.dcube.appendChild( refDesc )
-
-        testDesc = self.xmlNEW.createElement( "test_desc" )
-        testDescCDATA = self.xmlNEW.createTextNode( self.monDesc )
-        testDesc.appendChild( testDescCDATA )
-
-        self.dcube.appendChild( testDesc )
-
-        self.dcube.appendChild( self.plimit )
-
-        rootTDir = self.xmlNEW.createElement("TDirectory")
-        rootTDir.setAttribute("name", "/")
-
-        self.dcube.appendChild( rootTDir )
-
-        
-        self.refROOT = ROOT.TFile().Open( self.referenceURI, "READ" )
-
-        for hist in self.hists:
-            obj = self.refROOT.Get( hist.getAttribute("path"))
-            type = obj.ClassName()
-            path = hist.getAttribute("path").split("/")[1:]
-            tests = hist.getAttribute("tests")
-            self.__attach( path, rootTDir, tests, type )
-
-        self.refROOT.Close()
-        self.dcubeClientConfig.appendChild( self.dcube )
-
-        
-    ## create XML DOM Comment node
-    # @param self "Me, myself and Irene"
-    def __mainComment( self ):
-        com  = "\n  DCubeClient configuration file '%s'" % os.path.basename(self.newURI)
-        com += "\n  converted from old configuration file '%s'" % os.path.basename(self.oldURI)
-        com += "\n  using dcubeConvert.py script"
-        com += "\n  on %s " % strftime("%a, %d %b %Y %H:%M:%S %Z" , localtime())
-        return self.xmlNEW.createComment( com )
-        
-    ## create backup file
-    # @param self "Me, myself and Irene"
-    # @param path path to file
-    def __backup( self, path ):
-
-        backup = path + ".bak"
-        if ( os.path.exists( backup ) ):
-            print "BCK: backup file exists and will be overwitten!"
-        try:
-            os.rename( path, backup )
-        except:
-            print "BCK: creation of backup file failed"
-
-
-    ## save new DCubeClient configuration to file
-    # @param self "Me, myself and Irene"
-    def __save( self ):
-
-        if ( os.path.exists(self.newURI ) ): self.__backup( self.newURI )
-        try:
-            xmlNEW = open( self.newURI, "w+" )
-            xmlNEW.writelines( self.xmlNEW.toprettyxml() )
-            xmlNEW.close()
-            print "NEW: new config file '%s' has been created" % self.newURI
-        except:
-            sys.exit("error when saving new configuration file!")
-        
-
-    ## perform conversion
-    # @param self "Me, myself and Irene"
-    def convert( self ):
-        self.__parseOLD( )
-        self.__createNEW()
-        self.__save()
-        return 0
-
-    ## attach child node to parent
-    # @param self "Me, myself and Irene"
-    # @param what list of path chunks
-    # @param where path inside reference root file
-    def __attach( self, what, where, tests, type ):
-
-        histName = what[-1]
-        parent = what[0]
-        
-        if ( len(what) == 1 ):
-            nodeName = ""
-            if ( "TH1" in type ): nodeName = "hist1D"
-            if ( "TH2" in type ): nodeName = "hist2D"
-            if ( "TProfile" == type ): nodeName == "hist1D"
-            if ( "TProfile2D" == type ): nodeName = "hist2D"
-            if ( "TGraph" in type and "2" not in type ): nodeName = "graph"
-            
-            if ( nodeName != ""  ):
-                histNode = self.xmlNEW.createElement( nodeName )
-                histNode.setAttribute("tests", tests)
-                histNode.setAttribute("name", histName)
-                histNode.setAttribute( "type", type)
-                
-                print "NEW: attaching %s '%s' with tests '%s' at TDirectory '%s'" % ( type, histName, tests, where.getAttribute("name"))
-                where.appendChild( histNode )
-                return
-            else:
-                print "unsupported object name='%s' type='%s' found, skipping!" % ( histName, nodeName)
-        else:
-            subDirs = where.getElementsByTagName("TDirectory")
-            subDirsDict = { }
-
-            for subDir in subDirs:
-                subDirsDict[subDir.getAttribute( "name")] =  subDir 
-    
-            if ( parent not in subDirsDict ):
-                print "NEW: creating TDirectory name '%s'" % parent
-                tdir = self.xmlNEW.createElement( "TDirectory" )
-                tdir.setAttribute("name", parent )
-                where.appendChild( tdir )
-                print "NEW: calling recursive: what '%s' " % str(what[1:])
-                self.__attach( what[1:], tdir, tests, type )
-            else:
-                print "NEW: calling recursive: what '%s' " % str(what[1:])
-                self.__attach( what[1:], subDirsDict[parent], tests, type )
-        
-
-        pass
-
-## script execution
-if __name__ == "__main__":
-
-    usage = "Usage: $sh> dcubeConvert.py OLDCONFIG.XML [REFERENCE.root] [NEWCONFIG.XML]"
-
-    old = None
-    ref = None
-    new = None
-
-    if ( len(sys.argv) == 1 ):
-        print usage
-        sys.exit(0)
-    if ( len(sys.argv) == 2 ):
-        old = sys.argv[1]
-        DCubeConvert( old, ref, new )
-    if ( len(sys.argv) == 3 ):
-        old = sys.argv[1]
-        what = sys.argv[2]
-        if ( what.endswith(".root") ): ref = what
-        elif ( what.endswith(".xml") ): new = what
-        DCubeConvert( old, ref, new )
-    elif ( len(sys.argv) == 4 ):
-        old = sys.argv[1]
-        what1 = sys.argv[2]
-        what2  = sys.argv[3]
-        if ( what1.endswith(".root") ): ref = what1
-        if ( what1.endswith(".xml") ): new = what1
-        if ( not ref and what2.endswith(".root") ): ref = what2
-        if ( not new and what2.endswith(".xml") ): new = what2
-        DCubeConvert(old, ref, new)
-    else:
-        print "ERROR: wrong arguments!"
-        print usage
-        sys.exit(1)
-    
-        
-    
-
-    
-
-            
diff --git a/Tools/RunTimeTester/testsuite/DTD/RTT.dtd b/Tools/RunTimeTester/testsuite/DTD/RTT.dtd
deleted file mode 100755
index 14322bafd665f05b7e9e71f46f89a72c52771f5a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DTD/RTT.dtd
+++ /dev/null
@@ -1,95 +0,0 @@
-<!ELEMENT rttconfig (mode,                release,             refRelease,  
-                     cmtConfig,           branch,              runType,
-                     topProject,          otherProject?,       distArea,
-                     workBasePath,        resultsBasePath,     resultsBaseURL,
-                     site,                dataSetCatalog,      jobGroupConfig,    
-                     pathToVetoFile?,     dCubeCfgFile?,    
-                     confFileSrc, 
-                     packageAccept*,      packageVeto*,        webpageBranchVeto*,
-                     testMode?,           maxTime?,            versionDir?,  
-                     libToolsLoc?,        confFileSuffix?,    
-                     rttKitTest?,         releaseType?,        rttPilotJobs?,
-                     atnPilotJobs?,       concurrent?,         disableMailTo?,    devRun?, 
-                     deleteNonKeepFiles?, archive*,            chainStoreBasePath)>
-    
-   <!ELEMENT mode                   (#PCDATA)>
-   <!ELEMENT release                (#PCDATA)>
-   <!ELEMENT refRelease             (#PCDATA)>
-   <!ELEMENT cmtConfig              (#PCDATA)>
-   <!ELEMENT branch                 (#PCDATA)>
-   <!ELEMENT runType                (releaseRun | kitRun)>
-   <!ELEMENT topProject             (#PCDATA)>  
-   <!ELEMENT otherProject           (#PCDATA)>  
-   <!ELEMENT distArea               (#PCDATA)>
-   <!ELEMENT workBasePath           (#PCDATA)>
-   <!ELEMENT resultsBasePath        (#PCDATA)>
-   <!ELEMENT resultsBaseURL         (#PCDATA)>
-   <!ELEMENT site                   (#PCDATA)>
-   <!ELEMENT dataSetCatalog         (#PCDATA)>
-   <!ELEMENT jobGroupConfig         (#PCDATA)>
-   <!ELEMENT pathToVetoFile         (#PCDATA)>
-   <!ELEMENT dCubeCfgFile           (#PCDATA)>
-   <!ELEMENT chainStoreBasePath     (#PCDATA)>
-   <!ELEMENT confFileSrc            (  (rttProvides,  
-                                         localConfFiles?, 
-                                         installedConfFiles?) |
-                                       (rttProvides?, 
-                                         localConfFiles,  
-                                         installedConfFiles?) |
-                                       (rttProvides?, 
-                                         localConfFiles?, 
-                                         installedConfFiles) )>
-
-   <!ELEMENT rttRun                 EMPTY>
-   <!ELEMENT tctRun                 EMPTY>
-   <!ELEMENT fctRun                 EMPTY>
-
-   <!ELEMENT releaseRun             EMPTY>
-   <!ELEMENT kitRun                 (installArea)>
-   <!ELEMENT installArea            (#PCDATA)>
-   <!ELEMENT rttProvides            (packageConfFile+)>
-      <!ELEMENT packageConfFile     (package, pathToFile, cmtpath, version,
-                                     installArea, shareArea)>
-         <!ELEMENT package          (#PCDATA)>
-         <!ELEMENT pathToFile       (#PCDATA)>
-         <!ELEMENT cmtpath          (#PCDATA)>
-         <!ELEMENT version          (#PCDATA)>
-         <!ELEMENT shareArea        (#PCDATA)>
-   <!ELEMENT localConfFiles         (userHomeCmtDir, testProjectCmtDir)>
-   <!ELEMENT installedConfFiles     EMPTY>
-   <!ELEMENT packageAccept          (#PCDATA)>
-   <!ELEMENT packageVeto            (#PCDATA)>
-   <!ELEMENT webpageBranchVeto      (#PCDATA)>
-   <!ELEMENT testMode               (#PCDATA)>
-   <!ELEMENT maxTime                ((hours, minutes?) | minutes)>
-      <!ELEMENT hours               (#PCDATA)>
-      <!ELEMENT minutes             (#PCDATA)>
-   <!ELEMENT versionDir             (#PCDATA)>
-   <!ELEMENT libToolsLoc            (#PCDATA)>
-   <!ELEMENT confFileSuffix         (#PCDATA)>
-   <!ELEMENT rttKitTest             EMPTY>   
-   <!ELEMENT releaseType            (#PCDATA)> 
-   <!ELEMENT rttPilotJobs               EMPTY>
-   <!ELEMENT atnPilotJobs               EMPTY>
-   <!ELEMENT userHomeCmtDir         (#PCDATA)> 
-   <!ELEMENT testProjectCmtDir      (#PCDATA)> 
-   <!ELEMENT concurrent             (wpcServer)>
-   <!ELEMENT wpcServer            (#PCDATA)>
-   <!ELEMENT disableMailTo              EMPTY>
-   <!ELEMENT devRun                     EMPTY>
-   <!ELEMENT deleteNonKeepFiles         EMPTY>
-
-   <!ELEMENT archive                (type,base,nvers,condition+)>
-     <!ELEMENT type                    (#PCDATA)>
-     <!ELEMENT base                    (#PCDATA)>
-     <!ELEMENT nvers                   (#PCDATA)>
-     <!ELEMENT condition               (size?,name?,parentdir?)>
-        <!ELEMENT size                  (gt|lt|(gt,lt))>
-           <!ELEMENT gt                   (#PCDATA)>
-           <!ELEMENT lt                   (#PCDATA)> 
-        <!ELEMENT name                  (is+|isnot+|(is+,isnot+))>
-           <!ELEMENT is                   (#PCDATA)>
-           <!ELEMENT isnot                (#PCDATA)>
-        <!ELEMENT parentdir             (is+|isnot+|(is+,isnot+))>
-
-  
diff --git a/Tools/RunTimeTester/testsuite/DTD/unifiedTestConfiguration.dtd b/Tools/RunTimeTester/testsuite/DTD/unifiedTestConfiguration.dtd
deleted file mode 100755
index 7cec94394c911e6ec5f7c1e1f40e229ec33db733..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DTD/unifiedTestConfiguration.dtd
+++ /dev/null
@@ -1,189 +0,0 @@
-<!ELEMENT unifiedTestConfiguration ((doc?, atn,  kv?, rtt?) |
-                                    (doc?, atn?, kv,  rtt?) |
-                                    (doc?, atn?, kv?, rtt))>
-
-<!ELEMENT classification ((none) | (displayClass, displayProcess+, displayComponent+))?>
-<!ELEMENT displayClass (#PCDATA)>
-<!ELEMENT displayProcess (#PCDATA)>
-<!ELEMENT displayComponent (#PCDATA)>
-<!ELEMENT none EMPTY>
-
-
-<!ELEMENT atn         (classification?, TEST*)>
-  <!ELEMENT TEST     (doc?, classification?, options_atn*, furtherOptions_atn*, timelimit*, author*, mailto*, nightly_exclude*, prescript*, postscript*, package_atn*, rundir*, expectations*)>
-  <!ATTLIST TEST   name       CDATA                    #IMPLIED>
-  <!ATTLIST TEST   type       (script|athena|kv|rtt|makecheck)   #REQUIRED>
-  <!ATTLIST TEST   suite      CDATA                    "general">
-  <!ELEMENT options_atn     (#PCDATA | rttSpecs | kvSpecs)*>
-      <!ELEMENT rttSpecs  ANY>
-      <!ELEMENT kvSpecs   ANY>
-  <!ELEMENT furtherOptions_atn (#PCDATA)>
-  <!ELEMENT timelimit          (#PCDATA)>
-  <!ELEMENT author             (#PCDATA)>
-  <!ELEMENT mailto             (#PCDATA)>
-  <!ELEMENT prescript          (#PCDATA)>
-  <!ELEMENT nightly_exclude    (#PCDATA)>
-  <!ELEMENT postscript         (#PCDATA)>
-  <!ELEMENT package_atn        (#PCDATA)>
-  <!ELEMENT rundir             (#PCDATA)>
-  <!ELEMENT expectations       (errorMessage*, warningMessage*, successMessage*, returnValue*)>
-     <!ELEMENT errorMessage    (#PCDATA)>
-     <!ELEMENT warningMessage  (#PCDATA)>
-     <!ELEMENT successMessage  (#PCDATA)>
-     <!ELEMENT returnValue     (#PCDATA)>
-
-<!ELEMENT kv                (kvtest*)>
-    <!ELEMENT kvtest (release,priority?,kvsuite,trf,desc,author?,version?,
-                    inpath,infile,inpool,outpath,outfile,outntuple?,logfile,signature,copyfiles,checkfiles)>
-    <!ATTLIST kvtest                name    CDATA #REQUIRED>
-    <!ATTLIST kvtest                enabled CDATA #REQUIRED>
-
-    <!ELEMENT release     (#PCDATA)> 
-    <!ELEMENT priority    (#PCDATA)> 
-    <!ELEMENT kvsuite       (#PCDATA)> 
-    <!ELEMENT trf         (#PCDATA)> 
-    <!ELEMENT desc        (#PCDATA)> 
-    <!ELEMENT version     (#PCDATA)> 
-    <!ELEMENT inpath      (#PCDATA)> 
-    <!ELEMENT infile      (#PCDATA)> 
-    <!ELEMENT inpool      (#PCDATA)> 
-    <!ELEMENT outpath     (#PCDATA)> 
-    <!ELEMENT outfile     (#PCDATA)> 
-    <!ELEMENT outntuple   (#PCDATA)> 
-    <!ELEMENT logfile     (#PCDATA)> 
-    <!ELEMENT signature   (#PCDATA)> 
-    <!ELEMENT copyfiles   (#PCDATA)> 
-    <!ELEMENT checkfiles  (#PCDATA)> 
-
-
-<!ELEMENT rtt              ((EMPTY) | (rttContactPerson*,mailto?,refRelease, jobList, jobGroups?) | (jobGroups))?>
-   <!ELEMENT refRelease    (#PCDATA)>   
-   <!ELEMENT jobList       (classification?, athena*, rulechecker*, 
-                            qametrics*,  kitvalidation*, rttmonitor*,
-                            rttselftest*, jobTransform*, pseudo*, chain*)>
-      <!ELEMENT athena     ( doc?, doclink?, (numberedOnly |  limitedNightlies)?, rttKitTestJob?, 
-                            classification?, displayName?, commandLineFlags?, commandLineString?, preconfigOptions?, options, furtherOptions?,
-                            package?, group, confederation?, 
-                            queue?,  
-		            ((dataset*, dataset_info*) | chaindataset_info*), 
-                            dq2*, asciiRefFile*, binRefFile*,
-                            storeInDatabasePattern*, action*, monAction*, 
-                            test*, keepFilePattern*, auxFilePattern*, athenaFlavour*)>
-
-	 <!ELEMENT doc                    (#PCDATA)>
-	 <!ELEMENT doclink                (#PCDATA)>
-	 <!ELEMENT displayName            (#PCDATA)>
-	 <!ELEMENT rttContactPerson       (#PCDATA)>
-	 <!ELEMENT numberedOnly              EMPTY>
-	 <!ELEMENT limitedNightlies          EMPTY>
-	 <!ELEMENT limitedNightly         (nightly+)>
-         <!ELEMENT rttKitTestJob          EMPTY>
-         <!ELEMENT commandLineFlags       (#PCDATA)>
-         <!ELEMENT commandLineString      (#PCDATA)>
-         <!ELEMENT preconfigOptions       (#PCDATA)>
-         <!ELEMENT options                (#PCDATA)>
-          <!ELEMENT furtherOptions        (jobOptionItem+)>
-             <!ELEMENT jobOptionItem      (#PCDATA)>
-	 <!ELEMENT package                (#PCDATA)>
-	 <!ELEMENT group                  (#PCDATA)>
-	 <!ELEMENT confederation          (#PCDATA)>
-	 <!ELEMENT queue                  (#PCDATA)>
-	 <!ELEMENT nightly                (#PCDATA)>
-
- 	 <!-- The old dataset tag -->
-         <!ELEMENT dataset                  (#PCDATA)>	
-
- 	 <!-- The new dataset tag along with the definition of its children elements -->
-
-	 <!ELEMENT dataset_info                ((bytestream | dc2 | dc3 | jobTransformData), datasetName,datasetRepeat?)>
-	 <!ELEMENT bytestream                 EMPTY>
-	 <!ELEMENT dc2                        EMPTY>
-	 <!ELEMENT dc3                        EMPTY>
-	 <!ELEMENT jobTransformData           EMPTY>
-	 <!ELEMENT datasetName            (#PCDATA)>
-	 <!ELEMENT datasetRepeat          (#PCDATA)>
-
-	 <!ELEMENT chaindataset_info      ((bytestream | dc2 | dc3 | jobTransformData), chaindatasetName,datasetRepeat?, dataset_info)>
-	 <!ELEMENT chaindatasetName       (#PCDATA)>
-
-         <!ELEMENT dq2                    (dq2site, dq2dataset,datatype, dq2file+)>
-         <!ELEMENT dq2site                (#PCDATA)>
-         <!ELEMENT dq2dataset             (#PCDATA)>
-         <!ELEMENT datatype               (#PCDATA)>
-         <!ELEMENT dq2file                (#PCDATA)>
-	 <!ELEMENT asciiRefFile           (#PCDATA)>
-	 <!ELEMENT storeInDatabasePattern (#PCDATA)>
-	 <!ELEMENT athenaFlavour          (#PCDATA)>
-
-      <!ELEMENT rulechecker ( (numberedOnly | limitedNightlies)?, 
-                             package, queue)>
-      <!ELEMENT qametrics   ( (numberedOnly | limitedNightlies)?,
-                             package, queue)>
-      <!ELEMENT kitvalidation ((EMPTY) | ( (numberedOnly | limitedNightlies)?, queue))?>
-      <!ELEMENT rttmonitor EMPTY>
-      <!ELEMENT rttselftest EMPTY>
-      <!ELEMENT jobTransform ((numberedOnly |  limitedNightlies)?, 
-                              doc?, doclink?, classification?, rttKitTestJob?, displayName?, jobTransformJobName, 
-                              jobTransformCmd, group, (dataset_info* | chaindataset_info*), 
-		              queue?, keepFilePattern*, auxFilePattern*, (action*|test*)*)>
-
-         <!ELEMENT jobTransformJobName (#PCDATA)>
-         <!ELEMENT jobTransformCmd     (#PCDATA)>
-	 <!ELEMENT chain               (chainName, (chainElement | sequential | parallel))>
-         <!ELEMENT chainName           (#PCDATA)>
-         <!ELEMENT sequential          (sequential | parallel | chainElement)* >
-         <!ELEMENT parallel            (sequential | parallel | chainElement)* >
-         <!ELEMENT chainElement        (( athena | jobTransform | pseudo), chainfileout*)>
-	 <!ELEMENT chainfileout        (#PCDATA) >
-         <!ELEMENT pseudo              (athena)>
-
-   <!ELEMENT jobGroups (jobGroup+)>
-      <!ELEMENT jobGroup (keepFilePattern*,   asciiRefFile*, 
-                          binRefFile*,        storeInDatabasePattern*, 
-                          checker*,           
-                          postScript*,        auxFilePattern*, 
-                          monAction*,
-                          (action*|test*)*,   
-                          stringToFile*,     
-                          summary*,           logChecker*)>
-
-            <!ELEMENT keepFilePattern (#PCDATA)>
-	    <!ELEMENT binRefFile      (#PCDATA)>
-            <!ELEMENT checker         ((searchString*, vetoString*) | (comparisonType?))>
-	       <!ELEMENT searchString     (#PCDATA)>
-	       <!ELEMENT vetoString       (#PCDATA)>
-               <!ELEMENT comparisonType   (#PCDATA)>
-            <!ELEMENT postScript          EMPTY>
-	    <!ELEMENT auxFilePattern      (#PCDATA)>
-            <!ELEMENT action             (modulename,testname, arg*, listarg*)>
-            <!ELEMENT monAction          (modulename,testname, arg*, listarg*)>
-            <!ELEMENT test               (modulename,testname, arg*, listarg*)>
-               <!ELEMENT modulename       (#PCDATA)>
-	       <!ELEMENT testname         (#PCDATA)>
-               <!ELEMENT arg              (argname, argvalue)>
-	          <!ELEMENT argname       (#PCDATA)>
-		  <!ELEMENT argvalue      (#PCDATA)>
-               <!ELEMENT listarg          (listargname, listargvalue+)>
-	          <!ELEMENT listargname   (#PCDATA)>
-		  <!ELEMENT listargvalue  (#PCDATA)>
-            <!ELEMENT stringToFile        (line+)>
-	       <!ELEMENT line             (#PCDATA)>
-            <!ELEMENT logChecker          ((searchString+, vetoString*) | 
-	                                   (searchString*, vetoString+))>
-            <!ELEMENT summary             (#PCDATA)>
-
-<!ATTLIST rtt               xmlns  CDATA #REQUIRED>
-<!ATTLIST jobGroup          name   CDATA #REQUIRED>
-<!ATTLIST jobGroup          parent CDATA #REQUIRED>
-<!ATTLIST athena            trendId  CDATA "">
-<!ATTLIST jobTransform      trendId  CDATA "">
-<!ATTLIST action            position  CDATA "">
-<!ATTLIST action            runInAtlasEnv  CDATA "">
-<!ATTLIST test              position  CDATA "">
-<!ATTLIST keepFilePattern   info   CDATA "">
-<!ATTLIST keepFilePattern   displayColor   CDATA "black">
-<!ATTLIST checker           class  CDATA #REQUIRED>
-<!ATTLIST logChecker        class  CDATA #REQUIRED>
-<!ATTLIST postScript        class  CDATA #REQUIRED>
-<!ATTLIST stringToFile      file   CDATA #REQUIRED>
-
diff --git a/Tools/RunTimeTester/testsuite/DTD/vetoFile.dtd b/Tools/RunTimeTester/testsuite/DTD/vetoFile.dtd
deleted file mode 100644
index e629ebef14c3c83d28698ac06088ec6b8f3c69a8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/DTD/vetoFile.dtd
+++ /dev/null
@@ -1,9 +0,0 @@
-<!ELEMENT vetos (CPC?, ids?)>
-
-<!ELEMENT CPC              (displayClass*, displayProcess*, displayComponent*)>
-<!ELEMENT ids              (id*)>
-<!ELEMENT id               (#PCDATA)>
-<!ELEMENT displayClass     (#PCDATA)>
-<!ELEMENT displayProcess   (#PCDATA)>
-<!ELEMENT displayComponent (#PCDATA)>
-
diff --git a/Tools/RunTimeTester/testsuite/data/simplemergejob/simple1.data b/Tools/RunTimeTester/testsuite/data/simplemergejob/simple1.data
deleted file mode 100644
index e38a7c972470281c7c7983dcf36edc9b398f259e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/data/simplemergejob/simple1.data
+++ /dev/null
@@ -1 +0,0 @@
-31 32 33 34 35 36 37 38 39 40
diff --git a/Tools/RunTimeTester/testsuite/data/simplemergejob/simple2.data b/Tools/RunTimeTester/testsuite/data/simplemergejob/simple2.data
deleted file mode 100644
index ddb508295e08455490da5543d6a2ad6d6c602adf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/data/simplemergejob/simple2.data
+++ /dev/null
@@ -1 +0,0 @@
-41 42 43 44 45 46 47 48 49 50
diff --git a/Tools/RunTimeTester/testsuite/install.py b/Tools/RunTimeTester/testsuite/install.py
deleted file mode 100644
index d1db3d01e7fa2ebd49d34e5f845503f5319eb982..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/install.py
+++ /dev/null
@@ -1,99 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-import os
-import os.path
-import getopt
-
-def usage():
-    print "Please supply path to directory in which to install."
-    sys.exit(1)
-    
-def getargs():
-    try:
-        return sys.argv[1].strip()
-    except:
-        usage()
-        
-    return basepath
-
-def yesnoQuestion(question):
-    ans = raw_input(question)
-    while ans.lower() not in ['y','n']:
-        ans = raw_input(question)
-    return ans.lower()    
-        
-def checkExists(basepath):
-    if not os.path.exists(basepath):
-        if yesnoQuestion('%s: inexistant dir. Create it? [y|n]: ' % basepath)=='n': sys.exit(1)
-        os.system('mkdir -p %s' % basepath)
-    elif os.listdir(basepath):
-        if yesnoQuestion('%s: exists, but dir not empty, really install there? [y|n]: ' % basepath) == 'n': sys.exit(1)
-
-def checkEnvVars():
-    try:
-        os.environ['RTTShareDir']
-        os.environ['RTTSrcDir']
-    except KeyError:
-        print 'Please first set up the RTTShareDir, RTTSrcDir and RTTLogDir env vars, if not done.'
-        sys.exit(1)
-
-def setupInstall(basepath):
-    print 'Copying stuff to %s' % basepath
-    pathToThisModule = os.path.dirname(os.path.abspath(__file__))
-    os.system('cd %s;cp -r %s/* .' % (basepath, pathToThisModule))
-
-def tplToFile(tplpath, srcToDstDict):
-    f = open(tplpath)
-    conts = f.read()
-    f.close()
-
-    for src,dst in srcToDstDict.items():
-        conts = conts.replace(src,dst)
-
-    f = open(tplpath.split('.TPL')[0], 'w')
-    f.write(conts)
-    f.close()
-
-    os.remove(tplpath)
-
-def installFramework(basepath):
-    print 'Creating files from templates...'
-
-    dict = {'$_BASE_$':basepath,
-            '$_LIB_TOOLS_FROM_ENV_$':os.environ['RTTShareDir']}
-
-    tplToFile(os.path.join(basepath, 'xml/TopLevelJobGroup.xml.TPL'), dict)
-    tplToFile(os.path.join(basepath, 'xml/rttDataSetCatalog.xml.TPL'), dict)
-
-    for testname in os.listdir(os.path.join(basepath, 'testsuites')):
-        dict['$_TESTNAME_$'] = testname
-        start = os.path.join(basepath, 'testsuites', testname)
-        for root, dirs, files in os.walk(start):
-            tplFiles = [os.path.join(root, f) for f in files if f.endswith('.TPL')]
-            [tplToFile(t, dict) for t in tplFiles]
-
-def cleanup(basepath):
-    print 'Cleaning up...'
-    # Now remove the install.py file, no longer relevant
-    os.remove(os.path.join(basepath, 'install.py'))
-    print 'Done.'
-
-def howtorun(basepath):
-    print '************************************************************************'
-    print 'To run the framework:'
-    print '\t1) cd %s' % basepath
-    print '\t2) ./run.sh --jobs'
-    print '\t3) ./run.sh --tests -d <path_to_testsuites_dir>'
-    print '************************************************************************'
-
-if __name__ == '__main__':
-    checkEnvVars()
-    basepath = getargs()
-    checkExists(basepath)
-    setupInstall(basepath)
-    installFramework(basepath)
-    cleanup(basepath)
-    howtorun(basepath)
-
-
diff --git a/Tools/RunTimeTester/testsuite/run.sh b/Tools/RunTimeTester/testsuite/run.sh
deleted file mode 100755
index 86d36a32cd5e2c46a93bc6764022599240de50ac..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run.sh
+++ /dev/null
@@ -1,27 +0,0 @@
-#!/bin/bash
-
-whatToDo=$1
-
-if test "$whatToDo" = "-h" || test "$whatToDo" = "--help";then
-    echo "./run.sh --jobs ==> to run jobs"
-    echo "./run.sh --tests ==> to run tests (assumes you have a testsuites/ dir in the current dir)"
-    echo "./run.sh --tests -d <path_to_testsuites_dir> ==> to run tests with different testsuites dir"
-else
-    if test "$whatToDo" != "--tests" && test "$whatToDo" != "--jobs"; then
-	echo "Unknown argument", $whatToDo, "must be one of --jobs or --tests.";
-    fi
-fi
-
-export PYTHONPATH=`pwd`/src:$RTTSrcDir:$PYTHONPATH
-
-if test "$whatToDo" = "--jobs";then
-    python run/runSuites.py
-elif test "$whatToDo" = "--tests";then
-    if [ $# -eq 1 ]; then
-	python run/runTests.py -d testsuites	
-    else
-	shift;
-	python run/runTests.py $@	
-    fi
-fi
-
diff --git a/Tools/RunTimeTester/testsuite/run/coverage.py b/Tools/RunTimeTester/testsuite/run/coverage.py
deleted file mode 100755
index 49336db30f340562eaa74c031e361a92b5990b5b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/coverage.py
+++ /dev/null
@@ -1,1163 +0,0 @@
-#!/usr/bin/python
-#
-#             Perforce Defect Tracking Integration Project
-#              <http://www.ravenbrook.com/project/p4dti/>
-#
-#                   COVERAGE.PY -- COVERAGE TESTING
-#
-#             Gareth Rees, Ravenbrook Limited, 2001-12-04
-#                     Ned Batchelder, 2004-12-12
-#         http://nedbatchelder.com/code/modules/coverage.html
-#
-#
-# 1. INTRODUCTION
-#
-# This module provides coverage testing for Python code.
-#
-# The intended readership is all Python developers.
-#
-# This document is not confidential.
-#
-# See [GDR 2001-12-04a] for the command-line interface, programmatic
-# interface and limitations.  See [GDR 2001-12-04b] for requirements and
-# design.
-
-r"""Usage:
-
-coverage.py -x [-p] MODULE.py [ARG1 ARG2 ...]
-    Execute module, passing the given command-line arguments, collecting
-    coverage data. With the -p option, write to a temporary file containing
-    the machine name and process ID.
-
-coverage.py -e
-    Erase collected coverage data.
-
-coverage.py -c
-    Collect data from multiple coverage files (as created by -p option above)
-    and store it into a single file representing the union of the coverage.
-
-coverage.py -r [-m] [-o dir1,dir2,...] FILE1 FILE2 ...
-    Report on the statement coverage for the given files.  With the -m
-    option, show line numbers of the statements that weren't executed.
-
-coverage.py -a [-d dir] [-o dir1,dir2,...] FILE1 FILE2 ...
-    Make annotated copies of the given files, marking statements that
-    are executed with > and statements that are missed with !.  With
-    the -d option, make the copies in that directory.  Without the -d
-    option, make each copy in the same directory as the original.
-
--o dir,dir2,...
-  Omit reporting or annotating files when their filename path starts with
-  a directory listed in the omit list.
-  e.g. python coverage.py -i -r -o c:\python23,lib\enthought\traits
-
-Coverage data is saved in the file .coverage by default.  Set the
-COVERAGE_FILE environment variable to save it somewhere else."""
-
-__version__ = "2.85.20080914"    # see detailed history at the end of this file.
-
-import compiler
-import compiler.visitor
-import glob
-import os
-import re
-import string
-import symbol
-import sys
-import threading
-import token
-import types
-import zipimport
-from socket import gethostname
-
-# Python version compatibility
-try:
-    strclass = basestring   # new to 2.3
-except:
-    strclass = str
-
-# 2. IMPLEMENTATION
-#
-# This uses the "singleton" pattern.
-#
-# The word "morf" means a module object (from which the source file can
-# be deduced by suitable manipulation of the __file__ attribute) or a
-# filename.
-#
-# When we generate a coverage report we have to canonicalize every
-# filename in the coverage dictionary just in case it refers to the
-# module we are reporting on.  It seems a shame to throw away this
-# information so the data in the coverage dictionary is transferred to
-# the 'cexecuted' dictionary under the canonical filenames.
-#
-# The coverage dictionary is called "c" and the trace function "t".  The
-# reason for these short names is that Python looks up variables by name
-# at runtime and so execution time depends on the length of variables!
-# In the bottleneck of this application it's appropriate to abbreviate
-# names to increase speed.
-
-class StatementFindingAstVisitor(compiler.visitor.ASTVisitor):
-    """ A visitor for a parsed Abstract Syntax Tree which finds executable
-        statements.
-    """
-    def __init__(self, statements, excluded, suite_spots):
-        compiler.visitor.ASTVisitor.__init__(self)
-        self.statements = statements
-        self.excluded = excluded
-        self.suite_spots = suite_spots
-        self.excluding_suite = 0
-        
-    def doRecursive(self, node):
-        for n in node.getChildNodes():
-            self.dispatch(n)
-
-    visitStmt = visitModule = doRecursive
-    
-    def doCode(self, node):
-        if hasattr(node, 'decorators') and node.decorators:
-            self.dispatch(node.decorators)
-            self.recordAndDispatch(node.code)
-        else:
-            self.doSuite(node, node.code)
-            
-    visitFunction = visitClass = doCode
-
-    def getFirstLine(self, node):
-        # Find the first line in the tree node.
-        lineno = node.lineno
-        for n in node.getChildNodes():
-            f = self.getFirstLine(n)
-            if lineno and f:
-                lineno = min(lineno, f)
-            else:
-                lineno = lineno or f
-        return lineno
-
-    def getLastLine(self, node):
-        # Find the first line in the tree node.
-        lineno = node.lineno
-        for n in node.getChildNodes():
-            lineno = max(lineno, self.getLastLine(n))
-        return lineno
-    
-    def doStatement(self, node):
-        self.recordLine(self.getFirstLine(node))
-
-    visitAssert = visitAssign = visitAssTuple = visitPrint = \
-        visitPrintnl = visitRaise = visitSubscript = visitDecorators = \
-        doStatement
-    
-    def visitPass(self, node):
-        # Pass statements have weird interactions with docstrings.  If this
-        # pass statement is part of one of those pairs, claim that the statement
-        # is on the later of the two lines.
-        l = node.lineno
-        if l:
-            lines = self.suite_spots.get(l, [l,l])
-            self.statements[lines[1]] = 1
-        
-    def visitDiscard(self, node):
-        # Discard nodes are statements that execute an expression, but then
-        # discard the results.  This includes function calls, so we can't 
-        # ignore them all.  But if the expression is a constant, the statement
-        # won't be "executed", so don't count it now.
-        if node.expr.__class__.__name__ != 'Const':
-            self.doStatement(node)
-
-    def recordNodeLine(self, node):
-        # Stmt nodes often have None, but shouldn't claim the first line of
-        # their children (because the first child might be an ignorable line
-        # like "global a").
-        if node.__class__.__name__ != 'Stmt':
-            return self.recordLine(self.getFirstLine(node))
-        else:
-            return 0
-    
-    def recordLine(self, lineno):
-        # Returns a bool, whether the line is included or excluded.
-        if lineno:
-            # Multi-line tests introducing suites have to get charged to their
-            # keyword.
-            if lineno in self.suite_spots:
-                lineno = self.suite_spots[lineno][0]
-            # If we're inside an excluded suite, record that this line was
-            # excluded.
-            if self.excluding_suite:
-                self.excluded[lineno] = 1
-                return 0
-            # If this line is excluded, or suite_spots maps this line to
-            # another line that is exlcuded, then we're excluded.
-            elif self.excluded.has_key(lineno) or \
-                 self.suite_spots.has_key(lineno) and \
-                 self.excluded.has_key(self.suite_spots[lineno][1]):
-                return 0
-            # Otherwise, this is an executable line.
-            else:
-                self.statements[lineno] = 1
-                return 1
-        return 0
-    
-    default = recordNodeLine
-    
-    def recordAndDispatch(self, node):
-        self.recordNodeLine(node)
-        self.dispatch(node)
-
-    def doSuite(self, intro, body, exclude=0):
-        exsuite = self.excluding_suite
-        if exclude or (intro and not self.recordNodeLine(intro)):
-            self.excluding_suite = 1
-        self.recordAndDispatch(body)
-        self.excluding_suite = exsuite
-        
-    def doPlainWordSuite(self, prevsuite, suite):
-        # Finding the exclude lines for else's is tricky, because they aren't
-        # present in the compiler parse tree.  Look at the previous suite,
-        # and find its last line.  If any line between there and the else's
-        # first line are excluded, then we exclude the else.
-        lastprev = self.getLastLine(prevsuite)
-        firstelse = self.getFirstLine(suite)
-        for l in range(lastprev+1, firstelse):
-            if self.suite_spots.has_key(l):
-                self.doSuite(None, suite, exclude=self.excluded.has_key(l))
-                break
-        else:
-            self.doSuite(None, suite)
-        
-    def doElse(self, prevsuite, node):
-        if node.else_:
-            self.doPlainWordSuite(prevsuite, node.else_)
-    
-    def visitFor(self, node):
-        self.doSuite(node, node.body)
-        self.doElse(node.body, node)
-
-    visitWhile = visitFor
-
-    def visitIf(self, node):
-        # The first test has to be handled separately from the rest.
-        # The first test is credited to the line with the "if", but the others
-        # are credited to the line with the test for the elif.
-        self.doSuite(node, node.tests[0][1])
-        for t, n in node.tests[1:]:
-            self.doSuite(t, n)
-        self.doElse(node.tests[-1][1], node)
-
-    def visitTryExcept(self, node):
-        self.doSuite(node, node.body)
-        for i in range(len(node.handlers)):
-            a, b, h = node.handlers[i]
-            if not a:
-                # It's a plain "except:".  Find the previous suite.
-                if i > 0:
-                    prev = node.handlers[i-1][2]
-                else:
-                    prev = node.body
-                self.doPlainWordSuite(prev, h)
-            else:
-                self.doSuite(a, h)
-        self.doElse(node.handlers[-1][2], node)
-    
-    def visitTryFinally(self, node):
-        self.doSuite(node, node.body)
-        self.doPlainWordSuite(node.body, node.final)
-        
-    def visitWith(self, node):
-        self.doSuite(node, node.body)
-        
-    def visitGlobal(self, node):
-        # "global" statements don't execute like others (they don't call the
-        # trace function), so don't record their line numbers.
-        pass
-
-the_coverage = None
-
-class CoverageException(Exception):
-    pass
-
-class coverage:
-    # Name of the cache file (unless environment variable is set).
-    cache_default = ".coverage"
-
-    # Environment variable naming the cache file.
-    cache_env = "COVERAGE_FILE"
-
-    # A dictionary with an entry for (Python source file name, line number
-    # in that file) if that line has been executed.
-    c = {}
-    
-    # A map from canonical Python source file name to a dictionary in
-    # which there's an entry for each line number that has been
-    # executed.
-    cexecuted = {}
-
-    # Cache of results of calling the analysis2() method, so that you can
-    # specify both -r and -a without doing double work.
-    analysis_cache = {}
-
-    # Cache of results of calling the canonical_filename() method, to
-    # avoid duplicating work.
-    canonical_filename_cache = {}
-
-    def __init__(self):
-        global the_coverage
-        if the_coverage:
-            raise CoverageException("Only one coverage object allowed.")
-        self.usecache = 1
-        self.cache = None
-        self.parallel_mode = False
-        self.exclude_re = ''
-        self.nesting = 0
-        self.cstack = []
-        self.xstack = []
-        self.relative_dir = self.abs_file(os.curdir)+os.sep
-        self.exclude('# *pragma[: ]*[nN][oO] *[cC][oO][vV][eE][rR]')
-
-    # t(f, x, y).  This method is passed to sys.settrace as a trace function.  
-    # See [van Rossum 2001-07-20b, 9.2] for an explanation of sys.settrace and 
-    # the arguments and return value of the trace function.
-    # See [van Rossum 2001-07-20a, 3.2] for a description of frame and code
-    # objects.
-    
-    def t(self, f, w, unused):                                 #pragma: no cover
-        if w == 'line':
-            self.c[(f.f_code.co_filename, f.f_lineno)] = 1
-            #-for c in self.cstack:
-            #-    c[(f.f_code.co_filename, f.f_lineno)] = 1
-        return self.t
-    
-    def help(self, error=None):     #pragma: no cover
-        if error:
-            print error
-            print
-        print __doc__
-        sys.exit(1)
-
-    def command_line(self, argv, help_fn=None):
-        import getopt
-        help_fn = help_fn or self.help
-        settings = {}
-        optmap = {
-            '-a': 'annotate',
-            '-c': 'collect',
-            '-d:': 'directory=',
-            '-e': 'erase',
-            '-h': 'help',
-            '-i': 'ignore-errors',
-            '-m': 'show-missing',
-            '-p': 'parallel-mode',
-            '-r': 'report',
-            '-x': 'execute',
-            '-o:': 'omit=',
-            }
-        short_opts = string.join(map(lambda o: o[1:], optmap.keys()), '')
-        long_opts = optmap.values()
-        options, args = getopt.getopt(argv, short_opts, long_opts)
-        for o, a in options:
-            if optmap.has_key(o):
-                settings[optmap[o]] = 1
-            elif optmap.has_key(o + ':'):
-                settings[optmap[o + ':']] = a
-            elif o[2:] in long_opts:
-                settings[o[2:]] = 1
-            elif o[2:] + '=' in long_opts:
-                settings[o[2:]+'='] = a
-            else:       #pragma: no cover
-                pass    # Can't get here, because getopt won't return anything unknown.
-
-        if settings.get('help'):
-            help_fn()
-
-        for i in ['erase', 'execute']:
-            for j in ['annotate', 'report', 'collect']:
-                if settings.get(i) and settings.get(j):
-                    help_fn("You can't specify the '%s' and '%s' "
-                              "options at the same time." % (i, j))
-
-        args_needed = (settings.get('execute')
-                       or settings.get('annotate')
-                       or settings.get('report'))
-        action = (settings.get('erase') 
-                  or settings.get('collect')
-                  or args_needed)
-        if not action:
-            help_fn("You must specify at least one of -e, -x, -c, -r, or -a.")
-        if not args_needed and args:
-            help_fn("Unexpected arguments: %s" % " ".join(args))
-        
-        self.parallel_mode = settings.get('parallel-mode')
-        self.get_ready()
-
-        if settings.get('erase'):
-            self.erase()
-        if settings.get('execute'):
-            if not args:
-                help_fn("Nothing to do.")
-            sys.argv = args
-            self.start()
-            import __main__
-            sys.path[0] = os.path.dirname(sys.argv[0])
-            execfile(sys.argv[0], __main__.__dict__)
-        if settings.get('collect'):
-            self.collect()
-        if not args:
-            args = self.cexecuted.keys()
-        
-        ignore_errors = settings.get('ignore-errors')
-        show_missing = settings.get('show-missing')
-        directory = settings.get('directory=')
-
-        omit = settings.get('omit=')
-        if omit is not None:
-            omit = [self.abs_file(p) for p in omit.split(',')]
-        else:
-            omit = []
-        
-        if settings.get('report'):
-            self.report(args, show_missing, ignore_errors, omit_prefixes=omit)
-        if settings.get('annotate'):
-            self.annotate(args, directory, ignore_errors, omit_prefixes=omit)
-
-    def use_cache(self, usecache, cache_file=None):
-        self.usecache = usecache
-        if cache_file and not self.cache:
-            self.cache_default = cache_file
-        
-    def get_ready(self, parallel_mode=False):
-        if self.usecache and not self.cache:
-            self.cache = os.environ.get(self.cache_env, self.cache_default)
-            if self.parallel_mode:
-                self.cache += "." + gethostname() + "." + str(os.getpid())
-            self.restore()
-        self.analysis_cache = {}
-        
-    def start(self, parallel_mode=False):
-        self.get_ready()
-        if self.nesting == 0:                               #pragma: no cover
-            sys.settrace(self.t)
-            if hasattr(threading, 'settrace'):
-                threading.settrace(self.t)
-        self.nesting += 1
-        
-    def stop(self):
-        self.nesting -= 1
-        if self.nesting == 0:                               #pragma: no cover
-            sys.settrace(None)
-            if hasattr(threading, 'settrace'):
-                threading.settrace(None)
-
-    def erase(self):
-        self.get_ready()
-        self.c = {}
-        self.analysis_cache = {}
-        self.cexecuted = {}
-        if self.cache and os.path.exists(self.cache):
-            os.remove(self.cache)
-
-    def exclude(self, re):
-        if self.exclude_re:
-            self.exclude_re += "|"
-        self.exclude_re += "(" + re + ")"
-
-    def begin_recursive(self):
-        self.cstack.append(self.c)
-        self.xstack.append(self.exclude_re)
-        
-    def end_recursive(self):
-        self.c = self.cstack.pop()
-        self.exclude_re = self.xstack.pop()
-
-    # save().  Save coverage data to the coverage cache.
-
-    def save(self):
-        if self.usecache and self.cache:
-            self.canonicalize_filenames()
-            cache = open(self.cache, 'wb')
-            import marshal
-            marshal.dump(self.cexecuted, cache)
-            cache.close()
-
-    # restore().  Restore coverage data from the coverage cache (if it exists).
-
-    def restore(self):
-        self.c = {}
-        self.cexecuted = {}
-        assert self.usecache
-        if os.path.exists(self.cache):
-            self.cexecuted = self.restore_file(self.cache)
-
-    def restore_file(self, file_name):
-        try:
-            cache = open(file_name, 'rb')
-            import marshal
-            cexecuted = marshal.load(cache)
-            cache.close()
-            if isinstance(cexecuted, types.DictType):
-                return cexecuted
-            else:
-                return {}
-        except:
-            return {}
-
-    # collect(). Collect data in multiple files produced by parallel mode
-
-    def collect(self):
-        cache_dir, local = os.path.split(self.cache)
-        for f in os.listdir(cache_dir or '.'):
-            if not f.startswith(local):
-                continue
-
-            full_path = os.path.join(cache_dir, f)
-            cexecuted = self.restore_file(full_path)
-            self.merge_data(cexecuted)
-
-    def merge_data(self, new_data):
-        for file_name, file_data in new_data.items():
-            if self.cexecuted.has_key(file_name):
-                self.merge_file_data(self.cexecuted[file_name], file_data)
-            else:
-                self.cexecuted[file_name] = file_data
-
-    def merge_file_data(self, cache_data, new_data):
-        for line_number in new_data.keys():
-            if not cache_data.has_key(line_number):
-                cache_data[line_number] = new_data[line_number]
-
-    def abs_file(self, filename):
-        """ Helper function to turn a filename into an absolute normalized
-            filename.
-        """
-        return os.path.normcase(os.path.abspath(os.path.realpath(filename)))
-
-    def get_zip_data(self, filename):
-        """ Get data from `filename` if it is a zip file path, or return None
-            if it is not.
-        """
-        markers = ['.zip'+os.sep, '.egg'+os.sep]
-        for marker in markers:
-            if marker in filename:
-                parts = filename.split(marker)
-                try:
-                    zi = zipimport.zipimporter(parts[0]+marker[:-1])
-                except zipimport.ZipImportError:
-                    continue
-                try:
-                    data = zi.get_data(parts[1])
-                except IOError:
-                    continue
-                return data
-        return None
-
-    # canonical_filename(filename).  Return a canonical filename for the
-    # file (that is, an absolute path with no redundant components and
-    # normalized case).  See [GDR 2001-12-04b, 3.3].
-
-    def canonical_filename(self, filename):
-        if not self.canonical_filename_cache.has_key(filename):
-            f = filename
-            if os.path.isabs(f) and not os.path.exists(f):
-                if not self.get_zip_data(f):
-                    f = os.path.basename(f)
-            if not os.path.isabs(f):
-                for path in [os.curdir] + sys.path:
-                    g = os.path.join(path, f)
-                    if os.path.exists(g):
-                        f = g
-                        break
-            cf = self.abs_file(f)
-            self.canonical_filename_cache[filename] = cf
-        return self.canonical_filename_cache[filename]
-
-    # canonicalize_filenames().  Copy results from "c" to "cexecuted", 
-    # canonicalizing filenames on the way.  Clear the "c" map.
-
-    def canonicalize_filenames(self):
-        for filename, lineno in self.c.keys():
-            if filename == '<string>':
-                # Can't do anything useful with exec'd strings, so skip them.
-                continue
-            f = self.canonical_filename(filename)
-            if not self.cexecuted.has_key(f):
-                self.cexecuted[f] = {}
-            self.cexecuted[f][lineno] = 1
-        self.c = {}
-
-    # morf_filename(morf).  Return the filename for a module or file.
-
-    def morf_filename(self, morf):
-        if hasattr(morf, '__file__'):
-            f = morf.__file__
-        else:
-            f = morf
-        return self.canonical_filename(f)
-
-    # analyze_morf(morf).  Analyze the module or filename passed as
-    # the argument.  If the source code can't be found, raise an error.
-    # Otherwise, return a tuple of (1) the canonical filename of the
-    # source code for the module, (2) a list of lines of statements
-    # in the source code, (3) a list of lines of excluded statements,
-    # and (4), a map of line numbers to multi-line line number ranges, for
-    # statements that cross lines.
-    
-    def analyze_morf(self, morf):
-        if self.analysis_cache.has_key(morf):
-            return self.analysis_cache[morf]
-        filename = self.morf_filename(morf)
-        ext = os.path.splitext(filename)[1]
-        source, sourcef = None, None
-        if ext == '.pyc':
-            if not os.path.exists(filename[:-1]):
-                source = self.get_zip_data(filename[:-1])
-                if not source:
-                    raise CoverageException(
-                        "No source for compiled code '%s'." % filename
-                        )
-            filename = filename[:-1]
-        if not source:
-            sourcef = open(filename, 'rU')
-            source = sourcef.read()
-        try:
-            lines, excluded_lines, line_map = self.find_executable_statements(
-                source, exclude=self.exclude_re
-                )
-        except SyntaxError, synerr:
-            raise CoverageException(
-                "Couldn't parse '%s' as Python source: '%s' at line %d" %
-                    (filename, synerr.msg, synerr.lineno)
-                )
-        if sourcef:
-            sourcef.close()
-        result = filename, lines, excluded_lines, line_map
-        self.analysis_cache[morf] = result
-        return result
-
-    def first_line_of_tree(self, tree):
-        while True:
-            if len(tree) == 3 and type(tree[2]) == type(1):
-                return tree[2]
-            tree = tree[1]
-    
-    def last_line_of_tree(self, tree):
-        while True:
-            if len(tree) == 3 and type(tree[2]) == type(1):
-                return tree[2]
-            tree = tree[-1]
-    
-    def find_docstring_pass_pair(self, tree, spots):
-        for i in range(1, len(tree)):
-            if self.is_string_constant(tree[i]) and self.is_pass_stmt(tree[i+1]):
-                first_line = self.first_line_of_tree(tree[i])
-                last_line = self.last_line_of_tree(tree[i+1])
-                self.record_multiline(spots, first_line, last_line)
-        
-    def is_string_constant(self, tree):
-        try:
-            return tree[0] == symbol.stmt and tree[1][1][1][0] == symbol.expr_stmt
-        except:
-            return False
-        
-    def is_pass_stmt(self, tree):
-        try:
-            return tree[0] == symbol.stmt and tree[1][1][1][0] == symbol.pass_stmt
-        except:
-            return False
-
-    def record_multiline(self, spots, i, j):
-        for l in range(i, j+1):
-            spots[l] = (i, j)
-            
-    def get_suite_spots(self, tree, spots):
-        """ Analyze a parse tree to find suite introducers which span a number
-            of lines.
-        """
-        for i in range(1, len(tree)):
-            if type(tree[i]) == type(()):
-                if tree[i][0] == symbol.suite:
-                    # Found a suite, look back for the colon and keyword.
-                    lineno_colon = lineno_word = None
-                    for j in range(i-1, 0, -1):
-                        if tree[j][0] == token.COLON:
-                            # Colons are never executed themselves: we want the
-                            # line number of the last token before the colon.
-                            lineno_colon = self.last_line_of_tree(tree[j-1])
-                        elif tree[j][0] == token.NAME:
-                            if tree[j][1] == 'elif':
-                                # Find the line number of the first non-terminal
-                                # after the keyword.
-                                t = tree[j+1]
-                                while t and token.ISNONTERMINAL(t[0]):
-                                    t = t[1]
-                                if t:
-                                    lineno_word = t[2]
-                            else:
-                                lineno_word = tree[j][2]
-                            break
-                        elif tree[j][0] == symbol.except_clause:
-                            # "except" clauses look like:
-                            # ('except_clause', ('NAME', 'except', lineno), ...)
-                            if tree[j][1][0] == token.NAME:
-                                lineno_word = tree[j][1][2]
-                                break
-                    if lineno_colon and lineno_word:
-                        # Found colon and keyword, mark all the lines
-                        # between the two with the two line numbers.
-                        self.record_multiline(spots, lineno_word, lineno_colon)
-
-                    # "pass" statements are tricky: different versions of Python
-                    # treat them differently, especially in the common case of a
-                    # function with a doc string and a single pass statement.
-                    self.find_docstring_pass_pair(tree[i], spots)
-                    
-                elif tree[i][0] == symbol.simple_stmt:
-                    first_line = self.first_line_of_tree(tree[i])
-                    last_line = self.last_line_of_tree(tree[i])
-                    if first_line != last_line:
-                        self.record_multiline(spots, first_line, last_line)
-                self.get_suite_spots(tree[i], spots)
-
-    def find_executable_statements(self, text, exclude=None):
-        # Find lines which match an exclusion pattern.
-        excluded = {}
-        suite_spots = {}
-        if exclude:
-            reExclude = re.compile(exclude)
-            lines = text.split('\n')
-            for i in range(len(lines)):
-                if reExclude.search(lines[i]):
-                    excluded[i+1] = 1
-
-        # Parse the code and analyze the parse tree to find out which statements
-        # are multiline, and where suites begin and end.
-        import parser
-        tree = parser.suite(text+'\n\n').totuple(1)
-        self.get_suite_spots(tree, suite_spots)
-        #print "Suite spots:", suite_spots
-        
-        # Use the compiler module to parse the text and find the executable
-        # statements.  We add newlines to be impervious to final partial lines.
-        statements = {}
-        ast = compiler.parse(text+'\n\n')
-        visitor = StatementFindingAstVisitor(statements, excluded, suite_spots)
-        compiler.walk(ast, visitor, walker=visitor)
-
-        lines = statements.keys()
-        lines.sort()
-        excluded_lines = excluded.keys()
-        excluded_lines.sort()
-        return lines, excluded_lines, suite_spots
-
-    # format_lines(statements, lines).  Format a list of line numbers
-    # for printing by coalescing groups of lines as long as the lines
-    # represent consecutive statements.  This will coalesce even if
-    # there are gaps between statements, so if statements =
-    # [1,2,3,4,5,10,11,12,13,14] and lines = [1,2,5,10,11,13,14] then
-    # format_lines will return "1-2, 5-11, 13-14".
-
-    def format_lines(self, statements, lines):
-        pairs = []
-        i = 0
-        j = 0
-        start = None
-        pairs = []
-        while i < len(statements) and j < len(lines):
-            if statements[i] == lines[j]:
-                if start == None:
-                    start = lines[j]
-                end = lines[j]
-                j = j + 1
-            elif start:
-                pairs.append((start, end))
-                start = None
-            i = i + 1
-        if start:
-            pairs.append((start, end))
-        def stringify(pair):
-            start, end = pair
-            if start == end:
-                return "%d" % start
-            else:
-                return "%d-%d" % (start, end)
-        ret = string.join(map(stringify, pairs), ", ")
-        return ret
-
-    # Backward compatibility with version 1.
-    def analysis(self, morf):
-        f, s, _, m, mf = self.analysis2(morf)
-        return f, s, m, mf
-
-    def analysis2(self, morf):
-        filename, statements, excluded, line_map = self.analyze_morf(morf)
-        self.canonicalize_filenames()
-        if not self.cexecuted.has_key(filename):
-            self.cexecuted[filename] = {}
-        missing = []
-        for line in statements:
-            lines = line_map.get(line, [line, line])
-            for l in range(lines[0], lines[1]+1):
-                if self.cexecuted[filename].has_key(l):
-                    break
-            else:
-                missing.append(line)
-        return (filename, statements, excluded, missing,
-                self.format_lines(statements, missing))
-
-    def relative_filename(self, filename):
-        """ Convert filename to relative filename from self.relative_dir.
-        """
-        return filename.replace(self.relative_dir, "")
-
-    def morf_name(self, morf):
-        """ Return the name of morf as used in report.
-        """
-        if hasattr(morf, '__name__'):
-            return morf.__name__
-        else:
-            return self.relative_filename(os.path.splitext(morf)[0])
-
-    def filter_by_prefix(self, morfs, omit_prefixes):
-        """ Return list of morfs where the morf name does not begin
-            with any one of the omit_prefixes.
-        """
-        filtered_morfs = []
-        for morf in morfs:
-            for prefix in omit_prefixes:
-                if self.morf_name(morf).startswith(prefix):
-                    break
-            else:
-                filtered_morfs.append(morf)
-
-        return filtered_morfs
-
-    def morf_name_compare(self, x, y):
-        return cmp(self.morf_name(x), self.morf_name(y))
-
-    def report(self, morfs, show_missing=1, ignore_errors=0, file=None, omit_prefixes=[]):
-        if not isinstance(morfs, types.ListType):
-            morfs = [morfs]
-        # On windows, the shell doesn't expand wildcards.  Do it here.
-        globbed = []
-        for morf in morfs:
-            if isinstance(morf, strclass):
-                globbed.extend(glob.glob(morf))
-            else:
-                globbed.append(morf)
-        morfs = globbed
-        
-        morfs = self.filter_by_prefix(morfs, omit_prefixes)
-        morfs.sort(self.morf_name_compare)
-
-        max_name = max([5,] + map(len, map(self.morf_name, morfs)))
-        fmt_name = "%%- %ds  " % max_name
-        fmt_err = fmt_name + "%s: %s"
-        header = fmt_name % "Name" + " Stmts   Exec  Cover"
-        fmt_coverage = fmt_name + "% 6d % 6d % 5d%%"
-        if show_missing:
-            header = header + "   Missing"
-            fmt_coverage = fmt_coverage + "   %s"
-        if not file:
-            file = sys.stdout
-        print >>file, header
-        print >>file, "-" * len(header)
-        total_statements = 0
-        total_executed = 0
-        for morf in morfs:
-            name = self.morf_name(morf)
-            try:
-                _, statements, _, missing, readable  = self.analysis2(morf)
-                n = len(statements)
-                m = n - len(missing)
-                if n > 0:
-                    pc = 100.0 * m / n
-                else:
-                    pc = 100.0
-                args = (name, n, m, pc)
-                if show_missing:
-                    args = args + (readable,)
-                print >>file, fmt_coverage % args
-                total_statements = total_statements + n
-                total_executed = total_executed + m
-            except KeyboardInterrupt:                       #pragma: no cover
-                raise
-            except:
-                if not ignore_errors:
-                    typ, msg = sys.exc_info()[:2]
-                    print >>file, fmt_err % (name, typ, msg)
-        if len(morfs) > 1:
-            print >>file, "-" * len(header)
-            if total_statements > 0:
-                pc = 100.0 * total_executed / total_statements
-            else:
-                pc = 100.0
-            args = ("TOTAL", total_statements, total_executed, pc)
-            if show_missing:
-                args = args + ("",)
-            print >>file, fmt_coverage % args
-
-    # annotate(morfs, ignore_errors).
-
-    blank_re = re.compile(r"\s*(#|$)")
-    else_re = re.compile(r"\s*else\s*:\s*(#|$)")
-
-    def annotate(self, morfs, directory=None, ignore_errors=0, omit_prefixes=[]):
-        morfs = self.filter_by_prefix(morfs, omit_prefixes)
-        for morf in morfs:
-            try:
-                filename, statements, excluded, missing, _ = self.analysis2(morf)
-                self.annotate_file(filename, statements, excluded, missing, directory)
-            except KeyboardInterrupt:
-                raise
-            except:
-                if not ignore_errors:
-                    raise
-                
-    def annotate_file(self, filename, statements, excluded, missing, directory=None):
-        source = open(filename, 'r')
-        if directory:
-            dest_file = os.path.join(directory,
-                                     os.path.basename(filename)
-                                     + ',cover')
-        else:
-            dest_file = filename + ',cover'
-        dest = open(dest_file, 'w')
-        lineno = 0
-        i = 0
-        j = 0
-        covered = 1
-        while 1:
-            line = source.readline()
-            if line == '':
-                break
-            lineno = lineno + 1
-            while i < len(statements) and statements[i] < lineno:
-                i = i + 1
-            while j < len(missing) and missing[j] < lineno:
-                j = j + 1
-            if i < len(statements) and statements[i] == lineno:
-                covered = j >= len(missing) or missing[j] > lineno
-            if self.blank_re.match(line):
-                dest.write('  ')
-            elif self.else_re.match(line):
-                # Special logic for lines containing only 'else:'.  
-                # See [GDR 2001-12-04b, 3.2].
-                if i >= len(statements) and j >= len(missing):
-                    dest.write('! ')
-                elif i >= len(statements) or j >= len(missing):
-                    dest.write('> ')
-                elif statements[i] == missing[j]:
-                    dest.write('! ')
-                else:
-                    dest.write('> ')
-            elif lineno in excluded:
-                dest.write('- ')
-            elif covered:
-                dest.write('> ')
-            else:
-                dest.write('! ')
-            dest.write(line)
-        source.close()
-        dest.close()
-
-# Singleton object.
-the_coverage = coverage()
-
-# Module functions call methods in the singleton object.
-def use_cache(*args, **kw): 
-    return the_coverage.use_cache(*args, **kw)
-
-def start(*args, **kw): 
-    return the_coverage.start(*args, **kw)
-
-def stop(*args, **kw): 
-    return the_coverage.stop(*args, **kw)
-
-def erase(*args, **kw): 
-    return the_coverage.erase(*args, **kw)
-
-def begin_recursive(*args, **kw): 
-    return the_coverage.begin_recursive(*args, **kw)
-
-def end_recursive(*args, **kw): 
-    return the_coverage.end_recursive(*args, **kw)
-
-def exclude(*args, **kw): 
-    return the_coverage.exclude(*args, **kw)
-
-def analysis(*args, **kw): 
-    return the_coverage.analysis(*args, **kw)
-
-def analysis2(*args, **kw): 
-    return the_coverage.analysis2(*args, **kw)
-
-def report(*args, **kw): 
-    return the_coverage.report(*args, **kw)
-
-def annotate(*args, **kw): 
-    return the_coverage.annotate(*args, **kw)
-
-def annotate_file(*args, **kw): 
-    return the_coverage.annotate_file(*args, **kw)
-
-# Save coverage data when Python exits.  (The atexit module wasn't
-# introduced until Python 2.0, so use sys.exitfunc when it's not
-# available.)
-try:
-    import atexit
-    atexit.register(the_coverage.save)
-except ImportError:
-    sys.exitfunc = the_coverage.save
-
-def main():
-    the_coverage.command_line(sys.argv[1:])
-    
-# Command-line interface.
-if __name__ == '__main__':
-    main()
-
-
-# A. REFERENCES
-#
-# [GDR 2001-12-04a] "Statement coverage for Python"; Gareth Rees;
-# Ravenbrook Limited; 2001-12-04;
-# <http://www.nedbatchelder.com/code/modules/rees-coverage.html>.
-#
-# [GDR 2001-12-04b] "Statement coverage for Python: design and
-# analysis"; Gareth Rees; Ravenbrook Limited; 2001-12-04;
-# <http://www.nedbatchelder.com/code/modules/rees-design.html>.
-#
-# [van Rossum 2001-07-20a] "Python Reference Manual (releae 2.1.1)";
-# Guide van Rossum; 2001-07-20;
-# <http://www.python.org/doc/2.1.1/ref/ref.html>.
-#
-# [van Rossum 2001-07-20b] "Python Library Reference"; Guido van Rossum;
-# 2001-07-20; <http://www.python.org/doc/2.1.1/lib/lib.html>.
-#
-#
-# B. DOCUMENT HISTORY
-#
-# 2001-12-04 GDR Created.
-#
-# 2001-12-06 GDR Added command-line interface and source code
-# annotation.
-#
-# 2001-12-09 GDR Moved design and interface to separate documents.
-#
-# 2001-12-10 GDR Open cache file as binary on Windows.  Allow
-# simultaneous -e and -x, or -a and -r.
-#
-# 2001-12-12 GDR Added command-line help.  Cache analysis so that it
-# only needs to be done once when you specify -a and -r.
-#
-# 2001-12-13 GDR Improved speed while recording.  Portable between
-# Python 1.5.2 and 2.1.1.
-#
-# 2002-01-03 GDR Module-level functions work correctly.
-#
-# 2002-01-07 GDR Update sys.path when running a file with the -x option,
-# so that it matches the value the program would get if it were run on
-# its own.
-#
-# 2004-12-12 NMB Significant code changes.
-# - Finding executable statements has been rewritten so that docstrings and
-#   other quirks of Python execution aren't mistakenly identified as missing
-#   lines.
-# - Lines can be excluded from consideration, even entire suites of lines.
-# - The filesystem cache of covered lines can be disabled programmatically.
-# - Modernized the code.
-#
-# 2004-12-14 NMB Minor tweaks.  Return 'analysis' to its original behavior
-# and add 'analysis2'.  Add a global for 'annotate', and factor it, adding
-# 'annotate_file'.
-#
-# 2004-12-31 NMB Allow for keyword arguments in the module global functions.
-# Thanks, Allen.
-#
-# 2005-12-02 NMB Call threading.settrace so that all threads are measured.
-# Thanks Martin Fuzzey. Add a file argument to report so that reports can be 
-# captured to a different destination.
-#
-# 2005-12-03 NMB coverage.py can now measure itself.
-#
-# 2005-12-04 NMB Adapted Greg Rogers' patch for using relative filenames,
-# and sorting and omitting files to report on.
-#
-# 2006-07-23 NMB Applied Joseph Tate's patch for function decorators.
-#
-# 2006-08-21 NMB Applied Sigve Tjora and Mark van der Wal's fixes for argument
-# handling.
-#
-# 2006-08-22 NMB Applied Geoff Bache's parallel mode patch.
-#
-# 2006-08-23 NMB Refactorings to improve testability.  Fixes to command-line
-# logic for parallel mode and collect.
-#
-# 2006-08-25 NMB "#pragma: nocover" is excluded by default.
-#
-# 2006-09-10 NMB Properly ignore docstrings and other constant expressions that
-# appear in the middle of a function, a problem reported by Tim Leslie.
-# Minor changes to avoid lint warnings.
-#
-# 2006-09-17 NMB coverage.erase() shouldn't clobber the exclude regex.
-# Change how parallel mode is invoked, and fix erase() so that it erases the
-# cache when called programmatically.
-#
-# 2007-07-21 NMB In reports, ignore code executed from strings, since we can't
-# do anything useful with it anyway.
-# Better file handling on Linux, thanks Guillaume Chazarain.
-# Better shell support on Windows, thanks Noel O'Boyle.
-# Python 2.2 support maintained, thanks Catherine Proulx.
-#
-# 2007-07-22 NMB Python 2.5 now fully supported. The method of dealing with
-# multi-line statements is now less sensitive to the exact line that Python
-# reports during execution. Pass statements are handled specially so that their
-# disappearance during execution won't throw off the measurement.
-#
-# 2007-07-23 NMB Now Python 2.5 is *really* fully supported: the body of the
-# new with statement is counted as executable.
-#
-# 2007-07-29 NMB Better packaging.
-#
-# 2007-09-30 NMB Don't try to predict whether a file is Python source based on
-# the extension. Extensionless files are often Pythons scripts. Instead, simply
-# parse the file and catch the syntax errors.  Hat tip to Ben Finney.
-#
-# 2008-05-25 NMB Open files in rU mode to avoid line ending craziness.
-# Thanks, Edward Loper.
-#
-# 2008-09-14 NMB Add support for finding source files in eggs.
-# Don't check for morf's being instances of ModuleType, instead use duck typing
-# so that pseudo-modules can participate. Thanks, Imri Goldberg.
-# Use os.realpath as part of the fixing of filenames so that symlinks won't
-# confuse things.  Thanks, Patrick Mezard.
-#
-#
-# C. COPYRIGHT AND LICENCE
-#
-# Copyright 2001 Gareth Rees.  All rights reserved.
-# Copyright 2004-2008 Ned Batchelder.  All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions are
-# met:
-#
-# 1. Redistributions of source code must retain the above copyright
-#    notice, this list of conditions and the following disclaimer.
-#
-# 2. Redistributions in binary form must reproduce the above copyright
-#    notice, this list of conditions and the following disclaimer in the
-#    documentation and/or other materials provided with the
-#    distribution.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# HOLDERS AND CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
-# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
-# BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
-# OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
-# TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
-# USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH
-# DAMAGE.
-#
-# $Id: coverage.py,v 1.1 2009-04-02 15:00:33 rtt Exp $
diff --git a/Tools/RunTimeTester/testsuite/run/newtestsuite.py b/Tools/RunTimeTester/testsuite/run/newtestsuite.py
deleted file mode 100644
index 6eefcaa824e1f44f0b2a5a9b8c1e4ef8abb83172..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/newtestsuite.py
+++ /dev/null
@@ -1,398 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import sys
-import os, os.path
-
-# Helper function to grossly determine if testsuitepath ressembles a testsuite
-def isTestSuite(testsuitepath, templSuite=False):
-    # List of what we expect in the testsuitepath
-    if not os.path.isdir(testsuitepath): 
-        print '%s: not a directory' % testsuitepath
-        return False
-
-    things = ['tests.py', '__init__.py']
-    if templSuite:
-        things.append('cfg/cfg.xml.TPL')
-    else:
-        things.append('cfg/cfg.xml')
-
-    missing = [os.path.join(testsuitepath, t) for t in things if not os.path.exists(os.path.join(testsuitepath, t))]
-    for m in missing:
-        print '%s: missing from testsuite.' % m
-    return len(missing)==0
-
-# Write file helper
-def write(what, where):
-    h = open(where, 'w')
-    h.write(what)
-    h.close()
-
-# Turn a testsuite .TPL file into a normal file
-def tplToFile(tplpath, srcToDstDict):
-    f = open(tplpath)
-    conts = f.read()
-    f.close()
-
-    for src,dst in srcToDstDict.items():
-        conts = conts.replace(src,dst)
-
-    f = open(tplpath.split('.TPL')[0], 'w')
-    f.write(conts)
-    f.close()
-
-    os.remove(tplpath)
-
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-def usage():
-    print 'python %s </abs/or/rel/path/to/testsuite/dir_1>, [</abs/or/rel/path/to/testsuite/dir_2>, ...]' % sys.argv[0]
-    print
-    print 'At least one testsuite is required. Multiple suites should be space separated.'
-    sys.exit()
-
-def get():
-    if len(sys.argv)<2: usage()
-
-    suites = []
-    for ts in sys.argv[1:]:
-        ts.strip()
-        help = '-h' in ts or '--h' in ts or '-help' in ts or '--help' in ts
-        if help: usage()
-        suites.append(ts)
-    return [os.path.abspath(s) for s in suites]
-
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-def create(testsuitepath):
-    if os.path.exists(testsuitepath):
-        print '%s: already exists' % testsuitepath
-        sys.exit(1)
-        
-    testsuitedir  = os.path.dirname(testsuitepath)
-    testsuitename = os.path.basename(testsuitepath)
-
-    # create it
-    cmd  = 'mkdir -p %s;' % testsuitepath
-    cmd += 'cd %s;' % testsuitepath
-    cmd += 'mkdir Results Work cfg Package logs;'
-    os.system(cmd)
-
-    # now dump out various template files
-    write(cfg(), os.path.join(testsuitepath, 'cfg/cfg.xml.TPL'))
-    write(package(), os.path.join(testsuitepath, 'Package/package.xml.TPL'))
-    write(init(), os.path.join(testsuitepath, '__init__.py'))
-    write(tests(), os.path.join(testsuitepath, 'tests.py'))
-
-    print 'New %s testsuite skeleton created OK' % testsuitename
-    print 'You should now complete the relevant bits of:'
-    print os.path.join(testsuitepath, '__init__.py')
-    print os.path.join(testsuitepath, 'cfg/cfg.xml.TPL')
-    print os.path.join(testsuitepath, 'Package/package.xml.TPL')
-    print 
-    print
-
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-def install(testsuitepath):
-    def checkEnvVar():
-        try:
-            os.environ['RTTShareDir']
-        except KeyError:
-            print 'Please set up the RTTShareDir env var, then re-run to install.'
-            sys.exit(1)
-
-    checkEnvVar()
-
-    if not os.path.exists(testsuitepath):
-        print 'Cannot install inexistant testsuite: %s' % testsuitepath
-        sys.exit(1)
-
-    if not isTestSuite(testsuitepath, templSuite=True): 
-        print 'Cannot install, there were errors.'
-        sys.exit(1)
-
-    path = os.path.dirname(testsuitepath)
-    name = os.path.basename(testsuitepath)
-
-    m = 'Please give path to the _existing_ "testsuites" dir in which this should be installed:\n> '
-    testsuitesdir = os.path.abspath(raw_input(m).strip())
-    while not os.path.exists(testsuitesdir):
-        testsuitesdir = os.path.abspath(raw_input(m).strip())
-
-    installpath = os.path.join(testsuitesdir, name)
-    if os.path.exists(installpath):
-        print '%s: exists already, doing nothing.' % installpath
-        sys.exit(1)
-    
-    # now install in the testsuites dir given
-    cmd  = 'cd %s;' % testsuitesdir
-    cmd += 'cp -r %s .;' % testsuitepath
-    cmd += 'cd %s;' % name    
-    os.system(cmd)
-
-    dict = {'$_BASE_$' : os.path.dirname(testsuitesdir) ,
-            '$_LIB_TOOLS_FROM_ENV_$' : os.environ['RTTShareDir'],
-            '$_TESTNAME_$' : name}
-
-    for root, dirs, files in os.walk(installpath):
-        tplFiles = [os.path.join(root, f) for f in files if f.endswith('.TPL')]
-        [tplToFile(t, dict) for t in tplFiles]
-
-    print 'Installed: OK (Original %s untouched)' % testsuitepath
-
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-def upload(testsuitepath):
-    import tarfile
-
-    def cleanTestSuite():
-        for root, dirs, files in os.walk(testsuitepath):
-            for thing in [os.path.join(root, f) for f in files if f.endswith('~')]:
-                print 'Cleaning testsuite: %s: removing...' % thing
-                os.remove(thing)
-
-    # Method proper starts here
-    if not os.path.exists(testsuitepath):
-        print '%s: inexistant testsuite' % testsuitepath
-        sys.exit(1)
-    
-    if not isTestSuite(testsuitepath, templSuite=True): 
-        print 'Not uploading to tarball.'
-        sys.exit(1)
-
-    cleanTestSuite()
-    
-    # Finally ready to upload
-    tarballName = 'rttTesting.tgz'
-    tarballpath = '/afs/cern.ch/atlas/project/RTT/data/brinick/%s' % tarballName
-
-    tmpArea = '/tmp/rtt/tarball'    
-    os.system('mkdir -p %s;cp %s %s' % (tmpArea, tarballpath, tmpArea))
-    if not os.path.exists(os.path.join(tmpArea, tarballName)):
-        print 'Cannot grab tarball, exiting...'
-        sys.exit(1)
-
-    tgzfile = tarfile.open(tarballpath, 'r:gz')
-    for thing in tgzfile:
-        testsuitename = os.path.basename(testsuitepath)
-        if 'testsuites/%s' % testsuitename in thing.name:
-            print 'Testsuite %s already exists in official tarball. Not adding yours, sorry...' % testsuitename
-            tgzfile.close()
-            sys.exit(1)
-
-    cmd  = 'cd %s;' % tmpArea
-    cmd += 'tar xfz %s;' % tarballName
-    cmd += 'cd testsuites;'
-    cmd += 'cp -r %s .;' % testsuitepath
-    cmd += 'echo "Adding testsuite %s into tarball...";' % os.path.basename(testsuitepath)
-    cmd += 'cd ..;'
-    cmd += 'rm -rf %s;' % tarballName
-    cmd += 'echo "Building new tarball...";'
-    cmd += 'tar cfz %s *;' % tarballName
-    cmd += 'echo "Backing up previous official tarball...";'
-    cmd += 'cp %s %s.BACKUP;' % (tarballpath, tarballpath)
-    cmd += 'echo "Adding new tarball into official location...";'
-    cmd += 'cp %s %s;' % (tarballName, os.path.dirname(tarballpath))
-    cmd += 'cd %s;' % tmpArea
-    cmd += 'rm -rf *;'
-    cmd += 'echo "Done: OK."'
-    os.system(cmd)
-    
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-def ask(ts):
-    m  = 'What would you like to do with %s:\n' % ts
-    m += '\t[1] Create it ?\n'
-    m += '\t[2] Install it ? (so you can test it works)\n'
-    m += '\t[3] Add it? (to the official tarball repository)\n'
-    m += 'Choices: 1,2,3? '
-    ans = raw_input(m).strip()
-    while ans not in ['1','2','3']:
-        ans = raw_input(m).strip()
-    return ans
-
-def main():
-    for testsuitepath in get():
-        {'1':create,'2':install,'3':upload}.get(ask(testsuitepath))(testsuitepath)    
-
-# ====================================================================================
-# ====================================================================================
-# ====================================================================================
-
-#
-# Templates for creating new testsuites...
-#
-
-def cfg():
-    return """<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <refDB>
-      <file>$_BASE_$/testsuites/$_TESTNAME_$/refFile.db</file>
-      <enable>1</enable>
-   </refDB>
-   <moniDB>
-      <file>$_BASE_$/testsuites/$_TESTNAME_$/moniFile.db</file>
-      <enable>1</enable>
-   </moniDB>
-
-   <confFileSrc>
-
-   <!--
-      <rttProvides>
-         
-         <packageConfFile>
-            <package></package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath></cmtpath>
-            <version></version>
-            <installArea></installArea>
-            <shareArea></shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-         
-   -->
-
-   </confFileSrc>
-      
-   <packageAccept></packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
-"""
-
-def package():
-    return """<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-      <!--
-          <athena>
-             <doc>blah</doc>
-             <doclink>http://cern.ch</doclink>
-             <options>AthExHelloWorld/HelloWorldOptions.py</options>
-             <package>Control/AthenaExamples/AthExHelloWorld</package>
-             <group>HelloWorld</group>
-             <queue>short</queue>
-          </athena>
-
-       -->
-
-       </jobList>
-          
-       <jobGroups>
-          
-       <!-- <jobGroup name="AthenaHelloWorld" parent="Athena"></jobGroup> -->
-          
-       </jobGroups>
-          
-    </rtt>
-</unifiedTestConfiguration>
-"""
-
-def tests():
-    return """from utility import log
-from checker import Checker
-
-def run():
-    for klass in [ExampleChecker]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-'''Each new Checker child class that is created
-should be (if you want it to be run) added to the list
-in the run() method above.
-Checker child class constraints:
-1) Must inherit from Checker as shown in MyChecker example below
-2) Must take a tester instance in constructor 
-   (that Checker binds to self.tester. You also get self.summaryFile which contains an instance
-   of RTTSummaryFile (imported at top of this file. The source can be found in run/ dir.))
-3) Each test function called by the run() method in your Checker class should be decorated with the log function
-4) Each test function called by the run() method in your Checker class should return a boolean indicating outcome
-'''
-
-# !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
-# NOTE: This class (because of its name ExampleChecker) will not be run by the framework.
-# It is ignored, and is only here so as to provide, erm, an example...
-# Either write a new checker class, or modify this one including changing its name.
-# In either case, you should then add the checker class name to the list in run() above.
-# !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
-class ExampleChecker(Checker):
-    def __init__(self, tester):
-        Checker.__init__(tester)
-
-    def run(self):
-        self.myTestOne()
-        self.myTestTwo()
-
-    @log
-    def myTestOne(self):
-        "Each test function should be decorated by the log function."
-        return True
-
-    @log
-    def myTestTwo(self):
-        return False
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-"""
-
-def init():
-    return """'''Short testsuite description goes here'''
-
-def run():
-   "Any initialisation code goes in this method. Is run before RTT is launched."
-   pass
-"""
-
-if __name__ == '__main__':
-    main()
diff --git a/Tools/RunTimeTester/testsuite/run/orderByUsage.py b/Tools/RunTimeTester/testsuite/run/orderByUsage.py
deleted file mode 100644
index 46db8bf00515e95912d87d1318edb230189c480a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/orderByUsage.py
+++ /dev/null
@@ -1,19 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-lines = open('junk.log', 'r').readlines()
-tList = []
-for l in lines:
-    try:
-        usage = int(l.split()[3][:-1])
-    except:
-        continue
-    print usage
-    tList.append((usage, l))        
-
-tList.sort()
-lines = [l[1] for l in tList]
-for l in lines:
-    print l
-
-    
-    
diff --git a/Tools/RunTimeTester/testsuite/run/run.py.obsolete b/Tools/RunTimeTester/testsuite/run/run.py.obsolete
deleted file mode 100644
index ff303b4c052e4225e4cc3b12de9e557146f39e8c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/run.py.obsolete
+++ /dev/null
@@ -1,126 +0,0 @@
-import os, os.path, sys, time
-
-if not 'RTTSrcDir' in os.environ.keys():
-    print 'Please set up the RTTxxx env vars.'
-    sys.exit(1)
-
-sys.path.append(os.path.abspath(os.path.dirname(__file__)))
-sys.path.append(os.environ['RTTSrcDir'])
-
-from testsuite import TestSuite
-from utility import output
-
-# ====================================================
-# ====================================================
-# ====================================================
-# global logger used by many objects: RTTxxx.log
-from LoggerSetup import setupLoggers
-import logging
-setupLoggers(os.environ['RTTLogDir'])
-
-
-class TestFramework:
-    def __init__(self):
-        self.testIds = [int(id) for id in sys.argv[1:]]
-        self.findTestSuites()
-        self.chooseTestSuites()
-        self.runTestSuites()
-
-    def findTestSuites(self):
-        self.basepath = os.path.dirname(os.path.abspath(__file__))
-        suitesdir = os.path.join(self.basepath, '../testsuites') 
-        testsuites = [os.path.join(suitesdir, suite) for suite in os.listdir(suitesdir)]
-        self.testsuites = [TestSuite(t) for t in testsuites if os.path.isdir(t)]
-
-    def askUserForTests(self):
-        for i,suite in enumerate(self.testsuites):
-            info = suite.info()
-            testsuitename = suite.name
-            print '%d) %s: \n\t%s' % (i, testsuitename, info)
-
-        bad_choice = True
-        testIds = []
-        while bad_choice:
-            question  = '\nWhich testsuites should be run?\n'
-            question += '[where * = run all testsuites; otherwise a whitespace-separated list of integers (order entered = run order)]: \n'
-            question += '> '
-            ans = raw_input(question)
-
-            # Pick all testsuites
-            if ans.strip()=='*':
-                bad_choice = False
-                testIds = range(len(self.testsuites))
-                continue
-
-            # Pick a subset of the testsuites
-            try:
-                testIds = [int(i) for i in ans.split()]
-            except:
-                pass
-            else:
-                if not [t for t in testIds if t not in range(len(self.testsuites))]: bad_choice = False
-
-        return testIds
-
-    def chooseTestSuites(self):
-        if self.testIds:
-            testIds = self.testIds
-        else:
-            testIds = self.askUserForTests()
-
-        self.testsuites = [self.testsuites[id] for id in testIds]
-
-    def runTestSuites(self):
-        self.testsuitesToXML()
-        
-
-        tss = []
-        for t in self.testsuites:
-            # t.register(self.testsuitesToXML)
-            try:
-                t.start()
-                time.sleep(10)
-                # t.run()
-            except:
-                'Exception thrown while starting thread for testsuite %s' % t.name
-                continue
-
-            tss.append(t)
-                
-        print 'successfully started testsuite thread for:'
-        for t in tss: print '   ',t.name
-        [tr.join() for tr in tss]
-        print 'threads terminated for all testsuites'
-        
-    def testsuitesToXML(self):
-        from xml.dom.minidom import Document
-        doc = Document()
-        e = doc.createElement
-        t = doc.createTextNode
-    
-        tssEl = e('current_testsuites')
-
-        for suite in self.testsuites:
-            tsEl   = e('testsuite')
-            tspEl  = e('testsuitepath') 
-            tspEl.appendChild(t(os.path.abspath(os.path.join(self.basepath, '../testsuites/%s' % suite.name, suite.xmlFileDump))))
-            tsnEl  = e('testsuitename')
-            tsnEl.appendChild(t(suite.name))
-
-            statEl = e('testsuitestatus')
-            statEl.appendChild(t(suite.status))
-            tsEl.appendChild(statEl)
-            tsEl.appendChild(tsnEl)
-            tsEl.appendChild(tspEl)
-            tssEl.appendChild(tsEl)
-        doc.appendChild(tssEl)
-
-        outpath = os.path.abspath(os.path.join(self.basepath, '../web/current.testsuites.xml'))
-        output(doc.toprettyxml('  '), outpath)
-
-# ----------------------------------------------------
-# ----------------------------------------------------
-# ----------------------------------------------------
-
-if __name__ == '__main__':
-    TestFramework()
diff --git a/Tools/RunTimeTester/testsuite/run/runCoverage.py b/Tools/RunTimeTester/testsuite/run/runCoverage.py
deleted file mode 100644
index 376d7173b07718a9ef8a346b98e34feac1ad85ec..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/runCoverage.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import glob
-import os
-import coverage
-import sys
-import run
-
-def doit():
-    srcdir = os.environ['RTTSrcDir']
-    pyfiles = glob.glob(os.path.join(srcdir, '*.py'))
-
-    coverage.erase()
-    coverage.start()
-
-    run.TestFramework()
-
-    coverage.stop()
-    coverage.analysis(run)
-
-    coverage.report(pyfiles)
-
-    coverage.erase()
-
-if __name__ == '__main__':
-    doit()
diff --git a/Tools/RunTimeTester/testsuite/run/runSuites.py b/Tools/RunTimeTester/testsuite/run/runSuites.py
deleted file mode 100644
index 157e482703ae55168f8697546bb11ba5781a6c27..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/runSuites.py
+++ /dev/null
@@ -1,156 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, os.path, sys, time
-import popen2
-import string
-
-def nowSecs():
-    return time.strftime('%d %b %H:%M:%S', time.localtime())
-
-    
-if not 'RTTSrcDir' in os.environ.keys():
-    print 'Please set up the RTTxxx env vars.'
-    sys.exit(1)
-
-sys.path.append(os.path.abspath(os.path.dirname(__file__)))
-sys.path.append(os.environ['RTTSrcDir'])
-
-from testsuite import TestSuite
-from utility import output
-
-# ====================================================
-# ====================================================
-# ====================================================
-# global logger used by many objects: RTTxxx.log
-from LoggerSetup import setupLoggers
-import logging
-setupLoggers(os.environ['RTTLogDir'])
-
-
-def printJobStatus(d):
-    print '\n---------- Job Status ---------'
-    for k in d.keys():
-        jk = string.ljust(k, 40)
-        print '%s:     %s  %s' %  (jk, d[k][0], d[k][1])
-class TestFramework:
-    def __init__(self):
-        self.testIds = [int(id) for id in sys.argv[1:]]
-        self.findTestSuites()
-        self.chooseTestSuites()
-        self.runTestSuites()
-
-    def findTestSuites(self):
-        self.basepath = os.path.dirname(os.path.abspath(__file__))
-        suitesdir = os.path.normpath(os.path.join(self.basepath, '../testsuites') )
-        testsuites = [os.path.join(suitesdir, suite) for suite in os.listdir(suitesdir) if not suite.startswith('.svn')]
-        self.testsuites = [TestSuite(t) for t in testsuites if os.path.isdir(t)]
-
-    def askUserForTests(self):
-        for i,suite in enumerate(self.testsuites):
-            info = suite.info()
-            testsuitename = suite.name
-            print '%d) %s: \n\t%s' % (i, testsuitename, info)
-
-        bad_choice = True
-        testIds = []
-        while bad_choice:
-            question  = '\nWhich testsuites should be run?\n'
-            question += '[where q = quit; * = run all testsuites; otherwise a whitespace-separated list of integers (order entered = run order)]: \n'
-            question += '> '
-            ans = raw_input(question)
-
-            if ans.lower().strip()=='q':
-                sys.exit(0)
-                
-            # Pick all testsuites
-            if ans.strip()=='*':
-                bad_choice = False
-                testIds = range(len(self.testsuites))
-                continue
-
-            # Pick a subset of the testsuites
-            try:
-                testIds = [int(i) for i in ans.split()]
-            except:
-                pass
-            else:
-                if not [t for t in testIds if t not in range(len(self.testsuites))]: bad_choice = False
-
-        return testIds
-
-    def chooseTestSuites(self):
-        if self.testIds:
-            testIds = self.testIds
-        else:
-            testIds = self.askUserForTests()
-
-        self.testsuites = [self.testsuites[id] for id in testIds]
-
-    def runTestSuites(self):
-        """
-        Kludge in running testsuistes as a sub process.
-        Testsuite objects have already been made,and used to allow test selection.
-        Having done this, the objects are dumped, and new ones are started up in
-        sub shells.
-        """
-        
-        self.testsuitesToXML()
-        
-
-        tss = []
-        jobStatus = {}
-        for t in self.testsuites:
-            d = t.testsuitedir
-            runner = os.path.join(self.basepath, 'testSuiteRunner.py')
-            tdir  = os.path.join(d, 'testsuite.out')
-            cmd = 'python %s %s>& %s' % (runner, d,tdir)
-            #print 'about to popen the cmd: %s' % cmd
-            tss.append((t.name, popen2.Popen3(cmd)))
-            jobStatus[t.name] = ('running', nowSecs())
-        ntests = len(tss)
-        printJobStatus(jobStatus)
-
-        while tss:
-            toRemove = [p for p in tss if p[1].poll() != -1]
-            if toRemove:
-                [tss.remove(p) for p in toRemove]
-                for p in toRemove:
-                    jobStatus[p[0]] = ('completed', nowSecs())
-
-                printJobStatus(jobStatus)
-            time.sleep(10)
-
-        print 'all %d tests have completed' % ntests
-        
-    def testsuitesToXML(self):
-        from xml.dom.minidom import Document
-        doc = Document()
-        e = doc.createElement
-        t = doc.createTextNode
-    
-        tssEl = e('current_testsuites')
-
-        for suite in self.testsuites:
-            tsEl   = e('testsuite')
-            tspEl  = e('testsuitepath') 
-            tspEl.appendChild(t(os.path.abspath(os.path.join(self.basepath, '../testsuites/%s' % suite.name, suite.xmlFileDump))))
-            tsnEl  = e('testsuitename')
-            tsnEl.appendChild(t(suite.name))
-
-            statEl = e('testsuitestatus')
-            statEl.appendChild(t(suite.status))
-            tsEl.appendChild(statEl)
-            tsEl.appendChild(tsnEl)
-            tsEl.appendChild(tspEl)
-            tssEl.appendChild(tsEl)
-        doc.appendChild(tssEl)
-
-        outpath = os.path.abspath(os.path.join(self.basepath, '../web/current.testsuites.xml'))
-        output(doc.toprettyxml('  '), outpath)
-
-# ----------------------------------------------------
-# ----------------------------------------------------
-# ----------------------------------------------------
-
-if __name__ == '__main__':
-    TestFramework()
diff --git a/Tools/RunTimeTester/testsuite/run/runTests.py b/Tools/RunTimeTester/testsuite/run/runTests.py
deleted file mode 100644
index 66ac595897332b0b1096f33c816be8012a5853df..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/runTests.py
+++ /dev/null
@@ -1,70 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-
-import sys
-import os
-import commands
-import pprint
-import getopt
-
-def optusage():
-    print 'usage: python runTests -d [path to dir above testsuites]'
-    
-def getTopOfSuitesDir():
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'd:')
-    except getopt.GetoptError:
-        optusage()
-        sys.exit(2)
-
-    for o,a in opts:
-        if o == '-d': return a
-        
-    optusage()
-    sys.exit(2)
-
-class RunTests:
-    def __init__(self):
-        cwd = os.getcwd()
-        sys.path.append(cwd)
-        tspath = getTopOfSuitesDir()
-        assert(os.path.exists(tspath))
-        ls = os.listdir(tspath)
-        ls = [os.path.join(tspath,f) for f in ls] 
-        dirs = [(os.path.join(cwd, d), []) for d in ls if os.path.isdir(d) and d.find('.svn') == -1]
-
-        toFind = ['TestCase.py', 'RTTSummaryReference.xml']
-        for fn in toFind:
-            print
-            for (dir, missing) in dirs:
-                if not os.path.exists(os.path.join(dir,fn)): missing.append(fn)
-
-        fn = 'RTTSummary.xml'
-        for (dir, missing) in dirs:
-            if len(listFiles(dir, fn)) != 1:
-                missing.append(fn)
-
-        self.okdirs = [d for (d, missing) in dirs if len(missing) == 0]
-        self.baddirs = [ (d, missing) for (d, missing) in dirs if len(missing) != 0]
-
-    def runTests(self):
-        if len(self.baddirs) == 0:
-            print 'All directories have valid tests'
-            print
-        else:
-            print
-            for (d, missing) in self.baddirs:
-                print 'Could not run tests in %s: missing files:' % d
-                pprint.pprint(missing)
-                print
-        
-        for d in self.okdirs:
-            os.chdir(d)
-
-            s, o = commands.getstatusoutput('python TestCase.py')
-            print o
-
-
-runner = RunTests()
-runner.runTests()
diff --git a/Tools/RunTimeTester/testsuite/run/stats.py b/Tools/RunTimeTester/testsuite/run/stats.py
deleted file mode 100644
index 17b328bd9ccb53c6b964a26d2df7cc2069198426..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/stats.py
+++ /dev/null
@@ -1,71 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from xml.dom.minidom import Document
-
-class TestStatistics:
-    def __init__(self, testname):
-        self.testsuitetestname = testname
-        self.results = {}
-
-    def log(self): pass
-
-    def __iadd__(self, (name, result, message)):
-        self.results[name] = (result, message)
-        return self
-
-    def __str__(self): 
-        m  = '******************************\n'
-        m += '**** Results for tests.%s ****\n' % self.testsuitetestname
-        m += '******************************\n'
-        m += 'Total: %d, PASS: %d, FAIL: %d \n\n' % (self.summary())
-        
-        for testname, (status, message) in self.results.items():
-            m += '%s: %s' % ({True:'PASS', False:'FAIL'}.get(status), testname)
-            if message: m+=message
-            m += '\n'
-        return m
-
-    def summary(self):
-        totalTests = len(self.results.keys())
-        totalPass  = len([k for k,v in self.results.items() if v[0]])
-        totalFail  = len([k for k,v in self.results.items() if not v[0]])
-        return (totalTests, totalPass, totalFail)
-
-    def toxml(self):
-        doc = Document()
-        e = doc.createElement
-        t = doc.createTextNode
-        testsuitetestEl = e('testsuitetest')
-        testsuitetestnameEl = e('testsuitetestname')
-        testsuitetestnameEl.appendChild(t(self.testsuitetestname))
-        testsuitetestEl.appendChild(testsuitetestnameEl)
-
-        # Make the summary
-        tot, ok, fail = self.summary()
-        testsuitetestsummaryEl = e('summary')
-        totEl = e('totaltests')
-        totEl.appendChild(t(str(tot)))
-        totokEl = e('totalok')
-        totokEl.appendChild(t(str(ok)))
-        totfailEl = e('totalfail')
-        totfailEl.appendChild(t(str(fail)))
-
-        [testsuitetestEl.appendChild(child) for child in [totEl, totokEl, totfailEl]]
-
-        # Now the individual tests
-        for testname, (status, message) in self.results.items():
-            testEl = e('test')
-            testnameEl = e('testname')
-            testnameEl.appendChild(t(testname))
-            statEl = e('status')
-            statEl.appendChild(t(str(status)))
-            messEl = e('message')
-            messEl.appendChild(t(message))
-
-            [testEl.appendChild(f) for f in [testnameEl, statEl, messEl]]
-
-            testsuitetestEl.appendChild(testEl)
-
-        return testsuitetestEl
-    # doc.appendChild(testsuitetestEl)
-    #  return doc
diff --git a/Tools/RunTimeTester/testsuite/run/summaryFile.py b/Tools/RunTimeTester/testsuite/run/summaryFile.py
deleted file mode 100644
index 47bf17f915af59da742835c4d9a24d5b20e4be33..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/summaryFile.py
+++ /dev/null
@@ -1,70 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os
-import os.path
-import sys
-
-from Evaluate import Evaluate # from RTTSrcDir code base
-
-class RTTSummaryFile:
-    def __init__(self, path):
-        if not os.path.exists(path):
-            print '%s, inexistent' % path
-            sys.exit(1)
-
-        self.path = os.path.abspath(path)
-        self.docEl_ = self.docEl(path)
-        self.overviews = [Overview(o) for o in Evaluate('overview', self.docEl_)]
-        self.packages  = [Package(p) for p in Evaluate('package', self.docEl_)]
-
-    def docEl(self, path):
-        from xml.dom.minidom import parse
-        try:
-            return parse(path).documentElement
-        except:
-            print 'Unable to parse %s' % path
-            return None
-
-    def getPackageByName(self, name):
-        return [p for p in self.packages if p.name==name]
-
-
-class Thing: # convenience parent class
-    def __init__(self, elem):
-        self.xml = elem
-    def __getitem__(self, xpath):
-        return Evaluate(xpath, self.xml)
-
-class Package(Thing):
-    def __init__(self, packageEl):
-        Thing.__init__(self, packageEl)
-        self.jobs = [Job(j) for j in Evaluate('minder', packageEl)]
-        self.name = self['packageName/text()']
-
-    def getJobsByFullName(self, name):
-        return [j for j in self.jobs if j.fullname==fullname]
-    def getJobsByName(self, name):
-        return [j for j in self.jobs if j.name==name]
-    def getJobsByStatus(self, status):
-	return [j for j in self.jobs if j.status==status]
-    def getAllSuccessJobs(self):
-	return self.getJobsByStatus('success')
-    def getAllFailedJobs(self):
-	return self.getJobsByStatus('error')
-    def hasOnlySuccessJobs(self):
-	return len(self.getAllSuccessJobs()) == len(self.jobs)
-
-class Job(Thing):
-    def __init__(self, jobEl):
-        Thing.__init__(self, jobEl)
-        self.fullname = self['identifiedName/text()']
-        self.name     = self['jobName/text()']
-	self.status   = self['status/text()']
-    def isSuccess(self):
-        return self.status == 'success'
-    def isFailed(self):
-        return self.status == 'error'
-
-class Overview(Thing):
-    def __init__(self, ovEl):
-        Thing.__init__(self, ovEl)
diff --git a/Tools/RunTimeTester/testsuite/run/testSuiteRunner.py b/Tools/RunTimeTester/testsuite/run/testSuiteRunner.py
deleted file mode 100755
index 96a98ad454c3736e66ceb84e26f736002c22abfd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/testSuiteRunner.py
+++ /dev/null
@@ -1,13 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-'a glue function to allow running a testsuite via popen - ie as subprocess'
-
-import sys
-from testsuite import TestSuite
-def doit():
-    path = sys.argv[1]
-    ts = TestSuite(path)
-    ts.run()
-    
-if __name__ == '__main__':
-    doit()
diff --git a/Tools/RunTimeTester/testsuite/run/testsuite.py b/Tools/RunTimeTester/testsuite/run/testsuite.py
deleted file mode 100644
index 5c6c52adee715088222b5468e31b52896344a47a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/testsuite.py
+++ /dev/null
@@ -1,210 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-from utility         import timeit, SysPath, SysModules, output
-from summaryFile     import RTTSummaryFile
-from xml.dom.minidom import Document
-
-import os
-import sys
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class Status:
-    PENDING       = 'pending'
-    RUNNING_INIT  = 'running_init'
-    RUNNING_JOB   = 'running_job'
-    RUNNING_TESTS = 'running_tests'
-    DONE          = 'done'
-
-class TestSuite:
-    def __init__(self, path):
-        self.name         = os.path.basename(path)
-        self.syspath      = SysPath()
-        self.sysmodules   = SysModules()
-        self.testsuitedir = path  # base dir in which all test stuff sits
-        self.tests        = []
-        self.status       = Status.PENDING
-        self.xmlFileDump  = 'testsuite.status.xml'
-        self.listeners    = []
-        self.tester       = None
-        self.dumpXML()
-        
-    def register(self, callback):
-        self.listeners.append(callback)
-        
-    def inform(self):
-        [listener() for listener in self.listeners]
-
-    def info(self):
-        if not os.path.exists(os.path.join(self.testsuitedir, '__init__.py')):
-            return 'No __init__.py file found in %s, no test description available.' % self.name
-
-        self.syspath += self.testsuitedir
-        import __init__
-        info = __init__.__doc__
-        
-        self.syspath -= self.testsuitedir
-
-        self.sysmodules -= '__init__'
-
-        return info
-
-    def run(self):
-        print
-        print
-        print '********************************************'
-        print 'Launching testsuite: %s' % self.name
-        print '********************************************'
-        print
-        print
-        self.syspath += self.testsuitedir
-
-        os.chdir(self.testsuitedir)
-        self.runInit()
-        os.chdir(self.testsuitedir)
-        self.runJob()
-        #os.chdir(self.testsuitedir)
-        #self.runTests()
-        os.chdir(self.testsuitedir)
-        self.syspath -= self.testsuitedir
-        self.status = Status.DONE        
-        self.dumpXML()
-
-
-    @timeit
-    def runInit(self):
-        self.status = Status.RUNNING_INIT
-        self.dumpXML()
-        import __init__
-        initcode = None
-        try:
-            initcode = __init__.run
-        except AttributeError:
-            print '>>>> No init code to run in __init__.py'
-            return
-            
-        print '>>>> Running testsuite initialise code'
-        initcode()
-        self.sysmodules -= '__init__'
-
-    @timeit
-    def runJob(self):
-        self.status = Status.RUNNING_JOB
-        self.dumpXML()
-
-	# Output RTT run logs to testsuite dir
-        # this is really ugly.....
-        # RTTRunner gets the log dir from an envirnment variable in __init__
-        # but the testsuites are threaded, and all the threads are busy modifying the
-        # (global) environment variable RTTLogDir. Here I make the distasteful action
-        # of resetting the logdir in RTTDRunner after instantiation.
-        # old...
-
-
-
-
-
-
-        from RTTRunner_NoEnv import Runner
-        os.environ['RTTCfgFile'] = os.path.normpath(os.path.join(self.testsuitedir, 'cfg/cfg.xml'))
-        print
-        print '>>>>> Launching job using cfg file: %s ....' % os.environ['RTTCfgFile']
-        print
-        RTTSrcDir   = os.environ['RTTSrcDir']
-        RTTShareDir = os.environ['RTTShareDir']
-        RTTCfgFile  = os.path.normpath(os.path.join(self.testsuitedir, 'cfg/cfg.xml'))
-        RTTLogDir   = os.path.join(self.testsuitedir, 'logs')
-        useDB       = os.environ.has_key('RTTFactoryUseDB')
-        runner = Runner(RTTSrcDir, RTTShareDir, RTTCfgFile, RTTLogDir, useDB, self.name)
-        runner.logdir = os.path.join(self.testsuitedir, 'logs')
-        self.tester = runner.makeTester()        
-        self.dumpXML()
-        runner.runTester(self.tester)
-
-        print '>>>>> Finished running job from config file %s ....' % runner.topLevelRTTconfFileLocation
-
-        print
-        self.sysmodules -= 'RTTRunner' # remove from sys.modules
-
-    def loadTests(self):
-        import tests
-        self.tests = [t(self.tester) for t in tests.run() if t.__name__ != 'ExampleTest']
-
-    @timeit
-    def runTests(self):
-        print
-        print '>>>>> Launching post-RTT run tests.py...'
-
-        testsScript = os.path.join(self.testsuitedir, 'tests.py')
-        if not os.path.exists(testsScript):
-            print 'No tests were found in %s, nothing to do.' % self.testsuitedir
-            return
-
-        self.status = Status.RUNNING_TESTS
-        self.dumpXML()
-        
-        self.loadTests()
-
-        for test in self.tests:
-            print 'Running tests.%s' % test.__class__.__name__
-            test.run()
-            print str(test.stats)
-        print
-        print ' >>>>> Finished running tests.py'
-
-        self.dumpXML()
-        self.sysmodules -= 'tests' # remove from sys.modules
-        
-    def dumpXML(self):
-        doc  = Document()
-        e = doc.createElement
-        t = doc.createTextNode        
-        tsEl   = e('testsuite')
-        tsnEl  = e('testsuitename')
-        tsnEl.appendChild(t(self.name))
-        statEl = e('testsuitestatus')
-        statEl.appendChild(t(self.status))
-        tsEl.appendChild(tsnEl)
-        tsEl.appendChild(statEl)
-        
-        # Report on the job(s)
-        if self.tester:
-            if hasattr(self.tester, 'xmlConverter') and os.path.exists(self.tester.xmlConverter.RTTSummary):
-                h = open(self.tester.xmlConverter.RTTSummary, 'r')
-                xmlfile = h.read()
-                h.close()      
-                summFileEl = e('rttSummaryFile')
-
-                summFilePathEl = e('pathToSummaryFile')
-                
-                resdir = os.path.abspath(os.path.join(self.testsuitedir, 'Results'))
-                relativeSummFilePath = os.path.abspath(self.tester.xmlConverter.RTTSummary).split(resdir)[1].strip()
-                if relativeSummFilePath.startswith('/'): relativeSummFilePath = relativeSummFilePath[1:]
-                summFilePathEl.appendChild(t(relativeSummFilePath))
-                summFileEl.appendChild(summFilePathEl)
-
-                # summFileContentsEl = e('summaryFileContents')
-                # summFileContentsEl.appendChild(t(xmlfile))
-                # summFileEl.appendChild(summFileContentsEl)
-
-                tsEl.appendChild(summFileEl)
-        
-        # Report on the tests
-        if self.tests:
-            testsEl = e('tests')
-            for test in self.tests:
-                testsEl.appendChild(test.stats.toxml())
-            tsEl.appendChild(testsEl)
-
-        doc.appendChild(tsEl)
-        output(doc.toprettyxml('  '), os.path.join(self.testsuitedir, self.xmlFileDump))
-        self.inform() # update any listeners for state changes
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-
diff --git a/Tools/RunTimeTester/testsuite/run/testsuitetest.py b/Tools/RunTimeTester/testsuite/run/testsuitetest.py
deleted file mode 100644
index 75046478bb97c4f6273bd2c7df23501d86efea32..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/testsuitetest.py
+++ /dev/null
@@ -1,38 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from summaryFile import RTTSummaryFile
-from stats import TestStatistics
-
-class Test:
-    def __init__(self, tester):
-        self.tester      = tester
-        self.summaryFile = RTTSummaryFile(tester.xmlConverter.RTTSummary)
-        self.stats       = TestStatistics(self.__class__.__name__)
-
-    def run(self): 
-        for testcase in [getattr(self, t) for t in dir(self.__class__) if t.startswith('test') and callable(getattr(self,t))]:
-            retvalue = None
-            try:
-                retvalue = testcase()
-            except Exception, e:
-                self.stats += (testcase.__name__, False, 'Test code broken? Raised an exception: %s' % str(e))
-            else:            
-                self.stats += (testcase.__name__, retvalue, '')
-
-class SummaryFileTest(Test):
-    def __init__(self, tester):
-	Test.__init__(self, tester)
-
-    def hasExactlyNPackages(self, n): return len(self.summaryFile.packages) == n
-
-    def hasOnlyOnePackageCalled(self, name):
-        return len([p for p in self.summaryFile.packages if p.name==name])==1
-
-    def hasAtLeastNPackages(self, npacks):
-	return len(self.summaryFile.packages)>=npacks
-
-    def hasNJobs(self, njobs):
-        return reduce(lambda x,y:x+y, [len(p.jobs) for p in self.summaryFile.packages], 0) == njobs
-
-    def testHasOnlySuccessJobs(self):
-        return False not in [p.hasOnlySuccessJobs() for p in self.summaryFile.packages]
diff --git a/Tools/RunTimeTester/testsuite/run/utility.py b/Tools/RunTimeTester/testsuite/run/utility.py
deleted file mode 100644
index 048ca8737cbad680683819990ebf88409daf80bf..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/run/utility.py
+++ /dev/null
@@ -1,68 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os, sys, os.path, exc2string2
-
-# ----- Some useful decorator functions
-
-def timeit(func): # time the function
-    def wrapped(*args, **kwds):
-        import time
-        start = time.time()
-        ret = None
-        try:
-            ret = func(*args,**kwds)
-        except Exception, e:
-            print '%s threw an exception!' % func.__name__
-            print str(e)
-            print exc2string2.exc2string2()
-        print '%f secs to run %s.' % (time.time()-start, func.__name__)
-        return ret
-    return wrapped
-
-def log(func): # log the result of a check
-    def wrapped(*a,**k):
-        ret = None
-        try:
-            ret = func(*a,**k)
-        except Exception, e:
-            print 'FAIL : %s [%s]' % (func.__name__, 'Test code broken? Raised an exception: %s' % str(e))
-        else:            
-            print '%s%s' % ({True:'PASS : ', False:'FAIL : '}.get(ret), func.__name__)
-        return ret
-    return wrapped
-
-def output(what, where):
-    h = open(where, 'w')
-    h.write(what)
-    h.close()
-
-# --- Sys path manipulations
-
-class SysPath:
-    def __contains__(self, path):
-        return path in sys.path
-
-    def __iadd__(self, path):
-        path = os.path.abspath(path)
-        if not path in self:
-            sys.path = [path] + sys.path
-        return self
-
-    def __isub__(self, path):
-        path = os.path.abspath(path)
-        sys.path = [s for s in sys.path if s!=path]
-        return self
-
-    def __str__(self):
-        return '%d: %s' % (len(sys.path), str(sys.path))
-
-# --- Sys Modules manipulations
-class SysModules:
-    def __contains__(self, modulename):
-        return modulename in sys.modules.keys()
-
-    def __isub__(self, modulename):
-        if modulename in self:
-            del sys.modules[modulename]
-        return self
-
diff --git a/Tools/RunTimeTester/testsuite/src/RTTTestCase.py b/Tools/RunTimeTester/testsuite/src/RTTTestCase.py
deleted file mode 100644
index 2b52da91cbc43425a8ed8252d0f8e622aa69b428..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/RTTTestCase.py
+++ /dev/null
@@ -1,129 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-from parseHelpers import getText
-from parseHelpers import getSingleNodeText
-
-import unittest
-import xml.dom.minidom
-import difflib
-import pprint
-import string
-import os
-
-def addCR(l): return l+'\n'
-def toLines(cList):
-    line = ''
-    for c in cList: line += c
-    
-    lines = line.split('\n')
-    lines = [l.strip() for l in lines]
-    lines = [l for l in lines if l != '']
-    lines = [addCR(l)  for l in lines]
-
-    return lines
-
-def pruneDom(tNames, dom):
-        for t in tNames:
-            nodes = dom.getElementsByTagName(t)
-            for n in nodes:
-                parentNode = n.parentNode
-                # print 'removing child %s from parent %s' % (n.nodeName, parentNode.nodeName)
-                parentNode.removeChild(n)
-                n.unlink()
-        return dom
-    
-def fnsFromNodes(nodes):
-    # return a list of file names from a sequence of nodes
-    # constaining paths, with duplicates removed
-    names = {}
-    for n in nodes:
-        fn = string.strip(getText(n))
-        if len(fn) == 0: continue 
-        names[os.path.basename(fn)] = 1
-
-    return names.keys()
-    
-
-class RTTTestCase(unittest.TestCase):
-    
-    def setUp(self):
-        if 'dom' in self.__dict__.keys(): return
-        infileName = listFiles('.', 'RTTSummary.xml')[0]
-        infile = open(infileName, 'r')
-
-        refFile = open('RTTSummaryReference.xml')
-
-        self.dom     = xml.dom.minidom.parseString(infile.read())
-        self.refDom  = xml.dom.minidom.parseString(refFile.read())
-
-        self.logfile = open('TestCase.log','w')
-
-        
-    def rttTestNTags(self, node,  tag, n):
-        self.assert_(len(node.getElementsByTagName(tag)) == n)
-
-    def rttTestCheckTagContent(self, tag, content):
-        text =  getSingleNodeText(self.dom, tag)
-        self.assert_(text.find(content)>-1)
-
-    def rttTestNoErrorReport(self):
-        enodes = self.dom.getElementsByTagName('errors')
-        if len(enodes)>0:
-            for node in enodes:
-                self.assert_(len(getText(node)) == 0)
-
-    def rttKeepfileDiff(self):
-        tag = 'keepfile_newstyle'
-        skfiles = fnsFromNodes(self.dom.getElementsByTagName(tag))
-        rkfiles = fnsFromNodes(self.refDom.getElementsByTagName(tag))
-
-        mismatch = [s for s in skfiles if s not in rkfiles]
-        if(mismatch): self.logfile.write('\nkeepfile in summary and not in ref\n%s\n' % str(mismatch))
-        self.assert_(len(mismatch)==0)
-
-        mismatch = [s for s in rkfiles if s not in skfiles]
-        if(mismatch): self.logfile.write('\nkeepfile in ref and not in summary\n%s\n' % str( mismatch))
-        self.assert_(len(mismatch)==0)
-
-    def variableContentTags(self):
-        return ['statusText',
-                'latestUpdateTime1',
-                'latestUpdateTime2',
-                'endTime1',
-                'endTime2',
-                'startTime1',
-                'startTime2',
-                'nicosDate1',
-                'nicosDate2',
-                'stateHistory',
-                'wallTime',
-                'mem',
-                'vmem',
-                'cpuTime',
-                'resBaseDir',
-                'resReleaseDir',
-                'resultsPath',
-                'cpuTime2000',
-                'keepfile_newstyle',
-                'baseKeepFileDir',
-                'RTTLibVersion',
-                'RTTSrcVersion'
-                ]
-
-    def rttTestXMLComp(self):
-
-        tagsToPrune = self.variableContentTags()
-
-        prunedDom    = pruneDom(tagsToPrune, self.dom)
-        prunedRefDom = pruneDom(tagsToPrune, self.refDom)
-
-        newxml = toLines(prunedDom.toxml())
-        refxml = toLines(prunedRefDom.toxml())
-
-        d = difflib.Differ()
-        difflist = list(d.compare(newxml, refxml) )
-        difflist = [d for d in difflist if not d[0] == ' ']
-        pprint.pprint(difflist, self.logfile)
-
-        self.assert_(not difflist)
diff --git a/Tools/RunTimeTester/testsuite/src/keepFileDiff.py b/Tools/RunTimeTester/testsuite/src/keepFileDiff.py
deleted file mode 100644
index 2bbe408c104dd402d8629a44460d46442d7385f2..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/keepFileDiff.py
+++ /dev/null
@@ -1,91 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-
-import sys
-import os
-import getopt
-import string
-import xml.sax
-import pprint
-from xml.sax.handler import ContentHandler
-
-class KFHandler(ContentHandler):
-    def __init__(self):
-        self.kfiles = []
-        self.enable = False
-    def startElement(self, name, attr):
-        # print 'handling tag', name
-        if name == 'keepfile_newstyle':
-            self.enable = True
-    def endElement(self, name):
-            self.enable = False
-    def characters(self, ch):
-        if not self.enable:return
-        ch = string.strip(ch)
-        if len(ch)>0: self.kfiles.append(ch)
-        # if len(ch)>0: self.kfiles.append(os.path.basename(ch))
-    
-def optusage():
-    print 'usage: python keepFileDiff -d [path to dir with RTTSummaryReference.xml]'
-    
-def getSuiteDir():
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 'd:')
-    except getopt.GetoptError:
-        optuasge()
-        sys.exit(2)
-
-    for o,a in opts:
-        if o == '-d': return a
-        
-    optusage()
-    sys.exit(2)
-
-def getKeepFiles(fn):
-    lines = open(fn, 'r').readlines()
-    lines = [string.strip(l) for l in lines if l.find('<keepfile') != -1]
-    
-    print 'getting keep files from file ', fn
-
-    return lines
-
-def doit():
-    cwd = os.getcwd()
-    sdir = os.path.join(cwd, getSuiteDir())
-    if not os.path.exists(sdir):
-        print 'Non -existant test suite directory ', sdir
-        sys.exit(0)
-    rfile = os.path.join(sdir, 'RTTSummaryReference.xml')
-    if not os.path.exists(os.path.join(sdir,rfile)):
-        print 'No reference file ', rfile
-        sys.exit(0)
-
-    parser = xml.sax.make_parser()
-    handler = KFHandler()
-    parser.setContentHandler(handler)
-    parser.parse(rfile)
-    rkfiles = handler.kfiles
-
-    lfiles = listFiles(sdir, 'RTTSummary.xml')
-    if len(lfiles) != 1:
-        print 'error looking for RTTSummaryFile', lfiles
-        sys.exit(0)
-        
-    handler = KFHandler()
-    parser.setContentHandler(handler)
-    parser.parse(lfiles[0])
-    skfiles = handler.kfiles
-    
-    diff = [f for f in rkfiles if f not in skfiles]
-    pp = pprint.PrettyPrinter()
-    print
-    print ' in Reference and not in Summary'
-    pp.pprint(diff)
-
-    diff = [f for f in skfiles if f not in rkfiles]
-    print
-    print ' in Summary and not in Reference'
-    pp.pprint(diff)
-    
-doit()
diff --git a/Tools/RunTimeTester/testsuite/src/listFiles.py b/Tools/RunTimeTester/testsuite/src/listFiles.py
deleted file mode 100644
index 8d9c0235888beb6fa8db7e7a7c09433171400903..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/listFiles.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import os.path
-import fnmatch
-
-
-def listFiles(root, patterns='*', recurse=1, return_folders=0):
-    pattern_list = patterns.split(';')
-    class Bunch:
-        def __init__(self, **kwds): self.__dict__.update(kwds)
-
-    arg = Bunch(recursion=recurse, pattern_list=pattern_list,
-                return_folders=return_folders, results=[])
-
-    def visit(arg, dirname, files):
-        for name in files:
-            fullname = os.path.normpath(os.path.join(dirname, name))
-            if arg.return_folders or os.path.isfile(fullname):
-                for pattern in arg.pattern_list:
-                    if fnmatch.fnmatch(name, pattern):
-                        arg.results.append(fullname)
-                        break
-        if not arg.recursion: files[:] = []
-
-    os.path.walk(root, visit, arg)
-    return arg.results
diff --git a/Tools/RunTimeTester/testsuite/src/parseHelpers.py b/Tools/RunTimeTester/testsuite/src/parseHelpers.py
deleted file mode 100644
index ba6fd914aa8d0ce7e0a4d99065ad8383484913a8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/parseHelpers.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-
-def getText(node):
-
-    s = ''
-    for n in node.childNodes:
-        if n.nodeType == node.TEXT_NODE:
-            s += n.data
-    return s
-
-
-def getSingleNodeText(node, tag):
-            nodes = node.getElementsByTagName(tag)
-            snode = nodes[0]
-            return getText(snode)
-
diff --git a/Tools/RunTimeTester/testsuite/src/runTestCase.py b/Tools/RunTimeTester/testsuite/src/runTestCase.py
deleted file mode 100644
index 29e45a7c1df755b5474542b51df10ae9fe0b0bff..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/runTestCase.py
+++ /dev/null
@@ -1,27 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-import unittest
-import os
-def runTestCase(TestCase):
-    suite = unittest.makeSuite(TestCase, 'test')
-    result = unittest.TestResult()
-    suite.run(result)
-    suitePath = os.getcwd()
-    suiteName = os.path.basename(suitePath)
-    ntests    = result.testsRun
-    nerrors   = len(result.errors)
-    nfailures = len(result.failures)
-    
-    if result.errors:
-        of = open(os.path.join(suitePath, 'TestCaseErrors.log'), 'w')
-        [of.write(t[1]) for t in result.errors]
-        
-    if result.failures:
-        of = open(os.path.join(suitePath, 'TestCaseFailures.log'), 'w')
-        [of.write(t[1]) for t in result.failures]
-        
-    if result.wasSuccessful():
-        ok = 'OK'
-    else:
-        ok = 'check %s ' % os.path.join(suiteName,'TestCase.log')
-    print '%s Errors/Failures/Successes    %d %d %d %s' % (suiteName.ljust(30), nerrors, nfailures, (ntests-nerrors-nfailures), ok)
diff --git a/Tools/RunTimeTester/testsuite/src/updateRTTSummaryReferenceInSrcDir.py b/Tools/RunTimeTester/testsuite/src/updateRTTSummaryReferenceInSrcDir.py
deleted file mode 100644
index 4613bbec2aadbb687c4ee2b67f3f98c35459ed8e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/updateRTTSummaryReferenceInSrcDir.py
+++ /dev/null
@@ -1,56 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-import os
-import shutil
-import sys
-import getopt
-
-def optusage():
-    print 'usage: python updateRTTSummaryReferenceInSrcDir -t [rel path to testsuite dir] --live[to really make the copy]'
-    
-def getParams():
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 't:', ['live'])
-    except getopt.GetoptError:
-        optusage()
-        sys.exit(2)
-
-    suite = None
-    live = False
-    for o,a in opts:
-        if o == '-t':     suite = a
-        if o == '--live': live = True
-
-    if not suite:
-        optusage()
-        sys.exit(2)
-
-    return (suite, live)
-
-
-suite, live = getParams()
-suiteDir = os.path.abspath (suite)
-assert (os.path.exists(suite))
-
-files = listFiles(suiteDir, 'RTTSummary.xml')
-
-assert (len(files) == 1)
-dest1 = os.path.normpath(os.path.join(os.environ['RTTSrcDir'], '../testsuite/testsuites', os.path.basename(suiteDir), 'RTTSummaryReference.xml'))
-dest2 = os.path.join(suiteDir, 'RTTSummaryReference.xml')
-src = files[0]
-
-
-
-if live:
-    print 'Copier IS live, copying %s to %s' % (src, dest1)
-    shutil.copyfile(src, dest1)
-    print 'Copier IS live, copying %s to %s' % (src, dest2)
-    shutil.copyfile(src, dest2)
-            
-else:
-    print '\n\nCopier is NOT  live, copying from:\n%s\n' % (src)
-    print '\n\nCopier is NOT  live, copying to:\n%s' % (dest1)
-    print '\n\nCopier is NOT  live, copying to:\n%s' % (dest2)
-    print '\nCopier is NOT  live, NO copying was done'
-
diff --git a/Tools/RunTimeTester/testsuite/src/updateTestCase.py b/Tools/RunTimeTester/testsuite/src/updateTestCase.py
deleted file mode 100644
index fa22f4a33371a5920835e1ec25d445b44994eb3d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/src/updateTestCase.py
+++ /dev/null
@@ -1,53 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from listFiles import listFiles
-import os
-import shutil
-import sys
-import getopt
-
-def optusage():
-    print 'usage: python updateRTTSummaryReferenceInSrcDir -t [rel path to testsuite dir] --live[to really make the copy]'
-    
-def getParams():
-    try:
-        opts, args = getopt.getopt(sys.argv[1:], 't:', ['live'])
-    except getopt.GetoptError:
-        optusage()
-        sys.exit(2)
-
-    suite = None
-    live = False
-    for o,a in opts:
-        if o == '-t':     suite = a
-        if o == '--live': live = True
-
-    if not suite:
-        optusage()
-        sys.exit(2)
-
-    return (suite, live)
-
-
-suite, live = getParams()
-suiteDir = os.path.abspath (suite)
-assert (os.path.exists(suite))
-
-destDir = os.path.normpath(os.path.join(os.environ['RTTSrcDir'], '../testsuite/testsuites'))
-assert(os.path.exists(destDir))
-dest = os.path.normpath( os.path.join(destDir, os.path.basename(suiteDir), 'TestCase.py'))
-
-src = os.path.join(suiteDir, 'TestCase.py')
-assert(os.path.exists(src))
-
-
-
-if live:
-    print 'Copier IS live, copying %s to %s' % (src, dest)
-    shutil.copyfile(src, dest)
-            
-else:
-    print '\n\nCopier is NOT  live, copying from:\n%s\n' % (src)
-    print '\n\nCopier is NOT  live, copying to:\n%s'     % (dest)
-    print '\nCopier is NOT  live, NO copying was done'
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/Package/package.xml.TPL
deleted file mode 100644
index ef1df50381ae1d07498a4b226e2bf147733a4994..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/Package/package.xml.TPL
+++ /dev/null
@@ -1,56 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-        <chain>
-	  <chainName>MissingFileMergeJob</chainName>
-	  <sequential>
-	  <chainElement>
-
-	    <jobTransform>
-	      <doc>MergeJob</doc>
-	      <jobTransformJobName>MergeJob</jobTransformJobName>
-	      <jobTransformCmd>
-		MergePool_trf.py maxEvents=-1 inputAODFile=job1.AOD.pool.root,job2.AOD.pool.root geometryVersion=ATLAS-GEO-03-00-00 outputAODFile=MergedAOD.pool.root preInclude=RecExCommission/MinimalCommissioningSetup.py preExec='rec.doEgamma.set_Value_and_Lock(False)' --ignoreunknown
-	      </jobTransformCmd>
-
-	      <group>JobTransform</group>
-	      
-	      <chaindataset_info>
-		<jobTransformData/>
-		<chaindatasetName>job1.AOD.pool.root</chaindatasetName>
-		<dataset_info>
-		  <jobTransformData/>
-                  <datasetName>/afs/cern.ch/atlas/offline/external/FullChainTest/tier0-vol3/14.5.X/root_archive/archive/RPCwBeam_data08_cosmag.0091890.AOD.pool.root</datasetName>
-
-		</dataset_info>
-	      </chaindataset_info>
-	      
-	      <chaindataset_info>
-		<jobTransformData/>
-		<chaindatasetName>job2.AOD.pool.root</chaindatasetName>
-		<dataset_info>
-		  <jobTransformData/>
-                  <datasetName>/afs/cern.ch/atlas/offline/external/FullChainTest/tier0-vol3/14.5.X/root_archive/archive/RPCwBeam_data08_cosmag.0091890.AOD.pool.root</datasetName>
-		</dataset_info>
-	      </chaindataset_info>
-  
-	    </jobTransform>
-	    
-	  </chainElement>
-	</sequential>
-       </chain>
-    </jobList>
-          
-    <jobGroups>
-      <jobGroup name="JobTransform" parent="Transform"></jobGroup>     
-    </jobGroups>
-          
-  </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/RTTSummaryReference.xml
deleted file mode 100644
index cc42da2c7c6a4cb996b961a92adeaf42a7a88396..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/RTTSummaryReference.xml
+++ /dev/null
@@ -1,140 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1239389449.69</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0.5</releaseName>
-    <statusText>09/04/10@20:50: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/04/10@20:50</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/chainMergeMissingFiles/Results</resBaseDir>
-    <endTime1>09/04/10@20:50</endTime1>
-    <endTime2>1239389449.67</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-61</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/chainMergeMissingFiles/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1239388986.11</startTime2>
-    <startTime1>09/04/10@20:43</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1229680106</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-61</RTTSrcVersion>
-    <nicosDate1>2008/12/19 09:48:26</nicosDate1>
-    <release>14.5.0.5</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>RecJobTransforms</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Reconstruction/RecJobTransforms</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>RecJobTransforms-00-07-89</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <SequentialMinder>
-      <status>done</status>
-      <nActiveMinders>0</nActiveMinders>
-      <jobID>0</jobID>
-      <state>done</state>
-      <jobGroup>ContainerGroup</jobGroup>
-      <stateHistory>[('queued', '09/04/10 20:44'), ('running', '09/04/10 20:44'), ('done', '09/04/10 20:50')]</stateHistory>
-      <identifiedName>Sequential0</identifiedName>
-      <nContainedMinders>1</nContainedMinders>
-      <ChainJobMinder>
-        <status>done</status>
-        <nActiveMinders>0</nActiveMinders>
-        <jobID>1</jobID>
-        <state>done</state>
-        <jobGroup>ContainerGroup</jobGroup>
-        <stateHistory>[('queued', '09/04/10 20:44'), ('running', '09/04/10 20:44'), ('done', '09/04/10 20:50')]</stateHistory>
-        <identifiedName>Job1</identifiedName>
-        <nContainedMinders>1</nContainedMinders>
-        <minder>
-          <nUserTests>0</nUserTests>
-          <jobDocURL>
-          </jobDocURL>
-          <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/chainMergeMissingFiles/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2</resultsPath>
-          <batchStatus>DONE</batchStatus>
-          <jobDisplayName>
-          </jobDisplayName>
-          <cpuTime2000>446</cpuTime2000>
-          <jobID>2</jobID>
-          <chainLength>1</chainLength>
-          <state>done</state>
-          <postProcessingResult>no tests</postProcessingResult>
-          <jobGroup>JobTransform</jobGroup>
-          <identifiedName>MergeJob2</identifiedName>
-          <status>success</status>
-          <chainSuccesses>1</chainSuccesses>
-          <mem>667MB</mem>
-          <stateHistory>[('queued', '09/04/10 20:44'), ('running', '09/04/10 20:44'), ('postProcessingQueued', '09/04/10 20:49'), ('postProcessingRunning', '09/04/10 20:49'), ('done', '09/04/10 20:50')]</stateHistory>
-          <jobDocString>MergeJob</jobDocString>
-          <hashString>-819398929</hashString>
-          <nRetries>0</nRetries>
-          <cpuTime>166</cpuTime>
-          <jobName>MergeJob</jobName>
-          <vmem>1153MB</vmem>
-          <nUserActions>2</nUserActions>
-          <exitStatus>Unknown</exitStatus>
-          <processingResult>success</processingResult>
-          <wallTime>320</wallTime>
-          <datasets>
-            <datasetinfo>
-              <file>RPCwBeam_data08_cosmag.0091890.AOD.pool.root</file>
-              <directory>/afs/cern.ch/atlas/offline/external/FullChainTest/tier0-vol3/14.5.X/root_archive/archive</directory>
-              <mtime>09/02/16@16:45</mtime>
-              <size>18270897B</size>
-              <origin>fallback</origin>
-            </datasetinfo>
-            <datasetinfo>
-              <file>RPCwBeam_data08_cosmag.0091890.AOD.pool.root</file>
-              <directory>/afs/cern.ch/atlas/offline/external/FullChainTest/tier0-vol3/14.5.X/root_archive/archive</directory>
-              <mtime>09/02/16@16:45</mtime>
-              <size>18270897B</size>
-              <origin>fallback</origin>
-            </datasetinfo>
-          </datasets>
-          <missingDatasets>
-            <missingDataset>job1.AOD.pool.root</missingDataset>
-            <missingDataset>job2.AOD.pool.root</missingDataset>
-          </missingDatasets>
-          <sUserTests>0</sUserTests>
-          <deliveredkeepfiles>
-            <baseKeepFileDir>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2</baseKeepFileDir>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postProcessor.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postProcessorScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/package.xml</keepfile_newstyle>
-            <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/MergeJob2_log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postprocessing.debug.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postprocessing.warning.log</keepfile_newstyle>
-            <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/runScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postprocessing.error.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/postprocessing.info.log</keepfile_newstyle>
-            <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/env.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/RecJobTransforms/JobTransform/MergeJob/2/RTTtests.db</keepfile_newstyle>
-          </deliveredkeepfiles>
-          <errors/>
-          <done/>
-          <checkResults/>
-          <testResults/>
-        </minder>
-      </ChainJobMinder>
-    </SequentialMinder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/TestCase.py
deleted file mode 100644
index 3af32b5d47bbae3c8f6f678d2b947d507bcb1724..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/TestCase.py
+++ /dev/null
@@ -1,20 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self):
-        self.rttTestXMLComp()
-
-    def testKeepfileDiff(self):
-        self.rttKeepfileDiff()
-
-    def testNoErrorReport(self):
-        self.rttTestNoErrorReport()
-
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/__init__.py
deleted file mode 100644
index 9cfc0d958e61f3cdbc43132d3f480cf413bf356b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/__init__.py
+++ /dev/null
@@ -1,6 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Fallback file test for job chains. A chained merge job is run which
-requests non-existent chain store files. It then uses the fallback files as
-input files.
-"""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/cfg/cfg.xml.TPL
deleted file mode 100644
index 83e2ceb9329e17f2a1259412f5114eeb71d6ec2f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,51 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0.5</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Reconstruction/RecJobTransforms</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasProduction/14.5.0.5</cmtpath>
-            <version>RecJobTransforms-00-07-89</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasProduction/14.5.0.5/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasProduction/14.5.0.5/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-         
-
-   </confFileSrc>
-      
-   <packageAccept>RecJobTransforms</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/tests.py b/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/tests.py
deleted file mode 100644
index 3cce74edfb3662611d46ffbaa2ceb4354ea89a25..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainMergeMissingFiles/tests.py
+++ /dev/null
@@ -1,44 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from testsuitetest import Test, SummaryFileTest
-import os.path
-
-def run():
-    for klass in [XMLFileTest, LogTest]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class XMLFileTest(SummaryFileTest):
-    def __init__(self, tester):
-        SummaryFileTest.__init__(self, tester)
-
-    def testHasExactlyOnePackage(self):
-        return self.hasExactlyNPackages(1)
-
-    def testHasPackageCalled(self):
-	return self.hasOnlyOnePackageCalled('AthExHelloWorld')
-
-    def testHasOneJobOnly(self):
-        return self.hasNJobs(1)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class LogTest(Test):
-    def __init__(self, tester):
-        Test.__init__(self, tester)
-
-    def testLogFileExists(self):
-        minder = self.tester.testRuns[0].jobMinders[0]
-        logfile = os.path.join(minder.resPath, minder.log)
-        return os.path.exists(logfile)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/Package/package.xml.TPL
deleted file mode 100644
index c304091f386cac70f188b76ff696c324522bc3e6..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/Package/package.xml.TPL
+++ /dev/null
@@ -1,100 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-  <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-    <mailto>brinick.simmons@cern.ch</mailto>
-    <refRelease>14.5.0</refRelease>
-
-    <jobList>
-
-      <chain>
-	<chainName>TwoParallelThenMergeJob</chainName>
-	<sequential>
-	  <parallel>
-
-	    <chainElement>
-	      <jobTransform>
-		<doc>ParallelJob1</doc>
-		<jobTransformJobName>ParallelJob1</jobTransformJobName>
-		<jobTransformCmd>
-		  Reco_trf.py inputBSFile=/afs/cern.ch/user/g/gencomm/w0/RTT_INPUT_DATA/CosmicATN/daq.ATLAS.0091900.physics.IDCosmic.LB0001.SFO-1._0001.10EVTS.data maxEvents=-1 trigStream=IDCosmic autoConfiguration=FieldAndGeo,BeamType,ConditionsTag preInclude=RecExCommission/RecExCommissionRepro.py,RecExCommission/MinimalCommissioningSetup.py outputESDFile=myESD.pool.root outputAODFile=job1.AOD.pool.root outputMuonCalibNtp=muonCalib.root HIST=myMergedMonitoring.root RAW_IDPROJCOMM=IDPROJCOMM._0001.data DPD_PIXELCOMM=PIXELCOMM.pool.root DPD_SCTCOMM=SCTCOMM.pool.root DPD_IDCOMM=IDCOMM.pool.root DPD_IDPROJCOMM=IDPROJCOMM.pool.root DPD_CALOCOMM=CALOCOMM.pool.root DPD_EMCLUSTCOMM=EMCLUSTCOMM.pool.root DPD_EGAMTAUCOMM=EGamTauComm.pool.root DPD_RPCCOMM=RPCCOMM.pool.root DPD_TGCCOMM=TGCCOMM.pool.root DPD_TILECOMM=TILECOMM.pool.root postInclude=RecJobTransforms/reducedRegionSelectorPostConfig.py --ignoreunknown
-		</jobTransformCmd>
-		<group>JobTransform</group>
-
-		<dataset_info>
-		  <jobTransformData />
-		  <datasetName>/afs/cern.ch/user/g/gencomm/w0/RTT_INPUT_DATA/CosmicATN/daq.ATLAS.0091900.physics.IDCosmic.LB0001.SFO-1._0001.10EVTS.data</datasetName>
-		</dataset_info>
-
-	      </jobTransform>
-
-	      <chainfileout>job1.AOD.pool.root</chainfileout>
-	    </chainElement>
-
-	    <chainElement>
-
-	      <jobTransform>
-		<doc>ParallelJob2</doc>
-		<jobTransformJobName>ParallelJob2</jobTransformJobName>
-		<jobTransformCmd>
-		  Reco_trf.py inputBSFile=/afs/cern.ch/user/g/gencomm/w0/RTT_INPUT_DATA/CosmicATN/daq.ATLAS.0091900.physics.IDCosmic.LB0001.SFO-1._0001.10EVTS.data maxEvents=-1 trigStream=IDCosmic autoConfiguration=FieldAndGeo,BeamType,ConditionsTag preInclude=RecExCommission/RecExCommissionRepro.py,RecExCommission/MinimalCommissioningSetup.py outputESDFile=myESD.pool.root outputAODFile=job2.AOD.pool.root outputMuonCalibNtp=muonCalib.root HIST=myMergedMonitoring.root RAW_IDPROJCOMM=IDPROJCOMM._0001.data DPD_PIXELCOMM=PIXELCOMM.pool.root DPD_SCTCOMM=SCTCOMM.pool.root DPD_IDCOMM=IDCOMM.pool.root DPD_IDPROJCOMM=IDPROJCOMM.pool.root DPD_CALOCOMM=CALOCOMM.pool.root DPD_EMCLUSTCOMM=EMCLUSTCOMM.pool.root DPD_EGAMTAUCOMM=EGamTauComm.pool.root DPD_RPCCOMM=RPCCOMM.pool.root DPD_TGCCOMM=TGCCOMM.pool.root DPD_TILECOMM=TILECOMM.pool.root postInclude=RecJobTransforms/reducedRegionSelectorPostConfig.py --ignoreunknown
-		</jobTransformCmd>
-		<group>JobTransform</group>
-
-		<dataset_info>
-		  <jobTransformData />
-		  <datasetName>/afs/cern.ch/user/g/gencomm/w0/RTT_INPUT_DATA/CosmicATN/daq.ATLAS.0091900.physics.IDCosmic.LB0001.SFO-1._0001.10EVTS.data</datasetName>
-		</dataset_info>
-
-	      </jobTransform>
-
-	      <chainfileout>job2.AOD.pool.root</chainfileout>
-	    </chainElement>
-	    
-	  </parallel>
-
-	  <chainElement>
-
-	    <jobTransform>
-	      <doc>MergeJob</doc>
-	      <jobTransformJobName>MergeJob</jobTransformJobName>
-	      <jobTransformCmd>
-		MergePool_trf.py maxEvents=-1 inputAODFile=job1.AOD.pool.root,job2.AOD.pool.root geometryVersion=ATLAS-GEO-03-00-00 outputAODFile=MergedAOD.pool.root preInclude=RecExCommission/MinimalCommissioningSetup.py preExec='rec.doEgamma.set_Value_and_Lock(False)' --ignoreunknown
-	      </jobTransformCmd>
-
-	      <group>JobTransform</group>
-	      
-	      <chaindataset_info>
-		<jobTransformData/>
-		<chaindatasetName>job1.AOD.pool.root</chaindatasetName>
-		<dataset_info>
-		  <jobTransformData/>
-		  <datasetName>/castor/cern.ch/user/s/seuster/JetRec_RTT/J5/calib0.005014.J5_pythia_jetjet.digit.RDO.v12003101_tid003440._00006.pool.root.2</datasetName>
-		</dataset_info>
-	      </chaindataset_info>
-	      
-	      <chaindataset_info>
-		<jobTransformData/>
-		<chaindatasetName>job2.AOD.pool.root</chaindatasetName>
-		<dataset_info>
-		  <jobTransformData/>
-		  <datasetName>/castor/cern.ch/user/s/seuster/JetRec_RTT/J5/calib0.005014.J5_pythia_jetjet.digit.RDO.v12003101_tid003440._00006.pool.root.2</datasetName>
-		</dataset_info>
-	      </chaindataset_info>
-  
-	    </jobTransform>
-	    
-	  </chainElement>
-	</sequential>
-      </chain>
-
-    </jobList>
-          
-    <jobGroups>
-      <jobGroup name="JobTransform" parent="Transform"></jobGroup>     
-    </jobGroups>
-       
-  </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/RTTSummaryReference.xml
deleted file mode 100644
index d64d424774db3e43825e944d525083ca4f5488af..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/RTTSummaryReference.xml
+++ /dev/null
@@ -1,297 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1245228838.32</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0.5</releaseName>
-    <statusText>09/06/17@10:53: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/17@10:53</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/Results</resBaseDir>
-    <endTime1>09/06/17@10:53</endTime1>
-    <endTime2>1245228838.29</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1245225289.6</startTime2>
-    <startTime1>09/06/17@09:54</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1229680106</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/19 09:48:26</nicosDate1>
-    <release>14.5.0.5</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <SequentialMinder>
-      <status>done</status>
-      <nActiveMinders>0</nActiveMinders>
-      <jobID>0</jobID>
-      <state>done</state>
-      <jobGroup>ContainerGroup</jobGroup>
-      <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('done', '09/06/17 10:53')]</stateHistory>
-      <identifiedName>Sequential0</identifiedName>
-      <nContainedMinders>2</nContainedMinders>
-      <ChainJobMinder>
-        <status>done</status>
-        <nActiveMinders>0</nActiveMinders>
-        <jobID>6</jobID>
-        <state>done</state>
-        <jobGroup>ContainerGroup</jobGroup>
-        <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 10:47'), ('done', '09/06/17 10:53')]</stateHistory>
-        <identifiedName>Job6</identifiedName>
-        <nContainedMinders>1</nContainedMinders>
-        <minder>
-          <nUserTests>0</nUserTests>
-          <jobDocURL>
-          </jobDocURL>
-          <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7</resultsPath>
-          <batchStatus>DONE</batchStatus>
-          <jobDisplayName>
-          </jobDisplayName>
-          <cpuTime2000>164</cpuTime2000>
-          <jobID>7</jobID>
-          <chainLength>1</chainLength>
-          <state>done</state>
-          <postProcessingResult>no tests</postProcessingResult>
-          <jobGroup>JobTransform</jobGroup>
-          <identifiedName>MergeJob7</identifiedName>
-          <status>success</status>
-          <chainSuccesses>1</chainSuccesses>
-          <mem>631MB</mem>
-          <stateHistory>[('queued', '09/06/17 10:47'), ('running', '09/06/17 10:47'), ('postProcessingQueued', '09/06/17 10:50'), ('postProcessingRunning', '09/06/17 10:51'), ('done', '09/06/17 10:53')]</stateHistory>
-          <jobDocString>MergeJob</jobDocString>
-          <hashString>-816743624</hashString>
-          <nRetries>0</nRetries>
-          <cpuTime>61</cpuTime>
-          <jobName>MergeJob</jobName>
-          <vmem>1118MB</vmem>
-          <nUserActions>2</nUserActions>
-          <exitStatus>Unknown</exitStatus>
-          <processingResult>success</processingResult>
-          <wallTime>167</wallTime>
-          <datasets>
-            <datasetinfo>
-              <file>job1.AOD.pool.root</file>
-              <directory>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/ChainStore/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/TwoParallelThenMergeJob</directory>
-              <mtime>09/06/17@10:44</mtime>
-              <size>1297641B</size>
-              <origin>chainstore</origin>
-            </datasetinfo>
-            <datasetinfo>
-              <file>job2.AOD.pool.root</file>
-              <directory>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/ChainStore/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/TwoParallelThenMergeJob</directory>
-              <mtime>09/06/17@10:47</mtime>
-              <size>1297728B</size>
-              <origin>chainstore</origin>
-            </datasetinfo>
-          </datasets>
-          <missingDatasets/>
-          <archivers/>
-          <sUserTests>0</sUserTests>
-          <deliveredkeepfiles>
-            <baseKeepFileDir>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7</baseKeepFileDir>
-            <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/MergedAOD.pool.root.checkFile</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postprocessing.error.log</keepfile_newstyle>
-            <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/package.xml</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postprocessing.warning.log</keepfile_newstyle>
-            <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/env.log</keepfile_newstyle>
-            <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/runScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/MergeJob7_log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/RTTtests.db</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postProcessor.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postprocessing.debug.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postProcessorScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/MergeJob/7/postprocessing.info.log</keepfile_newstyle>
-          </deliveredkeepfiles>
-          <errors/>
-          <done/>
-          <checkResults/>
-          <testResults/>
-        </minder>
-      </ChainJobMinder>
-      <ParallelMinder>
-        <status>done</status>
-        <nActiveMinders>0</nActiveMinders>
-        <jobID>1</jobID>
-        <state>done</state>
-        <jobGroup>ContainerGroup</jobGroup>
-        <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('done', '09/06/17 10:47')]</stateHistory>
-        <identifiedName>Parallel1</identifiedName>
-        <nContainedMinders>2</nContainedMinders>
-        <ChainJobMinder>
-          <status>done</status>
-          <nActiveMinders>0</nActiveMinders>
-          <jobID>4</jobID>
-          <state>done</state>
-          <jobGroup>ContainerGroup</jobGroup>
-          <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('done', '09/06/17 10:47')]</stateHistory>
-          <identifiedName>Job4</identifiedName>
-          <nContainedMinders>1</nContainedMinders>
-          <minder>
-            <nUserTests>0</nUserTests>
-            <jobDocURL>
-            </jobDocURL>
-            <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5</resultsPath>
-            <batchStatus>DONE</batchStatus>
-            <jobDisplayName>
-            </jobDisplayName>
-            <cpuTime2000>1657</cpuTime2000>
-            <jobID>5</jobID>
-            <chainLength>1</chainLength>
-            <state>done</state>
-            <postProcessingResult>no tests</postProcessingResult>
-            <jobGroup>JobTransform</jobGroup>
-            <identifiedName>ParallelJob25</identifiedName>
-            <status>success</status>
-            <chainSuccesses>1</chainSuccesses>
-            <mem>1600MB</mem>
-            <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('postProcessingQueued', '09/06/17 10:16'), ('postProcessingRunning', '09/06/17 10:17'), ('done', '09/06/17 10:46')]</stateHistory>
-            <jobDocString>ParallelJob2</jobDocString>
-            <hashString>-1685739220</hashString>
-            <nRetries>0</nRetries>
-            <cpuTime>617</cpuTime>
-            <jobName>ParallelJob2</jobName>
-            <vmem>2098MB</vmem>
-            <nUserActions>2</nUserActions>
-            <exitStatus>Unknown</exitStatus>
-            <processingResult>success</processingResult>
-            <wallTime>1123</wallTime>
-            <datasets/>
-            <missingDatasets/>
-            <archivers/>
-            <sUserTests>0</sUserTests>
-            <deliveredkeepfiles>
-              <baseKeepFileDir>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5</baseKeepFileDir>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postProcessorScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/ParallelJob25_log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/RTTtests.db</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postProcessor.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/RPCCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/TGCCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postprocessing.info.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/myESD.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/IDCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/CALOCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/package.xml</keepfile_newstyle>
-              <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/env.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postprocessing.debug.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/EMCLUSTCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/runScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postprocessing.warning.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/IDPROJCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/SCTCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/TILECOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/EGamTauComm.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/PIXELCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/postprocessing.error.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob2/5/job2.AOD.pool.root.checkFile</keepfile_newstyle>
-            </deliveredkeepfiles>
-            <errors/>
-            <done/>
-            <checkResults/>
-            <testResults/>
-          </minder>
-        </ChainJobMinder>
-        <ChainJobMinder>
-          <status>done</status>
-          <nActiveMinders>0</nActiveMinders>
-          <jobID>2</jobID>
-          <state>done</state>
-          <jobGroup>ContainerGroup</jobGroup>
-          <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('done', '09/06/17 10:44')]</stateHistory>
-          <identifiedName>Job2</identifiedName>
-          <nContainedMinders>1</nContainedMinders>
-          <minder>
-            <nUserTests>0</nUserTests>
-            <jobDocURL>
-            </jobDocURL>
-            <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/chainjobtransforms/Results/14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3</resultsPath>
-            <batchStatus>DONE</batchStatus>
-            <jobDisplayName>
-            </jobDisplayName>
-            <cpuTime2000>1665</cpuTime2000>
-            <jobID>3</jobID>
-            <chainLength>1</chainLength>
-            <state>done</state>
-            <postProcessingResult>no tests</postProcessingResult>
-            <jobGroup>JobTransform</jobGroup>
-            <identifiedName>ParallelJob13</identifiedName>
-            <status>success</status>
-            <chainSuccesses>1</chainSuccesses>
-            <mem>1MB</mem>
-            <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('postProcessingQueued', '09/06/17 10:19'), ('postProcessingRunning', '09/06/17 10:19'), ('done', '09/06/17 10:44')]</stateHistory>
-            <jobDocString>ParallelJob1</jobDocString>
-            <hashString>1233200221</hashString>
-            <nRetries>0</nRetries>
-            <cpuTime>620</cpuTime>
-            <jobName>ParallelJob1</jobName>
-            <vmem>104MB</vmem>
-            <nUserActions>2</nUserActions>
-            <exitStatus>Unknown</exitStatus>
-            <processingResult>success</processingResult>
-            <wallTime>1282</wallTime>
-            <datasets/>
-            <missingDatasets/>
-            <archivers/>
-            <sUserTests>0</sUserTests>
-            <deliveredkeepfiles>
-              <baseKeepFileDir>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3</baseKeepFileDir>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postprocessing.warning.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postprocessing.info.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/IDCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/RTTtests.db</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/IDPROJCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postprocessing.error.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/TGCCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/SCTCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/env.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/TILECOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/CALOCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/myESD.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/runScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postProcessorScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/EGamTauComm.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/RPCCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/PIXELCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/job1.AOD.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postprocessing.debug.log</keepfile_newstyle>
-              <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/EMCLUSTCOMM.pool.root.checkFile</keepfile_newstyle>
-              <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/package.xml</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/postProcessor.log</keepfile_newstyle>
-              <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0.5/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ParallelJob1/3/ParallelJob13_log</keepfile_newstyle>
-            </deliveredkeepfiles>
-            <errors/>
-            <done/>
-            <checkResults/>
-            <testResults/>
-          </minder>
-        </ChainJobMinder>
-      </ParallelMinder>
-    </SequentialMinder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/TestCase.py
deleted file mode 100644
index 3e6d817e1dff147dd8c8e3f091530a0e496ddd44..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/TestCase.py
+++ /dev/null
@@ -1,31 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self):            self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-    
-    def variableContentTags(self):
-        """
-        Add the hashstring tag to the variable content
-        tag list. This job takes files from the
-        chain store, so the content of the file tag depends on
-        the (variable) name of the chain store.
-        """
-
-        tags = RTTTestCase.variableContentTags(self)
-        tags.extend(['directory', 'mtime'])
-        # the size of the output file changes by a few bytes from
-        # run to run. Why?
-        tags.append('size')
-        return tags
-
-
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/__init__.py
deleted file mode 100644
index f65ec7d0e1a7fe97a35ab073a2432e58f124fd5a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/__init__.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs a single chain comprising a sequence of 2 steps. First step is 2 parallel transform jobs, second step: single transform merge job."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/cfg/cfg.xml.TPL
deleted file mode 100644
index 7f4a3a8ac66867ade15c055e712bd56b25dfa00a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/chainjobtransforms/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,51 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0.5</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Control/AthenaExamples/AthExHelloWorld</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-            <version>AthExHelloWorld-01-01-01</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-
-   </confFileSrc>
-      
-   <packageAccept>AthExHelloWorld</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <deleteNonKeepFiles/>
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/Package/package.xml.TPL
deleted file mode 100644
index 3a845265e137a881debb7ed979283701f0f02f19..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/Package/package.xml.TPL
+++ /dev/null
@@ -1,28 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <jobTransform>
-             <doc>comprehensive job - test all RTT features</doc>
-	     <jobTransformJobName>ComprehensiveJob1</jobTransformJobName>
-             <jobTransformCmd>echo "1 2 3 4 5 6 7 8 9 10" &gt;&amp; comprehensiveJob.data;echo "ErrorCode=0 (OK)";</jobTransformCmd>
-             <group>JobTransform</group>
-             <keepFilePattern>comprehensiveJob.data</keepFilePattern>
-             <keepFilePattern>nonexistant.data</keepFilePattern>
-          </jobTransform>
-
-       </jobList>
-          
-       <jobGroups>
-          <jobGroup name="JobTransform" parent="Transform"/>
-       </jobGroups>
-          
-    </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/RTTSummaryReference.xml
deleted file mode 100644
index b8da186e1cb0a44f5f75201cf10c1c92dff70e8a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/RTTSummaryReference.xml
+++ /dev/null
@@ -1,104 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1244808647.5</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/06/12@14:10: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/12@14:10</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/comprehensivejob/Results</resBaseDir>
-    <endTime1>09/06/12@14:10</endTime1>
-    <endTime2>1244808647.49</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/comprehensivejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1244807889.53</startTime2>
-    <startTime1>09/06/12@13:58</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>1</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <nUserTests>0</nUserTests>
-      <jobDocURL>
-      </jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/comprehensivejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0</resultsPath>
-      <batchStatus>DONE</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>124</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>1</chainLength>
-      <state>done</state>
-      <postProcessingResult>no tests</postProcessingResult>
-      <jobGroup>JobTransform</jobGroup>
-      <identifiedName>ComprehensiveJob10</identifiedName>
-      <status>success</status>
-      <chainSuccesses>1</chainSuccesses>
-      <mem>1MB</mem>
-      <stateHistory>[('queued', '09/06/12 14:03'), ('running', '09/06/12 14:03'), ('postProcessingQueued', '09/06/12 14:09'), ('postProcessingRunning', '09/06/12 14:10'), ('done', '09/06/12 14:10')]</stateHistory>
-      <jobDocString>comprehensive job - test all RTT features</jobDocString>
-      <hashString>503309858</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>46</cpuTime>
-      <jobName>ComprehensiveJob1</jobName>
-      <vmem>104MB</vmem>
-      <nUserActions>2</nUserActions>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>success</processingResult>
-      <wallTime>371</wallTime>
-      <datasets/>
-      <missingDatasets/>
-      <archivers/>
-      <sUserTests>0</sUserTests>
-      <deliveredkeepfiles>
-        <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0</baseKeepFileDir>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postprocessing.warning.log</keepfile_newstyle>
-        <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/ComprehensiveJob10_log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postprocessing.info.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postprocessing.debug.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postProcessorScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/package.xml</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postProcessor.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/comprehensiveJob.data</keepfile_newstyle>
-        <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/runScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/postprocessing.error.log</keepfile_newstyle>
-        <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/env.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/ComprehensiveJob1/0/RTTtests.db</keepfile_newstyle>
-      </deliveredkeepfiles>
-      <errors/>
-      <done/>
-      <checkResults/>
-      <testResults/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/TestCase.py
deleted file mode 100644
index 1e1a9c26f92c5bc0b435f4633b37b60862e09a5c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/TestCase.py
+++ /dev/null
@@ -1,18 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-class TestCase(RTTTestCase):
-
-    def testCMLComp(self):
-        self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def testNoErrorReport(self):
-        self.rttTestNoErrorReport()
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/__init__.py
deleted file mode 100644
index 19356790a4a8fe019146288d89d01f21d7df3520..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/__init__.py
+++ /dev/null
@@ -1,7 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-'''Short testsuite description goes here'''
-
-def run():
-   "Any initialisation code goes in this method. Is run before RTT is launched."
-   pass
diff --git a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/cfg/cfg.xml.TPL
deleted file mode 100644
index 5ac6d936f30fe1f401abaa0259568b8315225f8c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/comprehensivejob/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Control/AthenaExamples/AthExHelloWorld</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-            <version>AthExHelloWorld-01-01-01</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-         
-   </confFileSrc>
-      
-   <packageAccept>AthExHelloWorld</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld/Package/package.xml.TPL
deleted file mode 100644
index c50ef90f4918d7b1172eaf42a9f0c075de84b3b9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/Package/package.xml.TPL
+++ /dev/null
@@ -1,31 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <athena>
-              <doc>blah</doc>
-              <doclink>http://cern.ch</doclink>
-              <options>AthExHelloWorld/HelloWorldOptions.py</options>
-              <package>Control/AthenaExamples/AthExHelloWorld</package>
-              <group>HelloWorld</group>
-              <queue>short</queue>
-          </athena>
-
-      </jobList>
-
-      <jobGroups>
-
-         <jobGroup name="AthenaHelloWorld" parent="Athena">
-         </jobGroup>
-
-    </jobGroups>
-
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/helloworld/RTTSummaryReference.xml
deleted file mode 100644
index e3a523de68701b6fa33137c8148928a547f67eb0..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/RTTSummaryReference.xml
+++ /dev/null
@@ -1,107 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1245225611.93</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/06/17@10:00: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/17@10:00</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/helloworld/Results</resBaseDir>
-    <endTime1>09/06/17@10:00</endTime1>
-    <endTime2>1245225611.91</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/helloworld/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1245225289.61</startTime2>
-    <startTime1>09/06/17@09:54</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>1</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <nUserTests>0</nUserTests>
-      <jobDocURL>http://cern.ch</jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/helloworld/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0</resultsPath>
-      <batchStatus>DONE</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>75</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>1</chainLength>
-      <state>done</state>
-      <postProcessingResult>no tests</postProcessingResult>
-      <jobGroup>AthenaHelloWorld</jobGroup>
-      <identifiedName>HelloWorldOptions0</identifiedName>
-      <status>success</status>
-      <chainSuccesses>1</chainSuccesses>
-      <mem>18MB</mem>
-      <stateHistory>[('queued', '09/06/17 09:57'), ('running', '09/06/17 09:57'), ('postProcessingQueued', '09/06/17 09:58'), ('postProcessingRunning', '09/06/17 09:59'), ('done', '09/06/17 09:59')]</stateHistory>
-      <jobDocString>blah</jobDocString>
-      <hashString>1079629659</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>28</cpuTime>
-      <jobName>HelloWorldOptions</jobName>
-      <vmem>297MB</vmem>
-      <nUserActions>3</nUserActions>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>success</processingResult>
-      <wallTime>Unobtainable</wallTime>
-      <datasets/>
-      <missingDatasets/>
-      <archivers/>
-      <sUserTests>0</sUserTests>
-      <deliveredkeepfiles>
-        <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0</baseKeepFileDir>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/RTTtests.db</keepfile_newstyle>
-        <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/runScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.warning.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postProcessorScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.info.log</keepfile_newstyle>
-        <keepfile_newstyle info='ls -altF command on run directory before and after athena command' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/lsaltF.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.debug.log</keepfile_newstyle>
-        <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/env.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postProcessor.log</keepfile_newstyle>
-        <keepfile_newstyle info='Job options called with the athena command' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/AthExHelloWorld/HelloWorldOptions.py</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.error.log</keepfile_newstyle>
-        <keepfile_newstyle info='cmt show uses command on package cmt dir' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/cmtShowUses.log</keepfile_newstyle>
-        <keepfile_newstyle info='Marker emitted at the end of the job' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/JobCompleteMarker.txt</keepfile_newstyle>
-        <keepfile_newstyle info='WARNING,ERROR and FATAL messages from job log' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/ERROR.log</keepfile_newstyle>
-        <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/package.xml</keepfile_newstyle>
-        <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/HelloWorldOptions0_log</keepfile_newstyle>
-      </deliveredkeepfiles>
-      <errors/>
-      <done/>
-      <checkResults/>
-      <testResults/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld/TestCase.py
deleted file mode 100644
index a43695cad26cbff172401d290382878c5d43d5d4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/TestCase.py
+++ /dev/null
@@ -1,19 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self):
-        self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def testNoErrorReport(self):
-        self.rttTestNoErrorReport()
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld/__init__.py
deleted file mode 100644
index 0e575bfe8300f2de51fe5a1ec7d00fcfc35d9126..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/__init__.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs single HelloWorld job on LSF batch using rttProvides."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld/cfg/cfg.xml.TPL
deleted file mode 100644
index c88d20e064ac3df0f555ff3d05b8eb78917e5c29..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,55 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-
-   <topProject>AtlasProduction</topProject>
-
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-        <rttProvides>
-
-            <packageConfFile>
-               <package>Control/AthenaExamples/AthExHelloWorld</package>
-               <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-               <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-               <version>AthExHelloWorld-01-01-01</version>
-               <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-               <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-            </packageConfFile> 
-
-        </rttProvides>
-
-   </confFileSrc>
-
-   <packageAccept>AthExHelloWorld</packageAccept>
-
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-
-   <releaseType>project</releaseType>
-
-   <disableMailTo />
-   <devRun />
-   <deleteNonKeepFiles/>
-
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld/tests.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld/tests.py
deleted file mode 100644
index 3cce74edfb3662611d46ffbaa2ceb4354ea89a25..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld/tests.py
+++ /dev/null
@@ -1,44 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from testsuitetest import Test, SummaryFileTest
-import os.path
-
-def run():
-    for klass in [XMLFileTest, LogTest]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class XMLFileTest(SummaryFileTest):
-    def __init__(self, tester):
-        SummaryFileTest.__init__(self, tester)
-
-    def testHasExactlyOnePackage(self):
-        return self.hasExactlyNPackages(1)
-
-    def testHasPackageCalled(self):
-	return self.hasOnlyOnePackageCalled('AthExHelloWorld')
-
-    def testHasOneJobOnly(self):
-        return self.hasNJobs(1)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class LogTest(Test):
-    def __init__(self, tester):
-        Test.__init__(self, tester)
-
-    def testLogFileExists(self):
-        minder = self.tester.testRuns[0].jobMinders[0]
-        logfile = os.path.join(minder.resPath, minder.log)
-        return os.path.exists(logfile)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/Package/package.xml.TPL
deleted file mode 100644
index c500e5170bf103f1413cca1016c4c92585e1181f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/Package/package.xml.TPL
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <athena>
-              <doc>blah</doc>
-              <doclink>http://cern.ch</doclink>
-              <options>AthExHelloWorld/WrongOnPurpose.py</options>
-              <package>Control/AthenaExamples/AthExHelloWorld</package>
-              <group>HelloWorld</group>
-              <queue>short</queue>
-	      <dataset_info>
-		<dc2/>
-		<datasetName>NonExistentDataSet</datasetName>
-		<datasetRepeat>2</datasetRepeat>	
-	      </dataset_info>	
-          </athena>
-
-      </jobList>
-
-      <jobGroups>
-
-         <jobGroup name="AthenaHelloWorld" parent="Athena">
-         </jobGroup>
-
-    </jobGroups>
-
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/RTTSummaryReference.xml
deleted file mode 100644
index ee34d8affc03c5012446cbbd6cf7f2d3d54fa76f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/RTTSummaryReference.xml
+++ /dev/null
@@ -1,76 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1239389064.69</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/04/10@20:44: jobs succ/done/tot: 0/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/04/10@20:44</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badDataSet/Results</resBaseDir>
-    <endTime1>09/04/10@20:44</endTime1>
-    <endTime2>1239389064.68</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-61</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badDataSet/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1239388986.4</startTime2>
-    <startTime1>09/04/10@20:43</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>0</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-61</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>1</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <identifiedName>WrongOnPurpose0</identifiedName>
-      <jobName>WrongOnPurpose</jobName>
-      <jobID>0</jobID>
-      <state>done</state>
-      <status>error</status>
-      <jobGroup>ErrorGroup</jobGroup>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badDataSet/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/WrongOnPurpose/0</resultsPath>
-      <stateHistory>[('queued', '09/04/10 20:44'), ('done', '09/04/10 20:44')]</stateHistory>
-      <datasets/>
-      <sUserTests/>
-      <deliveredkeepfiles/>
-      <errors>Error building descriptor in method handleAthenaDataSets
-Traceback:RTTSException.RTTInputErrorlogical data set name not known to  catalog NonExistentDataSet  File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobsXMLReader.py", line 93, in runXPathFn
-    fn(Evaluate(xpath, self.job), self.dict)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobsXMLReader.py", line 527, in handleAthenaDataSets
-    [self.handleAthenaDataSet(datasetNode, datasetsDict) for datasetNode in datasets]
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobsXMLReader.py", line 541, in handleAthenaDataSet
-    physDataSetName = self.getPhysicalDataset(datasetName)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobsXMLReader.py", line 422, in getPhysicalDataset
-    raise RTTInputError(msg)
-
-</errors>
-      <done/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/TestCase.py
deleted file mode 100644
index f420b1dc181506224dcc45a9c718b456f8b142c4..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/TestCase.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def variableContentTags(self):
-        vcTags = RTTTestCase.variableContentTags(self)
-        vcTags.append('errors')
-        return vcTags
-    
-    def test1ErrorTag(self): self.rttTestNTags(self.dom, 'errors', 1)
-
-if __name__ == '__main__':
-
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/__init__.py
deleted file mode 100644
index 80171fef9fe4b0efb966e29421ca4cf85f9d1a2d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/__init__.py
+++ /dev/null
@@ -1,6 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs single HelloWorld job on LSF batch using rttProvides.
-   Mentions a non-existant dataset in the package unidied configuration file. 
-   This should provoke the generation of an error message in the job descripor
-   which in turn should trigger the construction of an ErrorMinder"""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/cfg/cfg.xml.TPL
deleted file mode 100644
index 7f10de7cb2a5a2ef78359af219572f735a75220b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badDataSet/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,54 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-
-   <topProject>AtlasProduction</topProject>
-
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-        <rttProvides>
-
-            <packageConfFile>
-               <package>Control/AthenaExamples/AthExHelloWorld</package>
-               <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-               <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-               <version>AthExHelloWorld-01-01-01</version>
-               <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-               <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-            </packageConfFile> 
-
-        </rttProvides>
-
-   </confFileSrc>
-
-   <packageAccept>AthExHelloWorld</packageAccept>
-
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-
-   <releaseType>project</releaseType>
-
-    <disableMailTo />
-    <devRun />
-
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/Package/package.xml.TPL
deleted file mode 100644
index 287178c0fe590f1b3454ef21e34665c3cba0993d..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/Package/package.xml.TPL
+++ /dev/null
@@ -1,31 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <athena>
-              <doc>blah</doc>
-              <doclink>http://cern.ch</doclink>
-              <options>AthExHelloWorld/WrongOnPurpose.py</options>
-              <package>Control/AthenaExamples/AthExHelloWorld</package>
-              <group>HelloWorld</group>
-              <queue>short</queue>
-          </athena>
-
-      </jobList>
-
-      <jobGroups>
-
-         <jobGroup name="AthenaHelloWorld" parent="Athena">
-         </jobGroup>
-
-    </jobGroups>
-
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/RTTSummaryReference.xml
deleted file mode 100644
index fe2c5b9e1a3528210614eca6078d52840b27fd09..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/RTTSummaryReference.xml
+++ /dev/null
@@ -1,105 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1239389067.35</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/04/10@20:44: jobs succ/done/tot: 0/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/04/10@20:44</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badJobOptions/Results</resBaseDir>
-    <endTime1>09/04/10@20:44</endTime1>
-    <endTime2>1239389067.34</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-61</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badJobOptions/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1239388986.19</startTime2>
-    <startTime1>09/04/10@20:43</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>0</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-61</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>1</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <identifiedName>WrongOnPurpose0</identifiedName>
-      <jobName>WrongOnPurpose</jobName>
-      <jobID>0</jobID>
-      <state>done</state>
-      <status>error</status>
-      <jobGroup>ErrorGroup</jobGroup>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badJobOptions/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/WrongOnPurpose/0</resultsPath>
-      <stateHistory>[('queued', '09/04/10 20:44'), ('done', '09/04/10 20:44')]</stateHistory>
-      <datasets/>
-      <sUserTests/>
-      <deliveredkeepfiles/>
-      <errors>Error constructing minder from a Descriptor of type AthenaJobDescriptor, replacing with ErrorMinder:
-Exception:
-Error  setting up run directory by WrongOnPurpose0: TraceBack:
-exceptions.IOError[Errno 2] No such file or directory: u'/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badJobOptions/Work/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/WrongOnPurpose/0/AthExHelloWorld/WrongOnPurpose.py'  File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Minder.py", line 857, in setupJobGroupDirs
-    jobGroupDirectoryMaker.setupRunDir(self)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobGroupDirectoryMaker.py", line 181, in setupRunDir
-    self.descSetUpRunDirFn()
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/AthenaJobDescriptor.py", line 311, in setupRunDir
-    [addDataSets(self.runPath, self.datasets, jo) for jo in self.jobOptions]
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/AthenaJobDescriptor.py", line 298, in addDataSets
-    file = open(os.path.join(dir, jo))
-
-Traceback
-RTTSException.RTTCodingErrorError  setting up run directory by WrongOnPurpose0: TraceBack:
-exceptions.IOError[Errno 2] No such file or directory: u'/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badJobOptions/Work/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/WrongOnPurpose/0/AthExHelloWorld/WrongOnPurpose.py'  File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Minder.py", line 857, in setupJobGroupDirs
-    jobGroupDirectoryMaker.setupRunDir(self)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/JobGroupDirectoryMaker.py", line 181, in setupRunDir
-    self.descSetUpRunDirFn()
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/AthenaJobDescriptor.py", line 311, in setupRunDir
-    [addDataSets(self.runPath, self.datasets, jo) for jo in self.jobOptions]
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/AthenaJobDescriptor.py", line 298, in addDataSets
-    file = open(os.path.join(dir, jo))
-  File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Factory_Minder.py", line 131, in create
-    minder = self.createMinder(argBag, jDescriptor)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Factory_Minder.py", line 117, in createMinder
-    minder = minderClass(argBag, jDescriptor)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/LSFBatchJobMinder.py", line 33, in __init__
-    BatchJobMinder.__init__(self, argBag, jDescriptor)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/BatchJobMinder.py", line 20, in __init__
-    WorkerJobMinder.__init__(self, argBag, jDescriptor)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/WorkerJobMinder.py", line 40, in __init__
-    Minder.__init__(self, argBag, jDescriptor)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Minder.py", line 193, in __init__
-    self.setupJobGroupDirs(jobGroupDirectoryMaker)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/MethodTimer.py", line 9, in timed_f
-    result = f(self, *a, **k)
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/Minder.py", line 864, in setupJobGroupDirs
-    raise RTTCodingError(msg)
-
-</errors>
-      <done/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/TestCase.py
deleted file mode 100644
index cb426eb4f586a757a481090b5e7d4012e506e661..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/TestCase.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-
-    def testXMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def variableContentTags(self):
-        vcTags = RTTTestCase.variableContentTags(self)
-        vcTags.append('errors')
-        return vcTags
-
-    def test1ErrorTag(self): self.rttTestNTags(self.dom, 'errors', 1)
-
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/__init__.py
deleted file mode 100644
index 14e4595ec3646beebf5c1ca22c9f62a381e28e67..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/__init__.py
+++ /dev/null
@@ -1,4 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs single HelloWorld job on LSF batch using rttProvides.
-Uses bad job options file to check error recovery"""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/cfg/cfg.xml.TPL
deleted file mode 100644
index 7f10de7cb2a5a2ef78359af219572f735a75220b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badJobOptions/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,54 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-
-   <topProject>AtlasProduction</topProject>
-
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-        <rttProvides>
-
-            <packageConfFile>
-               <package>Control/AthenaExamples/AthExHelloWorld</package>
-               <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-               <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-               <version>AthExHelloWorld-01-01-01</version>
-               <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-               <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-            </packageConfFile> 
-
-        </rttProvides>
-
-   </confFileSrc>
-
-   <packageAccept>AthExHelloWorld</packageAccept>
-
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-
-   <releaseType>project</releaseType>
-
-    <disableMailTo />
-    <devRun />
-
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/Package/package.xml.TPL
deleted file mode 100644
index 838effe11253f5f8e4e02c3f0dbe702ab1fa9a82..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/Package/package.xml.TPL
+++ /dev/null
@@ -1,31 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <athena>
-              <doc>blah</doc>
-              <doclink>http://cern.ch</doclink>
-              <options>AthExHelloWorld/HelloWorldOptions.py</options>
-              <package>Control/AthenaExamples/AthExHelloWorld</package>
-              <group>HelloWorld</group>
-              <queue>short</queue>
-          </athena>
-
-      <jobList> <!-- intentionally error: should be </jobList> -->
-
-      <jobGroups>
-
-         <jobGroup name="AthenaHelloWorld" parent="Athena">
-         </jobGroup>
-
-    </jobGroups>
-
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/RTTSummaryReference.xml
deleted file mode 100644
index 9f4b6b84d61227ca58b91a86c5f0d55761f66d80..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/RTTSummaryReference.xml
+++ /dev/null
@@ -1,59 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1239393029.21</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/04/10@21:50: jobs succ/done/tot: 0/0/0</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/04/10@21:50</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badPackageFile/Results</resBaseDir>
-    <endTime1>09/04/10@21:50</endTime1>
-    <endTime2>1239393029.2</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-61</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/helloworld_badPackageFile/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1239393008.57</startTime2>
-    <startTime1>09/04/10@21:50</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-61</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>0</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>0</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>Initial</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>0</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>User xml file does not parse:
-Exception raised while parsing user input file:
-mismatched tag: line 30, column 5
-Traceback:
-RTTSException.RTTInputErrorException raised while parsing user input file:
-mismatched tag: line 30, column 5  File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/PackageXMLFile.py", line 30, in validate_
-    validator = Validator(self.path) # throws exception if parse fails
-   File "/afs/cern.ch/atlas/project/RTT/data/peter/RTTSrc_LI/Tools/RunTimeTester/src/validateXMLFile.py", line 32, in __init__
-    raise RTTInputError(m)</packageStatus>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/TestCase.py
deleted file mode 100644
index 7a8c27f91f59630791925bbd859f3360f2285e71..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/TestCase.py
+++ /dev/null
@@ -1,23 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def testCheckTagContent(self): self.rttTestCheckTagContent('packageStatus', 'User xml file does not parse:')
-
-    def variableContentTags(self):
-        vcTags = RTTTestCase.variableContentTags(self)
-        vcTags.append('packageStatus')
-        return vcTags
-    
-if __name__ == '__main__':
-
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/__init__.py
deleted file mode 100644
index 0005bc5af9ccb1fe50f6a355fc72e819de74a037..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/__init__.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs single HelloWorld job on LSF batch using rttProvides. Package file is not a valid xml file. Test package error recovery."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/cfg/cfg.xml.TPL
deleted file mode 100644
index 7f10de7cb2a5a2ef78359af219572f735a75220b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/helloworld_badPackageFile/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,54 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType>
-      <releaseRun/>
-   </runType>
-
-   <topProject>AtlasProduction</topProject>
-
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-        <rttProvides>
-
-            <packageConfFile>
-               <package>Control/AthenaExamples/AthExHelloWorld</package>
-               <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-               <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-               <version>AthExHelloWorld-01-01-01</version>
-               <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-               <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-            </packageConfFile> 
-
-        </rttProvides>
-
-   </confFileSrc>
-
-   <packageAccept>AthExHelloWorld</packageAccept>
-
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-
-   <releaseType>project</releaseType>
-
-    <disableMailTo />
-    <devRun />
-
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/Package/package.xml.TPL
deleted file mode 100644
index 657c4b0bdf3cb4a09ee747981a473b7def30f293..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/Package/package.xml.TPL
+++ /dev/null
@@ -1,51 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <jobTransform>
-             <doc>7218.singlepart_mu20</doc>
-             <jobTransformJobName>7218.singlepart_mu20</jobTransformJobName>
-             <jobTransformCmd>
-                   csc_evgen_trf.py -t runNumber=7218 firstEvent=1 maxEvents=1000 randomSeed=54298752 jobConfig=CSC.007218.singlepart_mu20.py outputEvgenFile=CSC.007218.singlepart_mu20.evgen.pool.root histogramFile=NONE ntupleFile=NONE inputGeneratorFile=NONE
-              </jobTransformCmd>
-              <group>JobTransform</group>
-          </jobTransform>
-
-       </jobList>
-          
-       <jobGroups>
-          
-          <jobGroup name="JobTransform" parent="Transform">
-              <keepFilePattern>*.root</keepFilePattern>
-              <keepFilePattern>*.txt</keepFilePattern>
-
-              <test>
-                  <modulename>RttLibraryTools</modulename>
-                  <testname>FileGrepper</testname>
-                  <arg>
-                     <argname>inputFile</argname>
-                     <argvalue>*_log</argvalue>
-                  </arg>
-                  <arg>
-                     <argname>outputFile</argname>
-                     <argvalue>RTT_search_results.txt</argvalue>
-                  </arg>
-                  <arg>
-                     <argname>searchList</argname>
-                     <argvalue>ErrorCode=0</argvalue>
-                  </arg>
-              </test>
-
-          </jobGroup>
-          
-       </jobGroups>
-          
-    </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/RTTSummaryReference.xml
deleted file mode 100644
index f607b486226da9b9df8960188cad7e92f47785a8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/RTTSummaryReference.xml
+++ /dev/null
@@ -1,114 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1245091275.93</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/06/15@20:41: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/15@20:41</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransform/Results</resBaseDir>
-    <endTime1>09/06/15@20:41</endTime1>
-    <endTime2>1245091275.91</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransform/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1245090745.69</startTime2>
-    <startTime1>09/06/15@20:32</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>1</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>1</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>1</nJobsSuccessInPackage>
-    <packageName>EvgenJobTransforms</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Generators/EvgenJobTransforms</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>EvgenJobTransforms-00-06-26</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>1</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <nUserTests>1</nUserTests>
-      <jobDocURL>
-      </jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransform/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0</resultsPath>
-      <batchStatus>DONE</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>118</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>1</chainLength>
-      <state>done</state>
-      <postProcessingResult>success</postProcessingResult>
-      <jobGroup>JobTransform</jobGroup>
-      <identifiedName>7218.singlepart_mu200</identifiedName>
-      <status>success</status>
-      <chainSuccesses>1</chainSuccesses>
-      <mem>1MB</mem>
-      <stateHistory>[('queued', '09/06/15 20:33'), ('running', '09/06/15 20:33'), ('postProcessingQueued', '09/06/15 20:40'), ('postProcessingRunning', '09/06/15 20:40'), ('done', '09/06/15 20:41')]</stateHistory>
-      <jobDocString>7218.singlepart_mu20</jobDocString>
-      <hashString>1518201472</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>44</cpuTime>
-      <jobName>7218.singlepart_mu20</jobName>
-      <vmem>104MB</vmem>
-      <nUserActions>2</nUserActions>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>success</processingResult>
-      <wallTime>363</wallTime>
-      <datasets/>
-      <missingDatasets/>
-      <archivers/>
-      <sUserTests>1</sUserTests>
-      <deliveredkeepfiles>
-        <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0</baseKeepFileDir>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postProcessorScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postprocessing.error.log</keepfile_newstyle>
-        <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/CSC.007218.singlepart_mu20.evgen.pool.root.checkFile</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/CSC.007218.singlepart_mu20.evgen.pool.root</keepfile_newstyle>
-        <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/env.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/jobReport_csc_evgen.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/RTTtests.db</keepfile_newstyle>
-        <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/runScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/jobReport.txt</keepfile_newstyle>
-        <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/7218.singlepart_mu200_log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/listDir.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/RTT_search_results.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postprocessing.warning.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postprocessing.info.log</keepfile_newstyle>
-        <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/package.xml</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postprocessing.debug.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/7218.singlepart_mu20/0/postProcessor.log</keepfile_newstyle>
-      </deliveredkeepfiles>
-      <errors/>
-      <done/>
-      <checkResults/>
-      <testResults>
-        <testResult>
-          <testName>FileGrepper_0</testName>
-          <testStatus>0</testStatus>
-        </testResult>
-      </testResults>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/TestCase.py
deleted file mode 100644
index b7c818e781733858adf3aafd31dfa8d672b7ae19..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/TestCase.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testCMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-    def testNKeepFiles(self): self.rttTestNTags(self.dom, 'keepfile_newstyle', 17)
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/__init__.py
deleted file mode 100644
index 3472e77997dec8ff56710383a16fd214ef19b02b..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/__init__.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs a single EvgenJobTransforms job."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/cfg/cfg.xml.TPL
deleted file mode 100644
index 82f3523e7bed6662027d0e25f5f20ceb5373c56f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Generators/EvgenJobTransforms</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0</cmtpath>
-            <version>EvgenJobTransforms-00-06-26</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-
-   </confFileSrc>
-      
-   <packageAccept>EvgenJobTransforms</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/tests.py b/Tools/RunTimeTester/testsuite/testsuites/jobtransform/tests.py
deleted file mode 100644
index 9683510d3c8e3dd8d18065519ff1adb1ad4706a9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransform/tests.py
+++ /dev/null
@@ -1,44 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from testsuitetest import Test, SummaryFileTest
-import os.path
-
-def run():
-    for klass in [XMLFileTest, LogTest]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class XMLFileTest(SummaryFileTest):
-    def __init__(self, tester):
-        SummaryFileTest.__init__(self, tester)
-
-    def testHasExactlyOnePackage(self):
-        return self.hasExactlyNPackages(1)
-
-    def testHasPackageCalled(self):
-	return self.hasOnlyOnePackageCalled('EvgenJobTransforms')
-
-    def testHasOneJobOnly(self):
-        return self.hasNJobs(1)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class LogTest(Test):
-    def __init__(self, tester):
-        Test.__init__(self, tester)
-
-    def testLogFileExists(self):
-        minder = self.tester.testRuns[0].jobMinders[0]
-        logfile = os.path.join(minder.resPath, minder.log)
-        return os.path.exists(logfile)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/Package/package.xml.TPL
deleted file mode 100644
index d7f73e4f404fe246e389f6f0ac6776bf5d5d050a..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/Package/package.xml.TPL
+++ /dev/null
@@ -1,55 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-  <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-    <mailto>brinick.simmons@cern.ch</mailto>
-    <refRelease>14.5.0</refRelease>
-
-    <jobList>
-
-      <jobTransform>
-	<doc>5224.AlpgenJimmyWenuNp3</doc>
-	<jobTransformJobName>5224.AlpgenJimmyWenuNp3</jobTransformJobName>
-	<jobTransformCmd>
-	  csc_evgen_trf.py -t runNumber=5224 firstEvent=1 maxEvents=50 randomSeed=54298752 jobConfig=CSC.005224.AlpgenJimmyWenuNp3.py outputEvgenFile=CSC.005224.AlpgenJimmyWenuNp3.evgen.pool.root histogramFile=NONE ntupleFile=NONE inputGeneratorFile=/afs/cern.ch/atlas/offline/ProdData/13.0.X/alpgen.005224.WenuNp3._00023.tar.gz
-	</jobTransformCmd>
-	<group>JobTransform</group>
-	<dataset_info>
-	  <jobTransformData />
-	  <datasetName>/afs/cern.ch/atlas/offline/ProdData/13.0.X/alpgen.005224.WenuNp3._00023.tar.gz</datasetName>
-	</dataset_info>
-      </jobTransform>
-
-    </jobList>
-       
-    <jobGroups>
-          
-      <jobGroup name="JobTransform" parent="Transform">
-	<keepFilePattern>*.root</keepFilePattern>
-	<keepFilePattern>*.txt</keepFilePattern>
-
-	<test>
-	  <modulename>RttLibraryTools</modulename>
-	  <testname>FileGrepper</testname>
-	  <arg>
-	    <argname>inputFile</argname>
-	    <argvalue>*_log</argvalue>
-	  </arg>
-	  <arg>
-	    <argname>outputFile</argname>
-	    <argvalue>RTT_search_results.txt</argvalue>
-	  </arg>
-	  <arg>
-	    <argname>searchList</argname>
-	    <argvalue>ErrorCode=0</argvalue>
-	  </arg>
-	</test>
-	
-      </jobGroup>
-      
-    </jobGroups>
-    
-  </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/RTTSummaryReference.xml
deleted file mode 100644
index cc706cbb1d338fdb2828135409b0cbe0be6d81d5..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/RTTSummaryReference.xml
+++ /dev/null
@@ -1,122 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1245091578.55</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/06/15@20:46: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/15@20:46</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransformWithData/Results</resBaseDir>
-    <endTime1>09/06/15@20:46</endTime1>
-    <endTime2>1245091578.53</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransformWithData/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1245090745.71</startTime2>
-    <startTime1>09/06/15@20:32</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>1</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>1</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>1</nJobsSuccessInPackage>
-    <packageName>EvgenJobTransforms</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Generators/EvgenJobTransforms</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>EvgenJobTransforms-00-06-26</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>1</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <nUserTests>1</nUserTests>
-      <jobDocURL>
-      </jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/jobtransformWithData/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0</resultsPath>
-      <batchStatus>DONE</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>486</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>1</chainLength>
-      <state>done</state>
-      <postProcessingResult>success</postProcessingResult>
-      <jobGroup>JobTransform</jobGroup>
-      <identifiedName>5224.AlpgenJimmyWenuNp30</identifiedName>
-      <status>success</status>
-      <chainSuccesses>1</chainSuccesses>
-      <mem>1MB</mem>
-      <stateHistory>[('queued', '09/06/15 20:33'), ('running', '09/06/15 20:33'), ('postProcessingQueued', '09/06/15 20:40'), ('postProcessingRunning', '09/06/15 20:40'), ('done', '09/06/15 20:46')]</stateHistory>
-      <jobDocString>5224.AlpgenJimmyWenuNp3</jobDocString>
-      <hashString>-1743466227</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>181</cpuTime>
-      <jobName>5224.AlpgenJimmyWenuNp3</jobName>
-      <vmem>104MB</vmem>
-      <nUserActions>2</nUserActions>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>success</processingResult>
-      <wallTime>364</wallTime>
-      <datasets>
-        <datasetinfo>
-          <file>alpgen.005224.WenuNp3._00023.tar.gz</file>
-          <directory>/afs/cern.ch/atlas/offline/ProdData/13.0.X</directory>
-          <mtime>Unknown</mtime>
-          <size>Unknown</size>
-          <origin>standard</origin>
-        </datasetinfo>
-      </datasets>
-      <missingDatasets/>
-      <archivers/>
-      <sUserTests>1</sUserTests>
-      <deliveredkeepfiles>
-        <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0</baseKeepFileDir>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/RTTtests.db</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postProcessor.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postprocessing.debug.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postprocessing.error.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/jobReport.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/RTT_search_results.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postprocessing.warning.log</keepfile_newstyle>
-        <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/env.log</keepfile_newstyle>
-        <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/5224.AlpgenJimmyWenuNp30_log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/listDir.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/CSC.005224.AlpgenJimmyWenuNp3.evgen.pool.root</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/jobReport_csc_evgen.txt</keepfile_newstyle>
-        <keepfile_newstyle info='Output from checkFile.py' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/CSC.005224.AlpgenJimmyWenuNp3.evgen.pool.root.checkFile</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postProcessorScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/postprocessing.info.log</keepfile_newstyle>
-        <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/runScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/EvgenJobTransforms/JobTransform/5224.AlpgenJimmyWenuNp3/0/package.xml</keepfile_newstyle>
-      </deliveredkeepfiles>
-      <errors/>
-      <done/>
-      <checkResults/>
-      <testResults>
-        <testResult>
-          <testName>FileGrepper_0</testName>
-          <testStatus>0</testStatus>
-        </testResult>
-      </testResults>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/TestCase.py
deleted file mode 100644
index f2c6351ec5e4496ace8a60f8ac7e5ca944d3a153..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/TestCase.py
+++ /dev/null
@@ -1,15 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testCMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/__init__.py
deleted file mode 100644
index 2ecbf8ab4b5a54b6fbf059fca93e1fb8a824a527..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/__init__.py
+++ /dev/null
@@ -1,3 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs a single job transform job that has uses a dataset_info tag."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/cfg/cfg.xml.TPL
deleted file mode 100644
index 82f3523e7bed6662027d0e25f5f20ceb5373c56f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/jobtransformWithData/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Generators/EvgenJobTransforms</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0</cmtpath>
-            <version>EvgenJobTransforms-00-06-26</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasOffline/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-
-   </confFileSrc>
-      
-   <packageAccept>EvgenJobTransforms</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/AthExHelloWorld/IHelloTool.h b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/AthExHelloWorld/IHelloTool.h
deleted file mode 100755
index 1ca87a4533d7c57293d78a788061cf3deb231a0e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/AthExHelloWorld/IHelloTool.h
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
-  Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-*/
-
-#ifndef ATHEXHELLOWORLD_IHELLOTOOL_H
-#define ATHEXHELLOWORLD_IHELLOTOOL_H 1
-
-#include "GaudiKernel/IAlgTool.h"
-
-class IHelloTool : virtual public IAlgTool {
-public:
-   virtual StatusCode saySomething() = 0;
-   static const InterfaceID& interfaceID() {
-     static const InterfaceID _IHelloToolID( "IHelloTool", 1, 0 );
-     return _IHelloToolID;
-   }
-};
-#endif // !ATHEXHELLOWORLD_IHELLOTOOL_H
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/ChangeLog b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/ChangeLog
deleted file mode 100755
index efb3b1f387fa21b0176e38acf501ba27654746b9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/ChangeLog
+++ /dev/null
@@ -1,119 +0,0 @@
-2008-09-30  Sebastien Binet  <binet@lblbox>
-
-	* tagging AthExHelloWorld-01-02-00
-	* migration to AthenaBaseComps
-	* M AthExHelloWorld/IHelloTool.h
-	* M cmt/requirements
-	* M src/HelloAlg.cxx
-	* M src/HelloAlg.h
-	* M src/HelloTool.cxx
-	* M src/HelloTool.h
-
-2007-12-14  Paolo Calafiura  <calaf@0-11-25-81-1e-12.dhcp.lbl.gov>
-
-	* src/HelloAlg.cxx (operator<<): OBO Steve Gowdy. Add a streamer
-	to pair<double, double> to make gcc41 happy.
-        * tag AthExHelloWorld-01-01-01
-
-2007-11-21  Alex Undrus <undrus@bnl.gov>
-        * test/AthExHelloWorld_TestConfiguration.xml
-	* in RTT part: replace AthExHelloWorld_jobOptions 
-	* with HelloWorldOptions job options.
-	* as an exception: retag AthExHelloWorld-01-01-00
-
-2007-11-08  Paolo Calafiura  <pcalafiura@lbl.gov>
-	* cmt/requirements: remove reference to remove job opt file
-        * tag AthExHelloWorld-01-01-00
-	
-2007-11-07  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * python-side cleanup and educational comments
-        * tagging as AthExHelloWorld-01-00-04
-
-2007-08-10  Alex Undrus <undrus@bnl.gov>
-        * tagging AthExHelloWorld-01-00-03
-	* in test/AthExHelloWorld_TestConfiguration.xml
-	* added doc, classification, mailto, rttContactPerson tags
-	
-2007-04-19  Sebastien Binet  <binet@lblbox>
-	* tagging AthExHelloWorld-01-00-02
-	* added a std::vector<std::pair<double, double> > property
-	* added a std::vector<std::vector<double> > property
-
-	* tagging AthExHelloWorld-01-00-01
-	* adding a std::map<std::string,std::string> property
-
-2007-02-02  Paolo Calafiura  <pcalafiura@lbl.gov>
-	* general cleanup. Renamed HelloWorld->HelloAlg. Introduced tool iface
-        * tag AthExHelloWorld-01-00-00
-	
-2006-10-20 Martin Woudstra <MJWoudstra@lbl.gov>
-	* share/AthExHelloWorld_jobOptions.py: use latest ToolHandle features
-	requires  AthenaCommon-02-04-26 or later
-	requires Gaudi-0.16.1.16-LCG47b or later
-	tagged with AthExHelloWorld-00-02-02
-
-2006-10-06 Martin Woudstra <MJWoudstra@lbl.gov>
-	* HelloWorld.h/.cxx:
-	  - use latest feature of ToolHandle: propertyName()
-	  - add missing #include	
-	
-2006-10-05 Martin Woudstra <MJWoudstra@lbl.gov>
-	* have working ToolHandle example. Needs AthenaCommon-02-04-24 or later
-        * tagged with AthExHelloWorld-00-02-01
-	
-2006-07-29  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * tagged with AthExHelloWorld-00-02-00
-
-2006-07-14  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * more elaborate job options settings
-
-2006-07-13  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * added tool to play and learn
-
-2006-07-05  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * removed .txt configuration file
-        * added genconf pattern and finished AthExHelloWorld_jobOptions.py accordingly
-
-2006-03-27  Alex Undrus <undrus@bnl.gov>
-	* test/AthExHelloWorld_TestConfiguration.xml : replace <options> with
-	* <options_atn> for ATN part (to avoid clash with RTT)
-	* tag AthExHelloWorld-00-01-16
-
-2005-10-31  Paolo Calafiura  <pcalafiura@lbl.gov>
-	* src/components/HelloWorld_entries.cxx : remove extra ; (Matthias) 
-        * tag AthExHelloWorld-00-01-15
-
-2005-09-04  Paolo Calafiura  <pcalafiura@lbl.gov>
-	* python/AthExHelloWorldConfig.py (HelloWorld.gaudiType):
-	required method added
-
-2005-07-14  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * added beginRun/endRun to example code
-
-2005-05-27  Wim Lavrijsen <WLavrijsen@lbl.gov>
-        * added example code making use of new config facilities
-
-2005-05-19  Alex Undrus <undrus@bnl.gov>
-	* cmt/requirements: added macro and runtime extras 
-	* for RTT system
-        * tag AthExHelloWorld-00-01-14
-	
-2005-05-19  Alex Undrus <undrus@bnl.gov>
-	* added test area with the unified configuration file
-        * (for automatic test running in ATN and RTT systems)
-	
-2004-12-03  David Quarrie <David.Quarrie@cern.ch>
-	* cmt/requirements: Remove redundant commented out statements
-
-2003-12-06 Paolo Calafiura <pcalafiura@lbl.gov>
-        * cmt/requirements: remoive redundant? use statements
-        tag AthExHelloWorld-00-01-11
-
-2003-11-17 Paolo Calafiura <pcalafiura@lbl.gov>
-        * cmt/requirements: explicitly install job opts in run area
-        tag AthExHelloWorld-00-01-09
-
-2003-11-14  Christian Arnault  <arnault@lal.in2p3.fr>
-
-	* share/HelloWorldOptions.txt: Adapt to the new convention
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/share/HelloWorldOptions.py b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/share/HelloWorldOptions.py
deleted file mode 100644
index 261907032c83553e63f0c58caa2a806b29e4426f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/share/HelloWorldOptions.py
+++ /dev/null
@@ -1,112 +0,0 @@
-###############################################################
-#
-# Job options file
-#
-#==============================================================
-
-#--------------------------------------------------------------
-# ATLAS default Application Configuration options
-#--------------------------------------------------------------
-
-# No event selector needed for basic 'Hello World!' Algorithm
-
-#--------------------------------------------------------------
-# Private Application Configuration options
-#--------------------------------------------------------------
-
-# Full job is a list of algorithms
-from AthenaCommon.AlgSequence import AlgSequence
-job = AlgSequence()
-
-# Add top algorithms to be run
-from AthExHelloWorld.AthExHelloWorldConf import HelloAlg
-job += HelloAlg( "HelloWorld" )   # 1 alg, named "HelloWorld"
-
-#--------------------------------------------------------------
-# Set output level threshold (DEBUG, INFO, WARNING, ERROR, FATAL)
-#--------------------------------------------------------------
-
-# Output level for HelloAlg only (note name: instance, not type)
-job.HelloWorld.OutputLevel = INFO
-
-# You can set the global output level on the message svc (not
-# recommended) or by using the -l athena CLI parameter
-
-#--------------------------------------------------------------
-# Event related parameters
-#--------------------------------------------------------------
-
-# Number of events to be processed (default is until the end of
-# input, or -1, however, since we have no input, a limit needs
-# to be set explicitly, here, choose 10)
-theApp.EvtMax = 10
-
-#--------------------------------------------------------------
-# Algorithms Private Options (all optional)
-#--------------------------------------------------------------
-
-# For convenience, get a reference to the HelloAlg Algorithm
-# named "HelloWorld" in the job
-HelloWorld = job.HelloWorld
-
-# Set an int property
-HelloWorld.MyInt = 42
-
-# Set a boolean property (False, True, 0, 1)
-HelloWorld.MyBool = True
-
-# Set a double property
-HelloWorld.MyDouble = 3.14159
-
-# Set a vector of strings property ...
-HelloWorld.MyStringVec = [ "Welcome", "to", "Athena", "Framework", "Tutorial" ]
-
-# ... and add one more:
-HelloWorld.MyStringVec += [ "!" ]
-
-# Set a map of strings to strings property ...
-HelloWorld.MyDict = { 'Bonjour'      : 'Guten Tag',
-                      'Good Morning' : 'Bonjour' , 'one' : 'uno' }
-
-# ... and add one more:
-HelloWorld.MyDict[ "Goeiedag" ] = "Ni Hao"
-
-# Set a table (a vector of pairs of doubles) ...
-HelloWorld.MyTable = [ ( 1 , 1 ) , ( 2 , 4 ) , ( 3 , 9 ) ]
-
-# ... and one more:
-HelloWorld.MyTable += [ ( 4, 16 ) ]
-
-# Set a matrix (a vector of vectors) ...
-HelloWorld.MyMatrix = [ [ 1, 2, 3 ],
-                        [ 4, 5, 6 ] ]
-
-# ... and some more:
-HelloWorld.MyMatrix += [ [ 7, 8, 9 ] ]
-
-#--------------------------------------------------------------
-# Algorithms Tool Usage Private Options (advanced and optional)
-#--------------------------------------------------------------
-
-# Import configurable for using our HelloTool
-from AthExHelloWorld.AthExHelloWorldConf import HelloTool
-
-# Setup a public tool so that it can be used (again, note name)
-ToolSvc += HelloTool( "PublicHello" )
-ToolSvc.PublicHello.MyMessage = "A Public Message!"
-
-# Tell "HelloWorld" to use this tool ("MyPublicHelloTool" is a
-# ToolHandle property of HelloAlg)
-HelloWorld.MyPublicHelloTool = ToolSvc.PublicHello
-
-# Hand "HelloWorld" a private HelloTool ("MyPrivateHelloTool" is
-# a ToolHandler property of HelloAlg)
-HelloWorld.MyPrivateHelloTool = HelloTool( "HelloTool" )
-HelloWorld.MyPrivateHelloTool.MyMessage = "A Private Message!"
-
-#==============================================================
-#
-# End of job options file
-#
-###############################################################
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.cxx b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.cxx
deleted file mode 100755
index af683e05f70312951e6c9ed95dfd42a06e77fb98..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.cxx
+++ /dev/null
@@ -1,182 +0,0 @@
-/*
-  Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-*/
-
-// STL includes
-#include <iterator>
-
-#include "AthExHelloWorld/IHelloTool.h"
-#include "HelloAlg.h"
-
-/////////////////////////////////////////////////////////////////////////////
-// FIXME Looks like we need operator<<( ostream&, pair<double, double > ) for gcc41
-std::ostream& operator<<( std::ostream& s, std::pair<double, double > p )
-{
-  s << p.first << " " << p.second;
-  return s;
-}
-
-HelloAlg::HelloAlg(const std::string& name, ISvcLocator* pSvcLocator) :
-  AthAlgorithm(name, pSvcLocator), m_myInt(0), m_myBool(0), m_myDouble(0),
-  m_myPrivateHelloTool("HelloTool",this), m_myPublicHelloTool("HelloTool"),
-  m_myDict(),
-  m_myTable(),
-  m_myMatrix()
-{
-  
-  // Part 2: Declare the properties
-  declareProperty("MyInt", m_myInt);
-  declareProperty("MyBool", m_myBool);
-  declareProperty("MyDouble", m_myDouble);
-
-  declareProperty("MyStringVec",m_myStringVec, "an entire vector of strings!");
-
-  declareProperty("MyPrivateHelloTool", m_myPrivateHelloTool, "private IHelloTool");
-  declareProperty("MyPublicHelloTool", m_myPublicHelloTool, "public, shared IHelloTool");
-
-  declareProperty("MyDict",
-		  m_myDict,
-		  "A little dictionary" );
-  // some default values;
-  m_myDict["Bonjour"]      = "Guten Tag";
-  m_myDict["Good Morning"] = "Bonjour";
-  m_myDict["one"]          = "uno";
-
-  declareProperty("MyTable", m_myTable, "A table of <double,double>" );
-  // some default values
-  m_myTable.push_back( std::make_pair( 1., 1. ) );
-  m_myTable.push_back( std::make_pair( 2., 2.*2. ) );
-  m_myTable.push_back( std::make_pair( 3., 3.*3. ) );
-
-  declareProperty("MyMatrix", m_myMatrix, "A matrix of doubles" );
-
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * 
-
-StatusCode HelloAlg::initialize() {
-
-  // Part 1: print where you are
-  ATH_MSG_INFO ("initialize()");
-
-  // Part 2: Print out the property values
-  ATH_MSG_INFO
-    (   "  MyInt =    " << m_myInt << endmsg
-     << "  MyBool =   " << (int)m_myBool << endmsg
-     << "  MyDouble = " << m_myDouble);
-
-  for (unsigned int i=0; i<m_myStringVec.size(); i++) {
-    ATH_MSG_INFO ("  MyStringVec[" << i << "] = " << m_myStringVec[i]);
-  }
-
-  for ( Dict_t::const_iterator itr = m_myDict.begin();
-	itr != m_myDict.end();
-	++itr ) {
-    ATH_MSG_INFO
-      ("  MyDict['" << itr->first << "'] = '" << itr->second << "'");
-  }
-  for ( Table_t::const_iterator itr = m_myTable.begin();
-	itr != m_myTable.end();
-	++itr ) {
-    ATH_MSG_INFO 
-      ("  MyTable['" << itr->first << "'] = '" << itr->second << "'");
-  }
-  for (unsigned int i=0; i<m_myMatrix.size(); i++) {
-    msg(MSG::INFO) << "  MyMatrix[" << i << "] = [ ";
-    std::copy( m_myMatrix[i].begin(), m_myMatrix[i].end(), 
-	       std::ostream_iterator<double>(msg().stream(), " ") );
-    msg() << "]" << endmsg;
-  }
-
-  ATH_MSG_INFO 
-    ("  " << m_myPrivateHelloTool.propertyName() 
-     << " = " << m_myPrivateHelloTool.type() 
-     << endmsg
-     << "  " << m_myPublicHelloTool.propertyName() 
-     << " = " << m_myPublicHelloTool.type());
-
-
-  // Part 3: Retrieve the tools using the ToolHandles
-  if ( m_myPrivateHelloTool.retrieve().isFailure() ) {
-    ATH_MSG_FATAL 
-      (m_myPrivateHelloTool.propertyName() << ": Failed to retrieve tool "
-       << m_myPrivateHelloTool.type());
-    return StatusCode::FAILURE;
-  } else {
-    ATH_MSG_INFO 
-      (m_myPrivateHelloTool.propertyName() << ": Retrieved tool " 
-       << m_myPrivateHelloTool.type());
-  }
-  if ( m_myPublicHelloTool.retrieve().isFailure() ) {
-    ATH_MSG_FATAL 
-      (m_myPublicHelloTool.propertyName() << ": Failed to retrieve tool "
-       << m_myPublicHelloTool);
-    return StatusCode::FAILURE;
-  } else {
-    ATH_MSG_INFO (m_myPublicHelloTool.propertyName() << ": Retrieved tool "
-		  << m_myPublicHelloTool.type());
-  }
-
-  return StatusCode::SUCCESS;
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * 
-
-StatusCode HelloAlg::execute() {
-
-  // Part 1: print where you are
-  ATH_MSG_INFO ("execute()");
-
-  // Part 1: Print out the different levels of messages
-  ATH_MSG_DEBUG   ("A DEBUG message");
-  ATH_MSG_INFO    ("An INFO message");
-  ATH_MSG_WARNING ("A WARNING message");
-  ATH_MSG_ERROR   ("An ERROR message");
-  ATH_MSG_FATAL   ("A FATAL error message");
-
-  ATH_MSG_FATAL   ("An RTT FATAL message, run for the hills!");
-
-  // Part 1a: Let publicly declared tool say something
-  ATH_MSG_INFO ("Let the tool " << m_myPublicHelloTool.propertyName() 
-		<< " say something:");
-  StatusCode sc1 = m_myPublicHelloTool->saySomething();
-
-    // Part 1b: Let privately declared tool say something
-  ATH_MSG_INFO ("Let the tool " << m_myPrivateHelloTool.propertyName()
-		<< " say something:");
-  StatusCode sc2 = m_myPrivateHelloTool->saySomething();
-
-  if ( sc1.isFailure() || sc2.isFailure() ) {
-    return StatusCode::FAILURE;
-  }
-
-  return StatusCode::SUCCESS;
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * 
-
-StatusCode HelloAlg::finalize() {
-
-  // Part 1: print where you are
-  ATH_MSG_INFO ("finalize()");
-
-  return StatusCode::SUCCESS;
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
-
-StatusCode HelloAlg::beginRun() {
-  // Part 1: print where you are
-  ATH_MSG_INFO ("beginRun()");
-
-  return StatusCode::SUCCESS;
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
-
-StatusCode HelloAlg::endRun() {
-  // Part 1: print where you are
-  ATH_MSG_INFO ("endRun()");
-
-  return StatusCode::SUCCESS;
-}
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.h b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.h
deleted file mode 100755
index de845b47812ecdcb97a04b08a8970378fd10b2c9..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloAlg.h
+++ /dev/null
@@ -1,49 +0,0 @@
-// -*- C++ -*-
-
-/*
-  Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-*/
-
-#ifndef ATHEXHELLOWORLD_HELLOALG_H
-#define ATHEXHELLOWORLD_HELLOALG_H 1
-
-#include "GaudiKernel/ToolHandle.h"
-#include "AthenaBaseComps/AthAlgorithm.h"
-
-#include <string>
-#include <vector>
-#include <utility>
-#include <map>
-
-/////////////////////////////////////////////////////////////////////////////
-
-class IHelloTool;
-
-class HelloAlg : public AthAlgorithm {
-public:
-  HelloAlg( const std::string& name, ISvcLocator* pSvcLocator );
-  StatusCode initialize();
-  StatusCode execute();
-  StatusCode finalize();
-
-  StatusCode beginRun();
-  StatusCode endRun();
-  
-private:
-  int    m_myInt;
-  bool   m_myBool;
-  double m_myDouble;
-  std::vector< std::string > m_myStringVec;
-
-  ToolHandle< IHelloTool > m_myPrivateHelloTool;
-  ToolHandle< IHelloTool > m_myPublicHelloTool;
-
-  typedef std::map<std::string, std::string> Dict_t;
-  Dict_t m_myDict;
-  typedef std::vector<std::pair<double, double> > Table_t;
-  Table_t m_myTable;
-  typedef std::vector<std::vector<double> > Matrix_t;
-  Matrix_t m_myMatrix;
-};
-
-#endif // ATHEXHELLOWORLD_HELLOALG_H
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.cxx b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.cxx
deleted file mode 100755
index 0e639c0096cebbf0360cca70a62a3ddcd5886b3c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.cxx
+++ /dev/null
@@ -1,36 +0,0 @@
-/*
-  Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-*/
-
-#include "HelloTool.h"
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
-
-HelloTool::HelloTool( const std::string& type, const std::string& name,
-		      const IInterface* parent ) 
-  : AthAlgTool( type, name, parent ), m_myMessage("Default message set in HelloTool.cxx")
-{
-  declareProperty( "MyMessage", m_myMessage, "something to say" );
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
-
-StatusCode HelloTool::queryInterface( const InterfaceID& riid, void** ppvIf )
-{
-   if ( riid == IHelloTool::interfaceID() )  {
-      *ppvIf = (IHelloTool*)this;
-      addRef();
-      return StatusCode::SUCCESS;
-   }
-
-   return AthAlgTool::queryInterface( riid, ppvIf );
-}
-
-// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
-
-StatusCode HelloTool::saySomething()
-{
-  ATH_MSG_INFO ("my message to the world: " << m_myMessage);
-
-   return StatusCode::SUCCESS;
-}
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.h b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.h
deleted file mode 100755
index 14c2191620afb9c6df62cd60085068e7190e4969..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/HelloTool.h
+++ /dev/null
@@ -1,27 +0,0 @@
-/*
-  Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-*/
-
-#ifndef ATHEXHELLOWORLD_HELLOTOOL_H
-#define ATHEXHELLOWORLD_HELLOTOOL_H 1
-
-#include "AthenaBaseComps/AthAlgTool.h"
-#include "AthExHelloWorld/IHelloTool.h"
-
-#include <string>
-
-class HelloTool : virtual public IHelloTool, virtual public AthAlgTool {
-public:
-   HelloTool( const std::string&, const std::string&, const IInterface* );
-
-// to allow access to the IHelloTool interface
-   StatusCode queryInterface( const InterfaceID& riid, void** ppvIf );
-
-// the magic method this tool provides
-   virtual StatusCode saySomething();
-
-private:
-   std::string m_myMessage;
-};
-
-#endif
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/components/AthExHelloWorld_entries.cxx b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/components/AthExHelloWorld_entries.cxx
deleted file mode 100644
index 2dc672dadc22a87b0e005f25e8f388010a6ebc79..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/src/components/AthExHelloWorld_entries.cxx
+++ /dev/null
@@ -1,7 +0,0 @@
-#include "../HelloAlg.h"
-#include "../HelloTool.h"
-
-
-DECLARE_COMPONENT( HelloAlg )
-DECLARE_COMPONENT( HelloTool )
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/test/AthExHelloWorld_TestConfiguration.xml b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/test/AthExHelloWorld_TestConfiguration.xml
deleted file mode 100755
index ae7d44f380f8841e8d19f98d360ed1a4385f99bc..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/Control/AthenaExamples/AthExHelloWorld/test/AthExHelloWorld_TestConfiguration.xml
+++ /dev/null
@@ -1,46 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-  <atn>
-     <TEST name="HelloWorld" type="athena" suite="Examples">
-        <options_atn>AthExHelloWorld/HelloWorldOptions.py</options_atn>
-        <timelimit>2</timelimit>
-        <author> Alex Undrus </author>
-        <mailto> undrus@bnl.gov </mailto>
-        <expectations>
-           <errorMessage>FAILURE (ERROR)</errorMessage>
-           <successMessage>FATAL A FATAL</successMessage>
-           <returnValue>0</returnValue>
-        </expectations>
-     </TEST>
-   </atn>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <rttContactPerson>Alex Undrus (undrus@bnl.gov)</rttContactPerson>
-      <mailto>undrus@bnl.gov</mailto>
-      <refRelease>13.0.0</refRelease>
-      
-      <jobList>
-
-      <classification>
-         <displayClass>OfflineValidation</displayClass>
-         <displayProcess>Core</displayProcess>
-         <displayComponent>Athena-Core</displayComponent>
-      </classification>
-
-         <athena>
-            <doc> Run Athena HelloWorld example </doc>
-            <options>AthExHelloWorld/HelloWorldOptions</options>
-            <package>Control/AthenaExamples/AthExHelloWorld</package>
-            <group>HelloWorld</group>
-            <queue>short</queue>
-         </athena>
-      </jobList>
-
-      <jobGroups>
-         <jobGroup name="AthenaHelloWorld" parent="Athena"/>
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/RTTSummaryReference.xml
deleted file mode 100644
index f9827f02eb37d1431855bafd3f0b32409c44aa33..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/RTTSummaryReference.xml
+++ /dev/null
@@ -1,109 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1239393297.89</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/04/10@21:54: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/04/10@21:54</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/linuxinteractive/Results</resBaseDir>
-    <endTime1>09/04/10@21:54</endTime1>
-    <endTime2>1239393297.87</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-61</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/linuxinteractive/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1239393253.95</startTime2>
-    <startTime1>09/04/10@21:54</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-61</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>1</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-02-00</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <displayClass>OfflineValidation</displayClass>
-      <displayComponent>Athena-Core</displayComponent>
-      <nUserTests>0</nUserTests>
-      <jobDocURL>
-      </jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/RTTTest_LI/testsuites/linuxinteractive/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0</resultsPath>
-      <batchStatus>Unknown</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>Unknown</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>1</chainLength>
-      <state>done</state>
-      <postProcessingResult>no tests</postProcessingResult>
-      <jobGroup>AthenaHelloWorld</jobGroup>
-      <identifiedName>HelloWorldOptions0</identifiedName>
-      <status>success</status>
-      <chainSuccesses>1</chainSuccesses>
-      <mem>Unknown</mem>
-      <stateHistory>[('queued', '09/04/10 21:54'), ('running', '09/04/10 21:54'), ('postProcessingQueued', '09/04/10 21:54'), ('postProcessingRunning', '09/04/10 21:54'), ('done', '09/04/10 21:54')]</stateHistory>
-      <jobDocString>Run Athena HelloWorld example</jobDocString>
-      <hashString>-1301877828</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>Unknown</cpuTime>
-      <jobName>HelloWorldOptions</jobName>
-      <vmem>Unknown</vmem>
-      <nUserActions>3</nUserActions>
-      <displayProcess>Core</displayProcess>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>success</processingResult>
-      <wallTime>Unknown</wallTime>
-      <datasets/>
-      <missingDatasets/>
-      <sUserTests>0</sUserTests>
-      <deliveredkeepfiles>
-        <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0</baseKeepFileDir>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.warning.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/RTTtests.db</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.info.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.debug.log</keepfile_newstyle>
-        <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/runScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/HelloWorldOptions0_log</keepfile_newstyle>
-        <keepfile_newstyle info='Job options called with the athena command' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/AthExHelloWorld/HelloWorldOptions.py</keepfile_newstyle>
-        <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/env.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postprocessing.error.log</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postProcessor.log</keepfile_newstyle>
-        <keepfile_newstyle info='ls -altF command on run directory before and after athena command' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/lsaltF.txt</keepfile_newstyle>
-        <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/postProcessorScript.sh</keepfile_newstyle>
-        <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/AthExHelloWorld_TestConfiguration.xml</keepfile_newstyle>
-        <keepfile_newstyle info='cmt show uses command on package cmt dir' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/cmtShowUses.log</keepfile_newstyle>
-        <keepfile_newstyle info='Marker emitted at the end of the job' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0/JobCompleteMarker.txt</keepfile_newstyle>
-      </deliveredkeepfiles>
-      <errors/>
-      <done/>
-      <checkResults/>
-      <testResults/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestCase.py
deleted file mode 100644
index 5be06e751d523fa96aea34acb3e9781e2e72e089..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestCase.py
+++ /dev/null
@@ -1,25 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-class TestCase(RTTTestCase):
-
-    def variableContentTags(self):
-        """
-        Add the hashstring tag to the variable content
-        tag list. This job takes files from the
-        chain store, so the content of the file tag depends on
-        the (variable) name of the chain store.
-        """
-
-        tags = RTTTestCase.variableContentTags(self)
-        tags.extend(['directory','hashString'])
-        return tags
-    
-    def testCMLComp(self): self.rttTestXMLComp()
-    def testNKeepFiles(self): self.rttTestNTags(self.dom, 'keepfile_newstyle', 16)
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestProject/ChangeLog b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestProject/ChangeLog
deleted file mode 100755
index dbb26014017e3435d2f2fa61ae3606dc428982e8..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/TestProject/ChangeLog
+++ /dev/null
@@ -1,11 +0,0 @@
-2006-05-25   David Quarrie <David.Quarrie@cern.ch>
-
-        * cmt/requirements: Add support for the AtlasProduction project
-
-2005-12-12  David Quarrie <David.Quarrie@cern.ch>
-
-	* cmt/requirements: Add support for all projects
-
-2005-08-19  David Quarrie <David.Quarrie@cern.ch>
-
-	* Initial version
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/__init__.py
deleted file mode 100644
index 55c4101300ab7b30dd755951168d460402988e6c..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/__init__.py
+++ /dev/null
@@ -1,17 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs a single helloworld job in linux interactive mode."""
-
-def run():
-    import os
-
-    # Setup the cmt home dir
-    cmd  = 'cd cmtHome;'
-    cmd += 'source /afs/cern.ch/sw/contrib/CMT/v1r20p20080222/mgr/setup.sh>& /dev/null;'
-    cmd += 'cmt config >& /dev/null;'
-
-    # Build the HW package
-    cmd += 'source setup.sh -tag=opt,32,gcc34,AtlasProduction,14.5.0,dev,setup,oneTest >& /dev/null;'
-    cmd += 'cd ../Control/AthenaExamples/AthExHelloWorld/cmt;'
-    cmd += 'gmake >& /dev/null;'
-    os.system(cmd)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cfg/cfg.xml.TPL
deleted file mode 100644
index fe786f659cf41c51c2917777ee58b0b4f6fb19dd..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,42 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LinuxInteractive</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <localConfFiles>
-         <userHomeCmtDir>$_BASE_$/testsuites/$_TESTNAME_$/cmtHome</userHomeCmtDir>
-         <testProjectCmtDir>$_BASE_$/testsuites/$_TESTNAME_$/TestProject/cmt</testProjectCmtDir>
-      </localConfFiles>
-
-   </confFileSrc>
-      
-   <packageAccept>AthExHelloWorld</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cmtHome/requirements.TPL b/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cmtHome/requirements.TPL
deleted file mode 100644
index 5b5308ccb87c03d9444cbb8f3570d30af8e564d1..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/linuxinteractive/cmtHome/requirements.TPL
+++ /dev/null
@@ -1,5 +0,0 @@
-set CMTSITE CERN
-set SITEROOT /afs/cern.ch
-macro ATLAS_DIST_AREA $(SITEROOT)/atlas/software/dist
-macro ATLAS_TEST_AREA $_BASE_$/testsuites/$_TESTNAME_$
-use AtlasLogin AtlasLogin-* $(ATLAS_DIST_AREA)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/pseudo/Package/package.xml.TPL
deleted file mode 100644
index 611899411642fb6d0337dbe0b31595988a9e7faa..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/Package/package.xml.TPL
+++ /dev/null
@@ -1,30 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-          <pseudo>
-              <athena>
-                  <doc>blah</doc>
-                  <doclink>http://cern.ch</doclink>
-                  <options>AthExHelloWorld/HelloWorldOptions.py</options>
-                  <package>Control/AthenaExamples/AthExHelloWorld</package>
-                  <group>HelloWorld</group>
-                  <queue>short</queue>
-              </athena>
-          </pseudo>
-
-       </jobList>
-
-	<jobGroups>          
-		<jobGroup name= "AthenaHelloWorld" parent="Athena"/>          
-	</jobGroups>          
-          
-    </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/pseudo/RTTSummaryReference.xml
deleted file mode 100644
index 009a26ddc64319e729e2b9cf50e81fc917f37a69..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/RTTSummaryReference.xml
+++ /dev/null
@@ -1,89 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1244808233.27</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/06/12@14:03: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/06/12@14:03</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/pseudo/Results</resBaseDir>
-    <endTime1>09/06/12@14:03</endTime1>
-    <endTime2>1244808233.25</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-65</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/pseudo/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1244807889.12</startTime2>
-    <startTime1>09/06/12@13:58</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-65</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>
-    </packageTag>
-    <nJobsDoneInPackage>0</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <minder>
-      <nUserTests>0</nUserTests>
-      <jobDocURL>http://cern.ch</jobDocURL>
-      <resultsPath>/afs/cern.ch/atlas/project/RTT/data/peter/NewTest/RTTTest/testsuites/pseudo/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/AthenaHelloWorld/HelloWorldOptions/0</resultsPath>
-      <batchStatus>Unknown</batchStatus>
-      <jobDisplayName>
-      </jobDisplayName>
-      <cpuTime2000>Unknown</cpuTime2000>
-      <jobID>0</jobID>
-      <chainLength>0</chainLength>
-      <state>done</state>
-      <postProcessingResult>unavailable</postProcessingResult>
-      <jobGroup>AthenaHelloWorld</jobGroup>
-      <identifiedName>HelloWorldOptions0</identifiedName>
-      <status>done</status>
-      <chainSuccesses>0</chainSuccesses>
-      <mem>Unknown</mem>
-      <stateHistory>[('queued', '09/06/12 14:03'), ('running', '09/06/12 14:03'), ('done', '09/06/12 14:03')]</stateHistory>
-      <jobDocString>blah</jobDocString>
-      <hashString>1079629659</hashString>
-      <nRetries>0</nRetries>
-      <cpuTime>Unknown</cpuTime>
-      <jobName>HelloWorldOptions</jobName>
-      <vmem>Unknown</vmem>
-      <nUserActions>3</nUserActions>
-      <exitStatus>Unknown</exitStatus>
-      <processingResult>unavailable</processingResult>
-      <wallTime>Unknown</wallTime>
-      <datasets/>
-      <missingDatasets/>
-      <archivers/>
-      <sUserTests>0</sUserTests>
-      <deliveredkeepfiles/>
-      <errors/>
-      <checkResults/>
-      <testResults/>
-    </minder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/pseudo/TestCase.py
deleted file mode 100644
index f1102c2a3db72a5f6d034f6293f91f3357122587..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/TestCase.py
+++ /dev/null
@@ -1,16 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-
-class TestCase(RTTTestCase):
-
-    def testXMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/pseudo/__init__.py
deleted file mode 100644
index 52b7d58c7bdd702d2f19023de150e5eb4faa0243..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/__init__.py
+++ /dev/null
@@ -1,4 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-"""Runs as a batch mode but all methods including command to submit
-job are dummies. Checks instantiation + dome code. Fast check."""
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/pseudo/cfg/cfg.xml.TPL
deleted file mode 100644
index 9f48cc74001fd5d8074d434a6b0a11b28782f189..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-         
-         <packageConfFile>
-            <package>Control/AthenaExamples/AthExHelloWorld</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-            <version></version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-         
-   </confFileSrc>
-      
-   <packageAccept>AthExHelloWorld</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-      
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/pseudo/checks.py b/Tools/RunTimeTester/testsuite/testsuites/pseudo/checks.py
deleted file mode 100644
index b87557d356be6a62855e1f6b924cfcd574e13816..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/pseudo/checks.py
+++ /dev/null
@@ -1,32 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from utility import log
-from summaryFile import RTTSummaryFile
-import os
-
-def run():
-    for klass in [XMLFileChecker]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class XMLFileChecker:
-    def __init__(self, tester):
-        self.tester   = tester
-        self.summfile = RTTSummaryFile(self.tester.paths.RTTSummary)
-
-    def run(self):
-        self.hasZeroJobs()
-        
-    @log
-    def hasZeroJobs(self):
-        return len(self.summfile.packages[0].jobs) == 0
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-
-
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/Package/package.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/Package/package.xml.TPL
deleted file mode 100644
index aa176ad211aeee7aad5d4fe41ae1a8defe50de92..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/Package/package.xml.TPL
+++ /dev/null
@@ -1,86 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-
-<unifiedTestConfiguration>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/rtt">
-
-      <mailto>brinick.simmons@cern.ch</mailto>
-      <refRelease>14.5.0</refRelease>
-
-      <jobList>
-
-         <chain>
-            <chainName>SimpleMergeJob</chainName>
-            <sequential>
-               <parallel>
-
-                  <chainElement>
-                      <jobTransform>
-                          <doc>MyJob</doc>
-                          <jobTransformJobName>SimpleJob1</jobTransformJobName>
-                          <jobTransformCmd>echo "1 2 3 4 5 6 7 8 9 10" &gt;&amp; simple1.data;echo "ErrorCode=0 (OK)";</jobTransformCmd>
-                          <group>JobTransform</group>
-                          <keepFilePattern>simple1.data</keepFilePattern>
-                      </jobTransform>
-
-                      <chainfileout>simple1.data</chainfileout>
-                  </chainElement>
-
-                  <chainElement>
-                       <jobTransform>
-                          <doc>MyJob</doc>
-                          <jobTransformJobName>SimpleJob2</jobTransformJobName>
-                          <jobTransformCmd>echo "11 12 13 14 15 16 17 18 19 20" &gt;&amp; simple2.data;echo "ErrorCode=0 (OK)";</jobTransformCmd>
-                          <group>JobTransform</group>
-                          <keepFilePattern>simple2.data</keepFilePattern>
-                      </jobTransform>
-
-                      <chainfileout>simple2.data</chainfileout>
-                  </chainElement>
-               </parallel>
-               <chainElement>
-
-                  <jobTransform>
-                     <doc>MyJob</doc>
-                     <jobTransformJobName>SimpleMergeJob0</jobTransformJobName>
-                     <jobTransformCmd>cat simple1.data simple2.data &gt;&amp; simple.merged.data;echo "ErrorCode=0 (OK)";</jobTransformCmd>
-                     <group>JobTransform</group>
-
-                     <chaindataset_info>
-                        <jobTransformData/>
-                        <chaindatasetName>simple1.data</chaindatasetName>
-                        <dataset_info>
-                           <jobTransformData/>
-                           <datasetName>$_BASE_$/data/$_TESTNAME_$/simple1.data</datasetName>
-                        </dataset_info>
-                     </chaindataset_info>
-
-                     <chaindataset_info>
-                        <jobTransformData/>
-                        <chaindatasetName>simple2.data</chaindatasetName>
-                        <dataset_info>
-                           <jobTransformData/>
-                           <datasetName>$_BASE_$/data/$_TESTNAME_$/simple2.data</datasetName>
-                        </dataset_info>
-                     </chaindataset_info>
-                     <keepFilePattern>simple.merged.data</keepFilePattern>
-                  </jobTransform>
-
-
-               </chainElement>
-
-            </sequential>
-         </chain>
-
-       </jobList>
-          
-       <jobGroups>
-          
-           <jobGroup name="JobTransform" parent="Transform">
-
-           </jobGroup>          
-
-       </jobGroups>
-          
-    </rtt>
-</unifiedTestConfiguration>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/RTTSummaryReference.xml b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/RTTSummaryReference.xml
deleted file mode 100644
index 82543d140d3f3c1901dd2b6a53fe67c86e90ac74..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/RTTSummaryReference.xml
+++ /dev/null
@@ -1,276 +0,0 @@
-
-<RTTSummary>
-  <overview>
-    <topProject>AtlasProduction</topProject>
-    <latestUpdateTime2>1247653513.87</latestUpdateTime2>
-    <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-    <releaseType>project</releaseType>
-    <site>CERNslc4</site>
-    <releaseName>14.5.0</releaseName>
-    <statusText>09/07/15@12:25: jobs succ/done/tot: 1/1/1</statusText>
-    <otherProject>
-    </otherProject>
-    <runType>build</runType>
-    <latestUpdateTime1>09/07/15@12:25</latestUpdateTime1>
-    <resBaseDir>/afs/cern.ch/user/r/rtt/test2/testsuites/simplemergejob/Results</resBaseDir>
-    <endTime1>09/07/15@12:25</endTime1>
-    <endTime2>1247653513.85</endTime2>
-    <RTTLibVersion>RunTimeTester-00-02-64</RTTLibVersion>
-    <resReleaseDir>/afs/cern.ch/user/r/rtt/test2/testsuites/simplemergejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline</resReleaseDir>
-    <startTime2>1247652686.48</startTime2>
-    <startTime1>09/07/15@12:11</startTime1>
-    <nPackages>1</nPackages>
-    <nPackagesAllJobsOK>1</nPackagesAllJobsOK>
-    <originalBranch>dev</originalBranch>
-    <nicosDate2>1228275841</nicosDate2>
-    <RTTSrcVersion>RunTimeTester-00-02-64</RTTSrcVersion>
-    <nicosDate1>2008/12/03 03:44:01</nicosDate1>
-    <release>14.5.0</release>
-  </overview>
-  <package>
-    <nPPSuccessInPackage>0</nPPSuccessInPackage>
-    <nTestsSuccessInPackage>0</nTestsSuccessInPackage>
-    <nJobsTotalPackage>1</nJobsTotalPackage>
-    <nJobsSuccessInPackage>0</nJobsSuccessInPackage>
-    <packageName>AthExHelloWorld</packageName>
-    <nJobsInPackage>1</nJobsInPackage>
-    <nJobsOpCloseInPackage>0</nJobsOpCloseInPackage>
-    <nTestsFailureInPackage>0</nTestsFailureInPackage>
-    <containerPackage>Control/AthenaExamples/AthExHelloWorld</containerPackage>
-    <nTimeOutsInPackage>0</nTimeOutsInPackage>
-    <phase>rttMainJobPhase</phase>
-    <packageTag>AthExHelloWorld-01-01-01</packageTag>
-    <nJobsDoneInPackage>1</nJobsDoneInPackage>
-    <nTestsInPackage>0</nTestsInPackage>
-    <nBatchErrorsInPackage>0</nBatchErrorsInPackage>
-    <nJobsFailureInPackage>0</nJobsFailureInPackage>
-    <nPPFailureInPackage>0</nPPFailureInPackage>
-    <nRetriesInPackage>0</nRetriesInPackage>
-    <packageStatus>done</packageStatus>
-    <SequentialMinder>
-      <status>done</status>
-      <nActiveMinders>0</nActiveMinders>
-      <jobID>0</jobID>
-      <state>done</state>
-      <jobGroup>ContainerGroup</jobGroup>
-      <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('done', '09/07/15 12:25')]</stateHistory>
-      <identifiedName>Sequential0</identifiedName>
-      <nContainedMinders>2</nContainedMinders>
-      <ChainJobMinder>
-        <status>done</status>
-        <nActiveMinders>0</nActiveMinders>
-        <jobID>6</jobID>
-        <state>done</state>
-        <jobGroup>ContainerGroup</jobGroup>
-        <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:21'), ('finishing', '09/07/15 12:24'), ('done', '09/07/15 12:24')]</stateHistory>
-        <identifiedName>Job6</identifiedName>
-        <nContainedMinders>1</nContainedMinders>
-        <minder>
-          <nUserTests>0</nUserTests>
-          <jobDocURL>
-          </jobDocURL>
-          <resultsPath>/afs/cern.ch/user/r/rtt/test2/testsuites/simplemergejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7</resultsPath>
-          <batchStatus>DONE</batchStatus>
-          <jobDisplayName>
-          </jobDisplayName>
-          <cpuTime2000>11</cpuTime2000>
-          <jobID>7</jobID>
-          <chainLength>1</chainLength>
-          <state>done</state>
-          <postProcessingResult>no tests</postProcessingResult>
-          <jobGroup>JobTransform</jobGroup>
-          <identifiedName>SimpleMergeJob07</identifiedName>
-          <status>success</status>
-          <chainSuccesses>1</chainSuccesses>
-          <mem>Unknown</mem>
-          <stateHistory>[('queued', '09/07/15 12:21'), ('running', '09/07/15 12:21'), ('postProcessingQueued', '09/07/15 12:22'), ('postProcessingRunning', '09/07/15 12:23'), ('done', '09/07/15 12:24')]</stateHistory>
-          <jobDocString>MyJob</jobDocString>
-          <hashString>-1946220797</hashString>
-          <nRetries>0</nRetries>
-          <cpuTime>4</cpuTime>
-          <jobName>SimpleMergeJob0</jobName>
-          <vmem>Unknown</vmem>
-          <nUserActions>2</nUserActions>
-          <exitStatus>Unknown</exitStatus>
-          <processingResult>success</processingResult>
-          <wallTime>Unobtainable</wallTime>
-          <datasets>
-            <datasetinfo>
-              <file>simple1.data</file>
-              <directory>/afs/cern.ch/user/r/rtt/test2/data/simplemergejob</directory>
-              <mtime>Unknown</mtime>
-              <size>Unknown</size>
-              <origin>fallback</origin>
-            </datasetinfo>
-            <datasetinfo>
-              <file>simple2.data</file>
-              <directory>/afs/cern.ch/user/r/rtt/test2/data/simplemergejob</directory>
-              <mtime>Unknown</mtime>
-              <size>Unknown</size>
-              <origin>fallback</origin>
-            </datasetinfo>
-          </datasets>
-          <missingDatasets>
-            <missingDataset>simple1.data</missingDataset>
-            <missingDataset>simple2.data</missingDataset>
-          </missingDatasets>
-          <archivers/>
-          <sUserTests>0</sUserTests>
-          <deliveredkeepfiles>
-            <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7</baseKeepFileDir>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/simple.merged.data</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/RTTtests.db</keepfile_newstyle>
-            <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/package.xml</keepfile_newstyle>
-            <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/runScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postprocessing.error.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postProcessor.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postprocessing.info.log</keepfile_newstyle>
-            <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/env.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postProcessorScript.sh</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postprocessing.debug.log</keepfile_newstyle>
-            <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/postprocessing.warning.log</keepfile_newstyle>
-            <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleMergeJob0/7/SimpleMergeJob07_log</keepfile_newstyle>
-          </deliveredkeepfiles>
-          <errors/>
-          <done/>
-          <checkResults/>
-          <testResults/>
-        </minder>
-      </ChainJobMinder>
-      <ParallelMinder>
-        <status>done</status>
-        <nActiveMinders>0</nActiveMinders>
-        <jobID>1</jobID>
-        <state>done</state>
-        <jobGroup>ContainerGroup</jobGroup>
-        <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('done', '09/07/15 12:21')]</stateHistory>
-        <identifiedName>Parallel1</identifiedName>
-        <nContainedMinders>2</nContainedMinders>
-        <ChainJobMinder>
-          <status>done</status>
-          <nActiveMinders>0</nActiveMinders>
-          <jobID>4</jobID>
-          <state>done</state>
-          <jobGroup>ContainerGroup</jobGroup>
-          <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('finishing', '09/07/15 12:21'), ('done', '09/07/15 12:21')]</stateHistory>
-          <identifiedName>Job4</identifiedName>
-          <nContainedMinders>1</nContainedMinders>
-          <minder>
-            <nUserTests>0</nUserTests>
-            <jobDocURL>
-            </jobDocURL>
-            <resultsPath>/afs/cern.ch/user/r/rtt/test2/testsuites/simplemergejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5</resultsPath>
-            <batchStatus>DONE</batchStatus>
-            <jobDisplayName>
-            </jobDisplayName>
-            <cpuTime2000>59</cpuTime2000>
-            <jobID>5</jobID>
-            <chainLength>1</chainLength>
-            <state>done</state>
-            <postProcessingResult>no tests</postProcessingResult>
-            <jobGroup>JobTransform</jobGroup>
-            <identifiedName>SimpleJob25</identifiedName>
-            <status>success</status>
-            <chainSuccesses>1</chainSuccesses>
-            <mem>1MB</mem>
-            <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('postProcessingQueued', '09/07/15 12:20'), ('postProcessingRunning', '09/07/15 12:20'), ('done', '09/07/15 12:21')]</stateHistory>
-            <jobDocString>MyJob</jobDocString>
-            <hashString>-2041773479</hashString>
-            <nRetries>0</nRetries>
-            <cpuTime>22</cpuTime>
-            <jobName>SimpleJob2</jobName>
-            <vmem>104MB</vmem>
-            <nUserActions>2</nUserActions>
-            <exitStatus>Unknown</exitStatus>
-            <processingResult>success</processingResult>
-            <wallTime>360</wallTime>
-            <datasets/>
-            <missingDatasets/>
-            <archivers/>
-            <sUserTests>0</sUserTests>
-            <deliveredkeepfiles>
-              <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5</baseKeepFileDir>
-              <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/env.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postProcessor.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postProcessorScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/runScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/RTTtests.db</keepfile_newstyle>
-              <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/package.xml</keepfile_newstyle>
-              <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/SimpleJob25_log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postprocessing.info.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postprocessing.debug.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postprocessing.warning.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob2/5/postprocessing.error.log</keepfile_newstyle>
-            </deliveredkeepfiles>
-            <errors/>
-            <done/>
-            <checkResults/>
-            <testResults/>
-          </minder>
-        </ChainJobMinder>
-        <ChainJobMinder>
-          <status>done</status>
-          <nActiveMinders>0</nActiveMinders>
-          <jobID>2</jobID>
-          <state>done</state>
-          <jobGroup>ContainerGroup</jobGroup>
-          <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('finishing', '09/07/15 12:18'), ('done', '09/07/15 12:19')]</stateHistory>
-          <identifiedName>Job2</identifiedName>
-          <nContainedMinders>1</nContainedMinders>
-          <minder>
-            <nUserTests>0</nUserTests>
-            <jobDocURL>
-            </jobDocURL>
-            <resultsPath>/afs/cern.ch/user/r/rtt/test2/testsuites/simplemergejob/Results/14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3</resultsPath>
-            <batchStatus>DONE</batchStatus>
-            <jobDisplayName>
-            </jobDisplayName>
-            <cpuTime2000>40</cpuTime2000>
-            <jobID>3</jobID>
-            <chainLength>1</chainLength>
-            <state>done</state>
-            <postProcessingResult>no tests</postProcessingResult>
-            <jobGroup>JobTransform</jobGroup>
-            <identifiedName>SimpleJob13</identifiedName>
-            <status>success</status>
-            <chainSuccesses>1</chainSuccesses>
-            <mem>Unknown</mem>
-            <stateHistory>[('queued', '09/07/15 12:14'), ('running', '09/07/15 12:14'), ('postProcessingQueued', '09/07/15 12:16'), ('postProcessingRunning', '09/07/15 12:17'), ('done', '09/07/15 12:18')]</stateHistory>
-            <jobDocString>MyJob</jobDocString>
-            <hashString>-1267101525</hashString>
-            <nRetries>0</nRetries>
-            <cpuTime>15</cpuTime>
-            <jobName>SimpleJob1</jobName>
-            <vmem>Unknown</vmem>
-            <nUserActions>2</nUserActions>
-            <exitStatus>Unknown</exitStatus>
-            <processingResult>success</processingResult>
-            <wallTime>35</wallTime>
-            <datasets/>
-            <missingDatasets/>
-            <archivers/>
-            <sUserTests>0</sUserTests>
-            <deliveredkeepfiles>
-              <baseKeepFileDir>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3</baseKeepFileDir>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/RTTtests.db</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postProcessor.log</keepfile_newstyle>
-              <keepfile_newstyle info='The script submitted to the batch machine' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/runScript.sh</keepfile_newstyle>
-              <keepfile_newstyle info='Package XML test configuration file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/package.xml</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postprocessing.info.log</keepfile_newstyle>
-              <keepfile_newstyle info='Job Log file' md5sum='' displayColor='#cc3333'>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/SimpleJob13_log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postprocessing.debug.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postprocessing.warning.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postprocessing.error.log</keepfile_newstyle>
-              <keepfile_newstyle info='Shows all the environmental variables set' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/env.log</keepfile_newstyle>
-              <keepfile_newstyle info='' md5sum='' displayColor=''>14.5.0/build/i686-slc4-gcc34-opt/offline/AthExHelloWorld/JobTransform/SimpleJob1/3/postProcessorScript.sh</keepfile_newstyle>
-            </deliveredkeepfiles>
-            <errors/>
-            <done/>
-            <checkResults/>
-            <testResults/>
-          </minder>
-        </ChainJobMinder>
-      </ParallelMinder>
-    </SequentialMinder>
-  </package>
-</RTTSummary>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/TestCase.py b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/TestCase.py
deleted file mode 100644
index 4b2ef6206ace7457199776122052ea3de62bd759..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/TestCase.py
+++ /dev/null
@@ -1,26 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from RTTTestCase import RTTTestCase
-import unittest
-import os
-class TestCase(RTTTestCase):
-
-    def variableContentTags(self):
-        """
-        Add the hashstring tag to the variable content
-        tag list. This job takes files from the
-        chain store, so the content of the file tag depends on
-        the (variable) name of the chain store.
-        """
-
-        tags = RTTTestCase.variableContentTags(self)
-        tags.extend(['directory','hashString', 'mtime'])
-        return tags
-    
-    def testCMLComp(self): self.rttTestXMLComp()
-
-    def testKeepfileDiff(self): self.rttKeepfileDiff()
-        
-if __name__ == '__main__':
-    from runTestCase import runTestCase
-    runTestCase(TestCase)
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/__init__.py b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/__init__.py
deleted file mode 100644
index 65210b009e66cb99c28e00c6a4d6bec83e62bd9f..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/__init__.py
+++ /dev/null
@@ -1,10 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-'''Short testsuite description goes here'''
-
-__length__ = "short" # < 10 mins
-
-
-def run():
-   "Any initialisation code goes in this method. Is run before RTT is launched."
-   pass
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/cfg/cfg.xml.TPL b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/cfg/cfg.xml.TPL
deleted file mode 100644
index a1ffab7988328d1c6c0636b3b9a881310f96940e..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/cfg/cfg.xml.TPL
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0"?>
-
-<!DOCTYPE rttconfig SYSTEM "file:$_BASE_$/DTD/RTT.dtd">
-
-<rttconfig>
-   <mode>LSFBatch</mode>
-   <release>14.5.0</release>
-   <refRelease>14.0.10</refRelease>
-   <cmtConfig>i686-slc4-gcc34-opt</cmtConfig>
-   <branch>dev</branch>
-   <runType><releaseRun/></runType>
-   <topProject>AtlasProduction</topProject>
-   <distArea>/afs/cern.ch/atlas/software/builds</distArea>
-   <workBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Work</workBasePath>
-   <resultsBasePath>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBasePath>
-   <resultsBaseURL>$_BASE_$/testsuites/$_TESTNAME_$/Results</resultsBaseURL>
-   <site>CERNslc4</site>
-   <dataSetCatalog>$_BASE_$/xml/rttDataSetCatalog.xml</dataSetCatalog>
-   <jobGroupConfig>$_BASE_$/xml/TopLevelJobGroup.xml</jobGroupConfig>
-   <dCubeCfgFile>$_BASE_$/DCubeClient-00-00-21/python/DCubeForRTTCfg.xml</dCubeCfgFile>
-
-   <confFileSrc>
-
-      <rttProvides>
-
-         <packageConfFile>
-            <package>Control/AthenaExamples/AthExHelloWorld</package>
-            <pathToFile>$_BASE_$/testsuites/$_TESTNAME_$/Package/package.xml</pathToFile>
-            <cmtpath>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0</cmtpath>
-            <version>AthExHelloWorld-01-01-01</version>
-            <installArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea</installArea>
-            <shareArea>/afs/cern.ch/atlas/software/builds/AtlasCore/14.5.0/InstallArea/share</shareArea>
-         </packageConfFile>
-
-      </rttProvides>
-
-   </confFileSrc>
-      
-   <packageAccept>AthExHelloWorld</packageAccept>
-      
-   <libToolsLoc>$_LIB_TOOLS_FROM_ENV_$</libToolsLoc>
-      
-   <releaseType>project</releaseType>
-   
-   <disableMailTo />
-   <devRun />
-   <deleteNonKeepFiles/>
-   <chainStoreBasePath>$_BASE_$/testsuites/$_TESTNAME_$/ChainStore</chainStoreBasePath>
-      
-</rttconfig>
diff --git a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/tests.py b/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/tests.py
deleted file mode 100644
index b283c36df2f782cd729addefab389566da39e388..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/testsuites/simplemergejob/tests.py
+++ /dev/null
@@ -1,30 +0,0 @@
-# Copyright (C) 2002-2017 CERN for the benefit of the ATLAS collaboration
-
-from testsuitetest import Test, SummaryFileTest
-import os.path
-
-def run():
-    for klass in [XMLFileTest]:
-        yield klass
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
-class XMLFileTest(SummaryFileTest):
-    def __init__(self, tester):
-        SummaryFileTest.__init__(self, tester)
-
-    def testHasExactlyOnePackage(self):
-        return self.hasExactlyNPackages(1)
-
-    def testHasPackageCalled(self):
-	return self.hasOnlyOnePackageCalled('AthExHelloWorld')
-
-    def testHasThreeJobsTotal(self):
-        return self.hasNJobs(3)
-
-# ==========================================================================================
-# ==========================================================================================
-# ==========================================================================================
-
diff --git a/Tools/RunTimeTester/testsuite/web/run.php b/Tools/RunTimeTester/testsuite/web/run.php
deleted file mode 100644
index 5c1e6826285cda4bfe31f1738bbd6d947b4ce952..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/web/run.php
+++ /dev/null
@@ -1,50 +0,0 @@
-<html>
-<head>
-<meta http-equiv="refresh" content="20" />
-
-<title>Framework tests: overview of current suites</title>
-<style>
-body{font-family:tahoma,sans-serif;font-size:1em;color:black;}
-tr.header{background-color:#8866ff;font-weight:bold;}
-tr.entry{background-color:#ddf;}
-</style>
-</head>
-<body>
-<?
-function getRootNodeFromXMLfile($theFile){
-  $xmlDoc = domxml_open_file($theFile);
-  $rootNode = $xmlDoc->document_element();
-  $xmlDoc->unlink();
-  unset($xmlDoc);
-  return $rootNode;
-}
-
-function getTagContent($parentNode, $tagName){
-  $thing = $parentNode->get_elements_by_tagname($tagName);
-  if (sizeOf($thing)==0) return "";
-  $thing = $thing[0]->get_content();
-  return trim($thing);
-}
-
-$current = getRootNodeFromXMLfile('current.testsuites.xml');
-$testsuites = $current->get_elements_by_tagname('testsuite');
-
-if (sizeOf($testsuites)==0){echo "No testsuite runs ongoing."; return;}
-?>
-
-<table>
-  <tr class="header"><td>Testsuite Name</td><td>Status</td><td>Testsuite location</td><td></td></tr>
-<?
-foreach($testsuites as $testsuite){
-  $name = getTagContent($testsuite, 'testsuitename');
-  $path = getTagContent($testsuite, 'testsuitepath');
-  $status = getTagContent($testsuite, 'testsuitestatus');
-  echo "<tr class='entry'><td>".$name."</td><td>".$status."</td><td>".$path."</td><td><a href='testsuite.php?name=".$name."'>Show me</a></td></tr>";
-  echo "";
-}
-
-?>
-
-</table>
-</body>
-</html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/testsuite/web/testsuite.php b/Tools/RunTimeTester/testsuite/web/testsuite.php
deleted file mode 100644
index f7d3baef1bbd8a525feaa93843372d95ba401c02..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/web/testsuite.php
+++ /dev/null
@@ -1,68 +0,0 @@
-<html>
-<head>
-<style>
-body{font-family:tahoma,sans-serif;font-size:1em;color:black;text-align:center;background-color:#fff}
-div.pageheader{text-align:center;font-weight:bold;}
-tr.tableheader{font-weight:bold;}
-tr.testname{background-color:#aaa}
-td.fail{background-color:red;}
-td.pass{background-color:#44aa33;}
-</style>
-<title></title>
-<body>
-<?
-function getRootNodeFromXMLfile($theFile){
-  $xmlDoc = domxml_open_file($theFile);
-  $rootNode = $xmlDoc->document_element();
-  $xmlDoc->unlink();
-  unset($xmlDoc);
-  return $rootNode;
-}
-
-function getTagContent($parentNode, $tagName){
-  $thing = $parentNode->get_elements_by_tagname($tagName);
-  if (sizeOf($thing)==0) return "";
-  $thing = $thing[0]->get_content();
-  return trim($thing);
-}
-
-$testsuitename = trim($_GET['name']);
-$pathToTestsuiteFile = '../testsuites/'.$testsuitename.'/testsuite.status.xml';
-$testsuite = getRootNodeFromXMLfile($pathToTestsuiteFile);
-
-$name   = getTagContent($testsuite, 'testsuitename');
-$status = getTagContent($testsuite, 'testsuitestatus');
-
-echo "<div class='pageheader'>".$testsuitename."</div><br />";
-
-$pathToSummaryFile = getTagContent($testsuite, 'pathToSummaryFile');
-if ($pathToSummaryFile==""){
-  echo "No job to show yet";
-  return;
-}
-
-
-echo "<a href='../testsuites/".$testsuitename."/Results/page1.php?xml=".$pathToSummaryFile."'>Show me the job(s)</a><br /><br />";
-
-echo "Tests:<br />";
-echo "<table align='center' border='1'>";
-foreach($testsuite->get_elements_by_tagname('testsuitetest') as $testsuitetest){
-  $testname = getTagContent($testsuitetest, 'testsuitetestname');
-  echo "<tr class='testname'><td colspan='3'>".$testname."</tr>";
-    foreach($testsuitetest->get_elements_by_tagname('test') as $test){
-    $testname    = getTagContent($test, 'testname');
-    $teststatus  = getTagContent($test, 'status');
-    $teststatus  = $teststatus?'PASS':'FAIL';
-    $statusCSS   = $teststatus=='PASS'?'pass':'fail';
-    $testmessage = getTagContent($test, 'message');
-    if (sizeOf($testmessage)){$testmessage='--';}
-
-    echo "<tr><td>".$testname."</td><td class='".$statusCSS."'>".$teststatus."</td><td>".$testmessage."</td></tr>";
-  }
-}
-
-echo "</table>";
-?>
-
-</body>
-</html>
\ No newline at end of file
diff --git a/Tools/RunTimeTester/testsuite/xml/TopLevelJobGroup.xml.TPL b/Tools/RunTimeTester/testsuite/xml/TopLevelJobGroup.xml.TPL
deleted file mode 100644
index e5e5f2bf3b2589af63f124ed1e2ada181e71c027..0000000000000000000000000000000000000000
--- a/Tools/RunTimeTester/testsuite/xml/TopLevelJobGroup.xml.TPL
+++ /dev/null
@@ -1,114 +0,0 @@
-<?xml version="1.0"?>
-<!DOCTYPE unifiedTestConfiguration SYSTEM "file:$_BASE_$/DTD/unifiedTestConfiguration.dtd">
-                                                                                                                                                            
-<unifiedTestConfiguration>
-   <atn/>
-   <kv/>
-   <rtt xmlns="http://www.hep.ucl.ac.uk/atlas/AtlasTesting/DTD/unifiedTestConfigurationFile">
-
-      <jobGroups>
-
-         <jobGroup name="Top" parent="">
-         
-            <keepFilePattern info="Shows all the environmental variables set">env.log</keepFilePattern>
-
-            <keepFilePattern info="cmt show uses command on package cmt dir">cmtShowUses.log</keepFilePattern>
-
-            <keepFilePattern info="ls -altF command on run directory before and after athena command">lsaltF.txt</keepFilePattern>
-
-            <keepFilePattern info="Gives a concise report of any problems encountered, checks failed, etc.">failureReport.html</keepFilePattern>
-
-            <keepFilePattern info="The script submitted to the batch machine">runScript.sh</keepFilePattern>
-
-            <keepFilePattern info="All ROOT macro output from your job goes in here">ROOToutput.log</keepFilePattern>
-
-            <keepFilePattern info="Exceptions raised by your Python script go in here">Python_Script_Output.log</keepFilePattern>
-
-            <keepFilePattern info="WARNING,ERROR and FATAL messages from job log">ERROR.log</keepFilePattern>
-
-            <keepFilePattern info="Output from checkFile.py">*.pool.root.checkFile</keepFilePattern>
-
-            <keepFilePattern info="Marker emitted at the end of the job">JobCompleteMarker.txt</keepFilePattern>
-
-            <keepFilePattern info="">DQWebDisplay.py.stdout</keepFilePattern>
-
-            <keepFilePattern info="">*.php</keepFilePattern>
-
-            <keepFilePattern info="">*dcube*</keepFilePattern>
-
-            <keepFilePattern info="">perfMonDozer_*.png</keepFilePattern>
-
-            <keepFilePattern info="">postprocessing.*.log</keepFilePattern>
-
-            <keepFilePattern info="">RTTtests.db</keepFilePattern>
-            <keepFilePattern info="">postProcessorScript.sh</keepFilePattern>
-            <keepFilePattern info="">postProcessor.log</keepFilePattern>
-
-            <keepFilePattern info="">DCube-*/*</keepFilePattern>
-            <keepFilePattern info="">DCube-*/plots/*.png</keepFilePattern>
-
-            <action position="zzzzzzzzzzz">
-               <modulename>RttLibraryTools</modulename>
-               <testname>CheckFileRunner</testname>
-            </action>
-
-            <action position="zzzzzzzzzzzz" runInAtlasEnv="True">
-               <modulename>RttLibraryTools</modulename>
-               <testname>PerfMonDozer</testname>
-            </action>
-
-
-         </jobGroup>
-      
-         <jobGroup name="Athena" parent="Top">
-             
-              <action>
-               <modulename>RttLibraryTools</modulename>
-               <testname>FileGrepper</testname>
-	       <arg>
-                <argname>inputFile</argname>
-                <argvalue>*_log</argvalue>
-               </arg>
-	      <arg>
-                <argname>outputFile</argname>
-                <argvalue>ERROR.log</argvalue>
-              </arg>
-	      <arg>
-                <argname>searchList</argname>
-                <argvalue>^(?! \-\+\-).*(WARN|ERROR|FATAL)</argvalue>
-              </arg>
-           </action>
-         
-            <logChecker class="SimpleLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <searchString>RTT determined exit status to be:  0</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-               <vetoString>job killed</vetoString>
-       
-              </logChecker>
-         </jobGroup>
-
-         <jobGroup name="AthenaATN" parent="Athena">
-             <logChecker class="ChainedJobsLogChecker">
-               <searchString>INFO Application Manager Finalized successfully</searchString>
-               <searchString>RTT determined exit status to be:  0</searchString>
-               <vetoString>ERROR Terminating event processing loop due to errors</vetoString>
-               <vetoString>FATAL Invalid initial state</vetoString>
-            </logChecker>
-         </jobGroup>
-
-         <jobGroup name="ErrorGroup" parent="Top" />
-         <jobGroup name="ContainerGroup" parent="Top" />
-
-         <jobGroup name="Transform" parent="Top">
-             <logChecker class="SimpleLogChecker"> 
-                <searchString>----- JobTransform shell script starts  ---------</searchString>
-                <searchString>----- JobTransform shell script end  ---------</searchString>
-                <searchString>ErrorCode=0 (OK)</searchString>
-             </logChecker>
-         </jobGroup>
-
-      </jobGroups>
-   </rtt>
-</unifiedTestConfiguration>