Skip to content

Resolve "create mini-analysis script to clean duplicate events"

Release notes

  • added cafDuplicateCleaner.py which performs a miminal "analysis" using the xAOD skimming feature to remove duplicate events from data xAODs. This needs to process all data samples that might have duplicate events (typically all of one processing campaign, e.g., 2015+16, 2017, 2018) in one run in order to also be able to find (and remove) duplicates across individual files. Usage: cafDuplicateChecker.py --output dirToWriteCleanedFilesTo input1.root input2.root --- (hint: use your shell's expansion of wildcards, i.e., input*.root). For first run use additional option --checkOnly to not perform the cleaning but only check for duplicates (much faster!)
  • fixed various issues regarding TQUniqueCut, mostly related to its structure being largely outdated before this revamp

Details

duplicate cleaner was tested by explicitly copying one input file (as to force duplicates to occur across file boundaries), working fine after various fixes to TQUniqueCut (for main fixes see #257 (closed))

Closes #256 (closed) and #257 (closed)

Edited by Ralf Gugel

Merge request reports