forked from lawrenceleejr/ZeroLeptonRun2
rsmith54/ZeroLeptonRun2
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
1/ Install and run ZeroLeptonRun2: ================================== The instructions for installing and compiling are in ZeroLeptonRun2/INSTALL The cafe config file is in ZeroLeptonRun2/config/zerolepton.config, you can copy it and modify it but for testing it is easier to just overwrite some configurations from the command line like: cafe ZeroLeptonRun2/config/zerolepton.config Events: 100 Input: root://eosatlas.cern.ch//eos/atlas/user/l/lduflot/atlasreadable/datafiles/mc14_13TeV.110401.PowhegPythia_P2012_ttbar_nonallhad.merge.DAOD_SUSY1.e2928_s1982_s2008_r5787_r5853_p1872/DAOD_SUSY1.05248705._000028.pool.root.1 This run a simple job with no systematics. A preliminary implementation of systematics that adds a tree for every systematics uses the zerolepton_with_syst.config config file. To run the truth code, an example is given below : cafe ZeroLeptonRun2/config/zeroleptontruth.config Events: 100 Input: root://eosatlas.cern.ch//eos/atlas/user/c/conti/atlasreadable/datafiles/mc15_13TeV.361007.Pythia8EvtGen_A14NNPDF23LO_jetjet_JZ7.evgen.EVNT.e3569_tid04904951_00/DAOD_TRUTH1.DAOD.04904951._000779.pool.root 2/ Extract dataset informations from AMI: ========================================= localSetupPyAMI ZeroLeptonRun2/python/makeDatasetList.py --help For example, ZeroLeptonRun2/python/makeDatasetList.py --baseline will make the list of mc14_13TeV corresponding to the baseline samples. 3/ Running jobs on the grid: ============================ #to reduce the size of the tarball it's better to clear libraries etc rc clean localSetupPandaClient ZeroLeptonRun2/python/launch_jobs_on_grid.py --help For example, with baseline.ds a text file with the list of dataset extracted with makeDatasetList.py: ZeroLeptonRun2/python/launch_jobs_on_grid.py --prunopts='--nGBPerJob=10 --excludedSite=ANALY_UIO,SIGNET,LUNARC ' --prefix='user.lduflot.' --suffix='_test2' --tmpDir=/scratch/duflot --inDSfile baseline.ds If the inDS contains the string data11/data12, the job will be configured to read real data. For MC, if the inDS still has the AMI tags, the script can figure out if ATLFAST was used. On the other hand, it knows about a limited pattern of signal datasets, to be safe use the --signal option. 4/ Creating an updated cross-section database: ============================================== This step is necessary to produce a file in the correct format for merging. Use the getMCInfo.py tools (slightly updated version of the Run1 tool) to build the extended cross-section database from the un-merged small ntuples. The tool will extract the number of events processed and the total sum of MC weights from the Counter and write an extended cross-section file that includes these informations. For example, one can list all downloaded small tuples in a file input.list (the file path should allow to extract the dataset IDs) an do ZeroLeptonRun2/python/getMCInfo.py --input input.list this will create MCBackgroundDB.dat (and .tex) in the local directory. You need to have the code compiled as it uses C++ code from SUSYTools. 5/ Merging small ntuples for ZeroLeptonHistFitter: ================================================== The filter/merge/update step is done by FilterUpdateMerge.py: ZeroLeptonRun2/python/FilterUpdateMerge.py --help For example: ZeroLeptonRun2/python/FilterUpdateMerge.py --doBackground --input_mc=small.list --doNormWeight --dbName updatedDB.dat Where the --dbName option points to an updated cross-section database built with the previous step (needed as the previous steps adds the number of events processed and corresponding sum weights. The file path of input files for MC should allow to retreive the dataset ID, so this is currently not possible for files in rucio. You need to have the code compiled as it uses C++ code.
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- C++ 88.1%
- Python 11.9%