Skip to content

Latest commit

 

History

History

cms-2016-simulated-datasets

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 

cms-2016-simulated-datasets

This directory contains helper scripts used to prepare CMS 2016 open data release regarding MC simulated datasets.

  • code/ folder contains the python code;
  • inputs/ folder contains input text files with the list of datasets for each year and input files;
  • outputs/ folder contains generated JSON records to be included as the CERN Open Data portal fixtures.

Every step necessary to produce the final *.json files is handled by the cmc-mc/interface.py script. Details about it can be queried with the command:

$ python3 code/interface.py --help

Please make sure to get the VOMS proxy file before running these scripts:

$ voms-proxy-init --voms cms --rfc --valid 190:00

Please make sure to set the EOS instance to EOSPUBLIC before running these scripts:

$ export EOS_MGM_URL=root://eospublic.cern.ch

Please make sure to have a valid userkey.nodes.pem certificate present in $HOME/.globus. If not, you have to run the following on top of the regular CMS certificate documentation:

$ cd $HOME/.globus
$ ls userkey.nodes.pem
$ openssl pkcs12 -in myCert.p12 -nocerts -nodes -out userkey.nodes.pem  # if not present
$ cd -

Warning: Creating the full local cache might take a long time.

First step is to create EOS file index cache:

$ python3 ./code/interface.py --create-eos-indexes inputs/CMS-2016-mc-datasets.txt

This requires the data files to be placed in their final location. However, for early testing on LXPLUS, all steps can be run without the EOS file index cache by means of adding the command-line option --ignore-eos-store to the commands below.

We can now build sample records by doing:

$ python3 ./code/interface.py --create-das-json-store --ignore-eos-store inputs/CMS-2016-mc-datasets.txt

$ auth-get-sso-cookie -u  https://cms-pdmv.cern.ch/mcm -o cookies.txt
$ python3 ./code/interface.py --create-mcm-store --ignore-eos-store inputs/CMS-2016-mc-datasets.txt

$ python3 ./code/interface.py --get-conf-files --ignore-eos-store inputs/CMS-2016-mc-datasets.txt

$ python3 code/lhe_generators.py

$ python3 ./code/interface.py --create-records --ignore-eos-store inputs/CMS-2016-mc-datasets.txt
$ python3 ./code/interface.py --create-conffiles-records --ignore-eos-store inputs/CMS-2016-mc-datasets.txt

Note that to build the test records an (empty) input file for DOIs and a recid info file must be present in the inputs directory.

Each step builds a subdirectory with a cache (das-json-store, mcm-store and config-store). They are large, do not upload them to the repository, respect the .gitignore.

The output JSON files for the dataset records will be generated in the outputs directory.

lhe_generators

$ python3 code/lhe_generators.py >& output
  • This will get lhe generator parameters from gridpacks for datasets listed in ./inputs/CMS-2016-mc-datasets.txt.
  • It works on LXPLUS or with mounted EOS.
  • Number of threads is set to 20 which is ideal for LXPLUS.

⚠️ There are many cases with various steps to get generator parameters for LHE -see #97-. Thus, in some few cases, the script MIGHT not work as expected so make sure to read it, check errors, and make any necessary tweaks