ALPGEN Production in 31X

Goals

  • Test the Alpgen Prouduction Workflow in CMSSW 3_1_X

Modus Operandi

  1. Creation of ASCII LHE files outside the CMS framework
  2. Publication of corresponding articles into MCDB (http://mcdb.cern.ch)
  3. Execution of CMSSW jobs (at CERN)for retrieving the files, translating the ASCII LHE files into EDM LHE files and publishing them to DBS
  4. Deployment of the files to the T2s for production.

  • The responsabilty of the Generator group is for steps 1 and 2. DMWM will execute the jobs for accomplishing 3 and 4.
  • Preferred file size from DataOps: 1-2 GB, ~ 1M events in the same LHE file.

Log

  • Try to generate large (1M) LHE fils and upload to MCDB.
    • 50 jobs of 2000000 events each, ttbar+1 jet, 40 GeV, should produce some 900K ttbar+1j unw events. Submitted to Condor.
    • Each job takes in average 3min 15sec!
    • Efficiency: 0.91% (average 18200 events obtained per job)
    • If I just concatenate the .unw files, AlpgenSource does not care... does it? No it doesn't =D

-- ThiagoTomei - 18 May 2009

  • Questions:
  1. What to do with the cross-sections and the luminosity, when concatening different files?
    • Sum the luminosities, average the cross-sections, average/sqrt(N) the errors?
  2. What is the workflow? For minimum startup conditions:
    1. ALPGEN Standalone + Condor --> generate many .wgt files, unweight them to many .unw files. Use the 1st script below for this.
    2. Script to concatenate all .unw files, write a single, unified _unw.par file, and write a .tar.gz file containing ALL the .wgt and .par files. Use the 2nd script below for this.
    3. cmsRun jobs + LHEWriter --> generate a single, large .lhe file. Needs CMSSW 3_1_X
    4. Write the HepML header. Needs HepML working
    5. upload to MCDB --> there is a Perl script from http://mcdb.cern.ch for that, but does it work from outside CERN? It uses SRM to access CASTOR.
    6. Coach somebody to test the following parts (MCDBSource, split and merge).

-- ThiagoTomei - 19 May 2009

  1. UPDATE: managed to run the full chain up MCDB source (article 239).
    • But there is a problem, CMSSW_3_1_X is not yet installed at SPRACE.
  2. UPDATE: things that should be in the test suite:
    • Check if grids are correct (maybe not needed now, for we are using LHE files?)
    • Check if split and merge works.

-- ThiagoTomei - 22 May 2009

Results from Marco (AlpgenTimeAndSizeStudies):

  • Running time is linear with number of events requested.
  • Running time is exponential with number of extra jets.
  • In worst case (ttbar + 4jets), 10M events runs on ~ 1h30min in a Phase III node. This is a good turnaround time.
  • I think that planning the production in chunks of 10M events is reasonable.
    • ttbar + 4jets case: from AlpgenGridProductionPage, one can say that the unweighting efficiency is ~ 5E-04. That, compound with a matching efficiency of 0.15, means that to get 100K events in the end one should ask for ~ 1.35G events. That translates into 135 jobs, a reasonable number.
    • ttbar + 0jets case: from AlpgenGridProductionPage, one can say that the unweighting efficiency is ~ 2E-02, matching efficiency is ~ 0.7. To get 100K events in the end, one needs to ask for ~ 7.1M events. Would rather ask for 100M, divided in 10 jobs, and aim for 1M events in the end.
    • CAVEAT: for ttbar + 4jets, 100K events is plenty, but for ttbar + 0 jets it is almost nothing. The two factors (ease of producing X number needed) compensate to a degree.
  • What would I quote as ''full farm capacity'':
    • 116 Phase III Worker Nodes.
    • Worst case scenario: ttbar + 4jets, 6.6M asked events / hour --> 500 UNW events / hour.
    • Total: 58000 ttbar + 4jets events / hour.

-- ThiagoTomei - 05 Jun 2009

  • Some long needed UPDATES:
    • CMSSW 3_1_1 now installed at SPRACE.
    • Managed to get files to CASTOR via srmcp and then upload them via upload2mcdb.pl
  • The workflow as of now:
    1. ALPGEN Standalone + Condor --> generate many .wgt files, unweight them to many .unw files. Use the 1st script below for this.
    2. Script to concatenate all .unw files, write a single, unified _unw.par file, and write a .tar.gz file containing ALL the .wgt and .par files. Use the 2nd script below for this.
    3. cmsRun jobs + LHEWriter --> generate a single, large .lhe file.
    4. Write the HepML header. Needs HepML working
    5. upload to CASTOR via srmcp
    6. upload to MCDB --> there is a Perl script from http://mcdb.cern.ch for that, use it from lxplus.
  • Action itens
    1. Must automatize everything
    2. Must check how to write a HepML header from the _unw.par file.

-- ThiagoTomei - 21 Jul 2009

  • W+jets in the 0-100 pT bracket needs parallelism. Other brackets do not.

-- ThiagoTomei - 22 Jul 2009

  • Managed to get upload2mcdb.pl working from SPRACE - also with multiple TCP streams, so it can work faster.
  • I retract my previous comment about paralellism.

-- ThiagoTomei - 23 Jul 2009

-- ThiagoTomei - 11 Nov 2009

Topic attachments
I Attachment History Action Size Date Who Comment
Texttxt sprace_ALPGEN_merge.py.txt r1 manage 3.0 K 2009-06-05 - 08:07 ThiagoTomei Script to merge the files produced by the sprace_ALPGEN_submit.py script. It produces a merged .unw file, a coherent _unw.par file, and a zipped directory contating the .wgt and .par files.
Texttxt sprace_ALPGEN_submit.py.txt r1 manage 3.3 K 2009-06-03 - 15:24 ThiagoTomei Script to submit many copies of the same Alpgen generation job to CONDOR. Automatically takes care of random seeds. Useful wiwh seq | xargs command.
Topic revision: r14 - 2009-11-11 - ThiagoTomei
 

This site is powered by the TWiki collaboration platform Powered by PerlCopyright © 2008-2024 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki? Send feedback

antalya escort bursa escort eskisehir escort istanbul escort izmir escort