Skip to content

Commit

Permalink
Merge pull request #1003 from franzoni/fixes-reqmgrChanges-62x
Browse files Browse the repository at this point in the history
Fixes reqmgr changes 62x
  • Loading branch information
davidlt committed Oct 16, 2013
2 parents e7fd164 + 1409256 commit dca3049
Showing 1 changed file with 13 additions and 15 deletions.
28 changes: 13 additions & 15 deletions Configuration/PyReleaseValidation/python/MatrixInjector.py
Expand Up @@ -80,8 +80,6 @@ def __init__(self,opt,mode='init',options=''):

self.defaultChain={
"RequestType" : "TaskChain", #this is how we handle relvals
"AcquisitionEra": {}, #Acq Era
"ProcessingString": {}, # processing string to label the dataset
"Requestor": self.user, #Person responsible
"Group": self.group, #group for the request
"CMSSWVersion": os.getenv('CMSSW_VERSION'), #CMSSW Version (used for all tasks in chain)
Expand All @@ -94,7 +92,7 @@ def __init__(self,opt,mode='init',options=''):
"DbsUrl": "http://cmsdbsprod.cern.ch/cms_dbs_prod_global/servlet/DBSServlet",
#"CouchDBName": self.couchDB, #Name of Couch Database containing config cache
#- Will contain all configs for all Tasks
"SiteWhitelist" : ["T2_CH_CERN", "T1_US_FNAL"], #Site whitelist
#"SiteWhitelist" : ["T2_CH_CERN", "T1_US_FNAL"], #Site whitelist
"TaskChain" : None, #Define number of tasks in chain.
"nowmTasklist" : [], #a list of tasks as we put them in
"unmergedLFNBase" : "/store/unmerged",
Expand All @@ -106,7 +104,7 @@ def __init__(self,opt,mode='init',options=''):
}

self.defaultHarvest={
"EnableDQMHarvest" : 1,
"EnableHarvesting" : "True",
"DQMUploadUrl" : self.dqmgui,
"DQMConfigCacheID" : None
}
Expand All @@ -115,8 +113,8 @@ def __init__(self,opt,mode='init',options=''):
"TaskName" : None, #Task Name
"ConfigCacheID" : None, #Generator Config id
"GlobalTag": None,
"SplittingAlgorithm" : "EventBased", #Splitting Algorithm
"SplittingArguments" : {"events_per_job" : None}, #Size of jobs in terms of splitting algorithm
"SplittingAlgo" : "EventBased", #Splitting Algorithm
"EventsPerJob" : None, #Size of jobs in terms of splitting algorithm
"RequestNumEvents" : None, #Total number of events to generate
"Seeding" : "AutomaticSeeding", #Random seeding method
"PrimaryDataset" : None, #Primary Dataset to be created
Expand All @@ -128,8 +126,8 @@ def __init__(self,opt,mode='init',options=''):
"ConfigCacheID" : None, #Processing Config id
"GlobalTag": None,
"InputDataset" : None, #Input Dataset to be processed
"SplittingAlgorithm" : "LumiBased", #Splitting Algorithm
"SplittingArguments" : {"lumis_per_job" : 10}, #Size of jobs in terms of splitting algorithm
"SplittingAlgo" : "LumiBased", #Splitting Algorithm
"LumisPerJob" : 10, #Size of jobs in terms of splitting algorithm
"nowmIO": {},
"KeepOutput" : False
}
Expand All @@ -139,8 +137,8 @@ def __init__(self,opt,mode='init',options=''):
"InputFromOutputModule" : None, #OutputModule name in the input task that will provide files to process
"ConfigCacheID" : None, #Processing Config id
"GlobalTag": None,
"SplittingAlgorithm" : "LumiBased", #Splitting Algorithm
"SplittingArguments" : {"lumis_per_job" : 10}, #Size of jobs in terms of splitting algorithm
"SplittingAlgo" : "LumiBased", #Splitting Algorithm
"LumisPerJob" : 10, #Size of jobs in terms of splitting algorithm
"nowmIO": {},
"KeepOutput" : False
}
Expand Down Expand Up @@ -214,7 +212,7 @@ def prepare(self,mReader, directories, mode='init'):
arg=s[2][index].split()
ns=map(int,arg[arg.index('--relval')+1].split(','))
chainDict['nowmTasklist'][-1]['RequestNumEvents'] = ns[0]
chainDict['nowmTasklist'][-1]['SplittingArguments']['events_per_job'] = ns[1]
chainDict['nowmTasklist'][-1]['EventsPerJob'] = ns[1]
if 'FASTSIM' in s[2][index] or '--fast' in s[2][index]:
thisLabel+='_FastSim'

Expand All @@ -227,9 +225,9 @@ def prepare(self,mReader, directories, mode='init'):
return -15
chainDict['nowmTasklist'][-1]['InputDataset']=nextHasDSInput.dataSet
splitForThisWf=nextHasDSInput.split
chainDict['nowmTasklist'][-1]['SplittingArguments']['lumis_per_job']=splitForThisWf
chainDict['nowmTasklist'][-1]['LumisPerJob']=splitForThisWf
if step in wmsplit:
chainDict['nowmTasklist'][-1]['SplittingArguments']['lumis_per_job']=wmsplit[step]
chainDict['nowmTasklist'][-1]['LumisPerJob']=wmsplit[step]
# get the run numbers or #events
if len(nextHasDSInput.run):
chainDict['nowmTasklist'][-1]['RunWhitelist']=nextHasDSInput.run
Expand All @@ -252,9 +250,9 @@ def prepare(self,mReader, directories, mode='init'):
print "Failed to find",'%s/%s.io'%(dir,step),".The workflows were probably not run on cfg not created"
return -15
if splitForThisWf:
chainDict['nowmTasklist'][-1]['SplittingArguments']['lumis_per_job']=splitForThisWf
chainDict['nowmTasklist'][-1]['LumisPerJob']=splitForThisWf
if step in wmsplit:
chainDict['nowmTasklist'][-1]['SplittingArguments']['lumis_per_job']=wmsplit[step]
chainDict['nowmTasklist'][-1]['LumisPerJob']=wmsplit[step]

#print step
chainDict['nowmTasklist'][-1]['TaskName']=step
Expand Down

0 comments on commit dca3049

Please sign in to comment.