00001 from GaudiPython import gbl, SUCCESS, FAILURE
00002 from multiprocessing import Event
00003 import pickle, time
00004
00005
00006
00007
00008
00009
00010
00011
00012
00013
00014
00015
00016
00017
00018
00019
00020
00021
00022
00023
00024
00025
00026
00027
00028
00029
00030
00031
00032
00033
00034
00035
00036
00037
00038
00039
00040
00041
00042
00043
00044
00045 aida2root = gbl.Gaudi.Utils.Aida2ROOT.aida2root
00046
00047
00048
00049 class HistoAgent( ) :
00050 def __init__( self, gmpComponent ) :
00051 self._gmpc = gmpComponent
00052 self.hvt = self._gmpc.hvt
00053 self.histos = []
00054 self.qin = self._gmpc.hq
00055 self.log = self._gmpc.log
00056
00057
00058
00059 self.bookingDict = {}
00060 self.bookingDict['DataObject'] = self.bookDataObject
00061 self.bookingDict['NTuple::Directory'] = self.bookDataObject
00062 self.bookingDict['NTuple::File'] = self.bookDataObject
00063 self.bookingDict['TH1D'] = self.bookTH1D
00064 self.bookingDict['TH2D'] = self.bookTH2D
00065 self.bookingDict['TH3D'] = self.bookTH3D
00066 self.bookingDict['TProfile'] = self.bookTProfile
00067 self.bookingDict['TProfile2D'] = self.bookTProfile2D
00068
00069 def register( self, tup ) :
00070
00071 assert tup.__class__.__name__ == 'tuple'
00072 self.histos.append( tup )
00073
00074 def Receive( self ) :
00075 hstatus = self._gmpc.nWorkers+1
00076 while True :
00077 tup = self.qin.get()
00078 if tup == 'HISTOS_SENT' :
00079 self.log.debug('received HISTOS_SENT message')
00080 hstatus -= 1
00081 if not hstatus : break
00082 else :
00083 self.register( tup )
00084 self._gmpc.sEvent.set()
00085 self.log.info('Writer received all histo bundles and set sync event')
00086 return SUCCESS
00087
00088
00089 def RebuildHistoStore( self ) :
00090 '''
00091 Rebuild the Histogram Store from the histos received by Receive()
00092 If we have a histo which is not in the store,
00093 book and fill it according to self.bookingDict
00094 If we have a histo with a matching histo in the store,
00095 add the two histos, remembering that aida2root must be used on
00096 the Stored histo for compatibility.
00097 '''
00098 errors = 0
00099 for tup in self.histos :
00100 workerID, histDict = tup
00101 added = 0 ; registered = 0; booked = 0
00102 for n in histDict.keys() :
00103 o = histDict[ n ]
00104 obj = self.hvt.retrieve( n )
00105 if obj :
00106 try :
00107 aida2root(obj).Add(o)
00108 except :
00109 self.log.warning('FAILED TO ADD : %s'%(str(obj)))
00110 errors += 1
00111 added += 1
00112 else :
00113 if o.__class__.__name__ in self.bookingDict.keys() :
00114 try :
00115 self.bookingDict[o.__class__.__name__](n, o)
00116 except :
00117 self.log.warning('FAILED TO REGISTER : %s\tto%s'\
00118 %(o.__class__.__name__, n))
00119 errors += 1
00120 else :
00121 self.log.warning( 'No booking method for: %s\t%s\t%s'\
00122 %(n,type(o),o.__class__.__name__) )
00123 errors += 1
00124 booked += 1
00125 hs = self.hvt.getHistoNames()
00126 self.log.info( 'Histo Store Rebuilt : ' )
00127 self.log.info( ' Contains %i objects.'%(len(hs)) )
00128 self.log.info( ' Errors in Rebuilding : %i'%(errors) )
00129 return SUCCESS
00130
00131
00132 def bookDataObject( self, n, o ):
00133 '''
00134 Register a DataObject to the Histo Store
00135 '''
00136 self._gmpc.hvt.registerObject( n, o )
00137
00138 def bookTH1D( self, n, o ) :
00139 '''
00140 Register a ROOT 1D THisto to the Histo Store
00141 '''
00142 obj = self.hvt._ihs.book( n, o.GetTitle(),\
00143 o.GetXaxis().GetNbins(),\
00144 o.GetXaxis().GetXmin(),\
00145 o.GetXaxis().GetXmax() )
00146 aida2root(obj).Add(o)
00147
00148 def bookTH2D( self, n, o ) :
00149 '''
00150 Register a ROOT 2D THisto to the Histo Store
00151 '''
00152 obj = self.hvt._ihs.book( n, o.GetTitle(),\
00153 o.GetXaxis().GetNbins(),\
00154 o.GetXaxis().GetXmin(),\
00155 o.GetXaxis().GetXmax(),\
00156 o.GetYaxis().GetNbins(),\
00157 o.GetYaxis().GetXmin(),\
00158 o.GetYaxis().GetXmax() )
00159 aida2root(obj).Add(o)
00160
00161 def bookTH3D( self, n, o ) :
00162 '''
00163 Register a ROOT 3D THisto to the Histo Store
00164 '''
00165 obj = self.hvt._ihs.book( n, o.GetTitle(),\
00166 o.GetXaxis().GetXbins(),\
00167 o.GetXaxis().GetXmin(),\
00168 o.GetXaxis().GetXmax(),\
00169 o.GetYaxis().GetXbins(),\
00170 o.GetYaxis().GetXmin(),\
00171 o.GetYaxis().GetXmax(),\
00172 o.GetZaxis().GetXbins(),\
00173 o.GetZaxis().GetXmin(),\
00174 o.GetZaxis().GetXmax() )
00175 aida2root(obj).Add(o)
00176
00177 def bookTProfile( self, n, o ) :
00178 '''
00179 Register a ROOT TProfile to the Histo Store
00180 '''
00181 obj = self.hvt._ihs.bookProf( n, o.GetTitle(),\
00182 o.GetXaxis().GetNbins(),\
00183 o.GetXaxis().GetXmin(),\
00184 o.GetXaxis().GetXmax(),\
00185 o.GetOption() )
00186 aida2root(obj).Add(o)
00187
00188 def bookTProfile2D( self, n, o ) :
00189 '''
00190 Register a ROOT TProfile2D to the Histo Store
00191 '''
00192 obj = self.hvt._ihs.bookProf( n, o.GetTitle(),\
00193 o.GetXaxis().GetNbins(),\
00194 o.GetXaxis().GetXmin(),\
00195 o.GetXaxis().GetXmax(),\
00196 o.GetYaxis().GetNbins(),\
00197 o.GetYaxis().GetXmin(),\
00198 o.GetYaxis().GetXmax() )
00199 aida2root(obj).Add(o)
00200
00201
00202
00203 class FileRecordsAgent( ) :
00204 def __init__( self, gmpComponent ) :
00205 self._gmpc = gmpComponent
00206 self.fsr = self._gmpc.fsr
00207 self.q = self._gmpc.fq
00208 self.log = self._gmpc.log
00209 self.objectsIn = []
00210 self.objectsOut = []
00211
00212 def localCmp( self, tupA, tupB ) :
00213
00214
00215 ind = 0
00216 valA = tupA[ind]
00217 valB = tupB[ind]
00218 if valA<valB : return -1
00219 elif valA>valB : return 1
00220 else : return 0
00221
00222
00223 def SendFileRecords( self ) :
00224
00225
00226
00227
00228
00229
00230
00231
00232
00233
00234
00235 self.log.info('Sending FileRecords...')
00236 lst = self.fsr.getHistoNames()
00237
00238
00239 if not lst :
00240 self.log.info('No FileRecords Data to send to Writer.')
00241 self.q.put( 'END_FSR' )
00242 return SUCCESS
00243
00244
00245 if '/FileRecords' in lst : lst.remove('/FileRecords')
00246
00247 for l in lst :
00248 o = self.fsr.retrieveObject( l )
00249 if hasattr(o, "configureDirectAccess") :
00250 o.configureDirectAccess()
00251
00252 if self._gmpc.nodeID == 0 :
00253 self.objectsOut.append( (0, l, pickle.dumps(o)) )
00254 else :
00255
00256
00257 if l == '/FileRecords/EventCountFSR' :
00258 tup = (self._gmpc.nodeID, l, pickle.dumps(o))
00259 self.objectsOut.append( tup )
00260 else :
00261
00262 assert "KeyedContainer" in o.__class__.__name__
00263 nObjects = o.numberOfObjects()
00264 if nObjects :
00265 self.log.debug("Keyed Container %s with %i objects"\
00266 %(l, nObjects))
00267 tup = (self._gmpc.nodeID, l, pickle.dumps(o))
00268 self.objectsOut.append( tup )
00269 self.log.debug('Done with FSR store, just to send to Writer.')
00270
00271 if self.objectsOut :
00272 self.log.debug('%i FSR objects to Writer'%(len(self.objectsOut)))
00273 for ob in self.objectsOut :
00274 self.log.debug('\t%s'%(ob[0]))
00275 self.q.put( self.objectsOut )
00276 else :
00277 self.log.info('Valid FSR Store, but no data to send to Writer')
00278 self.log.info('SendFSR complete')
00279 self.q.put( 'END_FSR' )
00280 return SUCCESS
00281
00282 def Receive( self ) :
00283
00284 self.log.info('Receiving FSR store data...')
00285 nc = self._gmpc.nWorkers
00286 while nc > 0 :
00287 objects = self.q.get( )
00288 if objects == 'END_FSR' :
00289 nc -= 1
00290 continue
00291 if nc==0 :
00292 break
00293
00294 for o in objects :
00295 self.objectsIn.append(o)
00296
00297
00298 self.objectsIn.sort(cmp=self.localCmp)
00299 self.log.info('All FSR data received')
00300 return SUCCESS
00301
00302 def Rebuild( self ) :
00303
00304 for sourceNode, path, serialob in self.objectsIn :
00305 self.log.debug('Working with %s'%(path))
00306 ob = pickle.loads(serialob)
00307 if hasattr( ob, "configureDirectAccess" ) :
00308 ob.configureDirectAccess()
00309 nCont = ob.numberOfObjects()
00310 self.log.debug( '\tcontainedObjects : %i'%(nCont) )
00311 if sourceNode == 0 :
00312 self.log.debug('Registering Object to : %s'%(path))
00313 self.fsr.registerObject( path, ob )
00314 else :
00315 self.log.debug('Merging Object to : %s'%(path))
00316 self.MergeFSRobject( sourceNode, path, ob )
00317
00318
00319
00320
00321 self.log.info('FSR Store Rebuilt. Correcting EventCountFSR')
00322 if bool( self.fsr._idp ) :
00323 ecount = '/FileRecords/EventCountFSR'
00324 if self.fsr[ecount] :
00325 self.fsr[ecount].setOutput( self._gmpc.nIn )
00326 self.log.info( 'Event Counter Output set : %s : %i'\
00327 %(ecount, self.fsr[ecount].output()) )
00328
00329 self.log.debug('FSR store reconstructed!')
00330 lst = self.fsr.getHistoNames()
00331 if lst :
00332 for l in lst :
00333 ob = self.fsr.retrieveObject(l)
00334 if hasattr( ob, 'configureDirectAccess' ) :
00335 ob.configureDirectAccess()
00336 if hasattr( ob, 'containedObjects' ) :
00337
00338 self.log.debug('\t%s (cont. objects : %i)'\
00339 %(l, ob.numberOfObjects()))
00340 else :
00341 self.log.debug('\t%s'%(l))
00342 self.log.info('FSR Store fully rebuilt.')
00343 return SUCCESS
00344
00345 def MergeFSRobject( self, sourceNode, path, ob ) :
00346
00347 if path == '/FileRecords/TimeSpanFSR' :
00348
00349 self.ProcessTimeSpanFSR( path, ob )
00350 elif path == '/FileRecords/EventCountFSR' :
00351
00352 self.ProcessEventCountFSR( path, ob )
00353
00354 elif "KeyedContainer" in ob.__class__.__name__ :
00355
00356
00357 if "LumiFSR" in ob.__class__.__name__ :
00358 self.MergeLumiFSR( path, ob )
00359 else:
00360 self.log.info("Skipping Merge of Keyed Container %s for %s"\
00361 %(ob.__class__.__name__,path))
00362
00363 def ProcessTimeSpanFSR( self, path, ob ) :
00364 if ob.containedObjects().size() :
00365 sz = ob.containedObjects().size()
00366 for j in xrange(sz) :
00367 cob = ob.containedObjects()[j]
00368 self.log.debug('Adding TimeSpanFSR')
00369
00370 tsfsr = gbl.LHCb.TimeSpanFSR()
00371 tsfsr.setEarliest( cob.earliest() )
00372 tsfsr.setLatest( cob.latest( ) )
00373 self.fsr[path].add(tsfsr)
00374 continue
00375
00376 def ProcessEventCountFSR( self, path, ob ) :
00377 self.log.debug('Event Count Input Addition')
00378 self.fsr[path].setInput( self.fsr[path].input()+ob.input() )
00379
00380 def MergeLumiFSR( self, path, keyedC ) :
00381 from ROOT import string
00382
00383 keyedContainer = self.fsr.retrieveObject(path)
00384
00385 assert keyedContainer.numberOfObjects() == 1
00386 l = keyedContainer.containedObject(0)
00387 baseLumi = LumiFSR( l )
00388
00389 nCont = keyedC.numberOfObjects()
00390 for i in xrange(nCont) :
00391 obj = keyedC.containedObject(i)
00392 nextLumi = LumiFSR( obj )
00393 baseLumi.merge( nextLumi )
00394
00395 newLumi = gbl.LHCb.LumiFSR()
00396 for r in baseLumi.runs :
00397 newLumi.addRunNumber( r )
00398 for f in baseLumi.files :
00399 newLumi.addFileID( string(f) )
00400 for k in baseLumi.keys :
00401 increment, integral = baseLumi.info[k]
00402 newLumi.addInfo(k, increment, integral)
00403
00404 self.fsr[path].clear()
00405
00406 self.fsr[path].add(newLumi)
00407 return SUCCESS
00408
00409
00410
00411 class LumiFSR( ) :
00412 def __init__(self, lumi) :
00413
00414
00415
00416
00417
00418
00419 self.runs = []
00420 self.files = []
00421 self.info = {}
00422 self.keys = []
00423
00424
00425 for r in lumi.runNumbers() :
00426 self.runs.append(r)
00427
00428 for f in lumi.fileIDs() :
00429 self.files.append(f)
00430
00431
00432 s = str(lumi)
00433 sa = s.split("info (key/incr/integral) : ")[-1]
00434 sa = sa.split('/')[:-1]
00435 for rec in sa :
00436 k,i,t = rec.split()
00437 k = int(k)
00438 i = int(i)
00439 t = int(t)
00440 self.info[k] = (i,t)
00441 self.keys = self.info.keys()
00442 def merge( self, otherLumi ) :
00443 assert otherLumi.__class__.__name__ == "LumiFSR"
00444
00445 for r in otherLumi.runs :
00446 if r in self.runs :
00447 pass
00448 else :
00449 self.runs.append( r )
00450 self.runs.sort()
00451
00452 for f in otherLumi.files :
00453 if f in self.files :
00454 pass
00455 else :
00456 self.files.append( f )
00457 self.files.sort()
00458
00459 for k in otherLumi.keys :
00460 increment, integral = otherLumi.info[k]
00461 if k in self.keys :
00462 myIncrement, myIntegral = self.info[k]
00463 self.info[k] = ( myIncrement+increment, myIntegral+integral )
00464 else :
00465 self.info[k] = ( increment, integral )
00466
00467 self.keys = self.info.keys()
00468 def __repr__( self ) :
00469 s = "LumiFSR Python class\n"
00470 s += "\tRuns : \n"
00471 for r in self.runs :
00472 s += "\t\t%i\n"%(r)
00473 s += "\tFiles : \n"
00474 for f in self.files :
00475 s += "\t\t%s\n"%(f)
00476 s += "\tInfo : \n"
00477 for k in self.keys :
00478 increment, integral = self.info[k]
00479 s += "\t\t%i\t%i\t%i\n"%(k,increment,integral)
00480 return s
00481
00482
00483
00484 class PackedCaloHypo() :
00485 def __init__(self, o) :
00486 cl = 'LHCb::PackedCaloHypo'
00487 assert o.__class__.__name__ == cl
00488 self.centX = o.centX
00489 self.centY = o.centY
00490 self.cerr = (o.cerr00,o.cerr10,o.cerr11)
00491 self.cov = (o.cov00,o.cov10,o.cov11,o.cov20,o.cov21,o.cov22)
00492 self.firstCluster = o.firstCluster
00493 self.firstDigit = o.firstDigit
00494 self.firstHypo = o.firstHypo
00495 self.hypothesis = o.hypothesis
00496 self.key = o.key
00497 self.lastCluster = o.lastCluster
00498 self.lastDigit = o.lastDigit
00499 self.lastHypo = o.lastHypo
00500 self.lh = o.lh
00501 self.pos = (o.posE, o.posX, o.posY)
00502 self.z = o.z
00503 def __repr__( self ) :
00504 s = "PackedCaloHypo : \n"
00505 s += "\tcentX : %s\n"%( str(self.centX) )
00506 s += "\tcentY : %s\n"%( str(self.centY) )
00507 s += "\tcerr : %s\n"%( str(self.cerr ) )
00508 s += "\tcov : %s\n"%( str(self.cov ) )
00509 s += "\tfirstCluster : %s\n"%( str(self.firstCluster) )
00510 s += "\tfirstDigit : %s\n"%( str(self.firstDigit) )
00511 s += "\tfirstHypo : %s\n"%( str(self.firstHypo) )
00512 s += "\thypothesis : %s\n"%( str(self.hypothesis) )
00513 s += "\tkey : %s\n"%( str(self.key) )
00514 s += "\tlastCluster : %s\n"%( str(self.lastCluster) )
00515 s += "\tlastDigit : %s\n"%( str(self.lastDigit) )
00516 s += "\tlastHypo : %s\n"%( str(self.lastHypo) )
00517 s += "\tlh : %s\n"%( str(self.lh ) )
00518 s += "\tpos : %s\n"%( str(self.pos ) )
00519 s += "\tz : %s\n"%( str(self.z ) )
00520 s += "---------------------------------------\n"
00521 return s
00522
00523
00524
00525 class SyncMini( object ) :
00526 def __init__( self, event, lastEvent=None ) :
00527 self.event = event
00528 self.t = 0.0
00529 self.lastEvent = None
00530 if lastEvent :
00531 self.lastEvent = lastEvent
00532 def check( self ) :
00533 return self.event.is_set()
00534 def checkLast( self ) :
00535 return self.lastEvent.is_set()
00536 def reset( self ) :
00537 self.event.clear()
00538 self.t = time.time()
00539 def getTime( self ) :
00540 return self.t
00541 def set( self ) :
00542 self.event.set()
00543 def __repr__( self ) :
00544 s = "---------- SyncMini --------------\n"
00545 s += " Status : %s\n"%(self.event.is_set())
00546 s += " t : %5.2f\n"%(self.t)
00547 if self.lastEvent :
00548 s += "Last Event : %s\n"%(self.lastEvent.is_set())
00549 s += "----------------------------------\n"
00550 return s
00551
00552
00553
00554 class Syncer( object ) :
00555 def __init__( self, nWorkers, log, manyEvents=False,
00556 limit=None, step=None, firstEvent=None ) :
00557
00558 self.limit = limit
00559 self.step = step
00560 self.d = {}
00561 self.manyEvents = manyEvents
00562 for i in xrange(-2, nWorkers) :
00563 self.d[ i ] = SyncMini( Event(), lastEvent=Event() )
00564 if self.manyEvents :
00565 self.limitFirst = firstEvent
00566 self.keys = self.d.keys()
00567 self.nWorkers = nWorkers
00568 self.log = log
00569
00570 def syncAll( self, step="Not specified" ) :
00571
00572
00573
00574 if self.manyEvents :
00575 sc = self.syncAllRolling( )
00576 return sc
00577
00578
00579 for i in xrange( 0, self.limit, self.step ) :
00580 if self.checkAll( ) :
00581 self.log.info('%s : All procs done @ %i s'%(step,i))
00582 break
00583 else :
00584 time.sleep(self.step)
00585
00586
00587 if self.checkAll() :
00588 self.log.info("All processes : %s ok."%(step))
00589 return SUCCESS
00590 else :
00591 self.log.critical('Some process is hanging on : %s'%(step))
00592 for k in self.keys :
00593 hangString= "%s : Proc/Stat : %i/%s"%(step,k,self.d[k].check())
00594 self.log.critical( hangString )
00595 return FAILURE
00596
00597 def syncAllRolling( self ) :
00598
00599
00600
00601
00602
00603
00604
00605 begin = time.time()
00606 firstEv = {}
00607 timers = {}
00608 for k in self.keys :
00609 self.d[k].reset()
00610 firstEv[k] = False
00611 timers[k] = 0.0
00612
00613 active = self.keys
00614 while True :
00615
00616 for k in active :
00617 sMini = self.d[k]
00618
00619 if sMini.check() :
00620 if sMini.checkLast() :
00621
00622 active.remove( k )
00623 alive = time.time()-begin
00624 self.log.info( "Audit : Node %i alive for %5.2f"\
00625 %(k,alive) )
00626 else :
00627 sMini.reset()
00628 else :
00629
00630
00631 wait = time.time()-sMini.getTime()
00632 cond = wait > self.limit
00633 if not firstEv[k] :
00634 cond = wait > self.limitFirst
00635 firstEv[k] = True
00636 if cond :
00637
00638 self.log.critical('Single event wait : %5.2f'%(wait))
00639 self.processHang()
00640 return FAILURE
00641
00642
00643 if self.checkLastEvents() :
00644 self.log.info('TC met for event loop')
00645 break
00646 else :
00647
00648 time.sleep(self.step)
00649
00650 self.log.info("All processes Completed all Events ok")
00651 return SUCCESS
00652
00653 def processHang( self ) :
00654 self.log.critical('Some proc is hanging during Event processing!')
00655 for k in self.keys :
00656 self.log.critical( "Proc/Stat : %i / %s"%(k,self.d[k].check()) )
00657 return
00658
00659 def checkAll( self ) :
00660
00661
00662 currentStatus = [ mini.check() for mini in self.d.values() ]
00663 return all( currentStatus )
00664
00665 def checkLastEvents( self ) :
00666
00667 stat = [ sMini.checkLast() for sMini in self.d.values() ]
00668 return all(stat)
00669
00670
00671
00672 def getEventNumber( evt ) :
00673
00674
00675 n = None
00676
00677 lst = [ '/Event/Gen/Header',
00678 '/Event/Rec/Header' ]
00679 for l in lst :
00680 try :
00681 n = evt[l].evtNumber()
00682 return n
00683 except :
00684
00685 continue
00686
00687
00688
00689 try :
00690 n = evt['/Event/DAQ/RawEvent'].banks(16)[0].data()[4]
00691 return n
00692 except :
00693 pass
00694
00695
00696 return n
00697
00698
00699
00700