wlcg lhcc mini-review lhcb summary. outline m activities in 2008: summary m status of dirac m...

Download WLCG LHCC mini-review LHCb Summary. Outline m Activities in 2008: summary m Status of DIRAC m Activities in 2009: outlook m Resources in 2009-10 PhC2

If you can't read please download the document

Upload: tyrone-flowers

Post on 18-Jan-2018

222 views

Category:

Documents


0 download

DESCRIPTION

Tier1s (re-)configuration m LFC mirror, ConditionsDB replication o DB replication using 3D from CERN to all Tier1s P In place for the whole year o LFC service for scalability and redundancy P Some problems at GridKa m Site SE migration (winter 08-09) o RAL (dCache to Castor2) P T0D1 migration went rather smoothly (FTS copy of files) P T1D0 migration extremely painful (staging tape by tape) d Took several months o PIC (Castor1 to dCache for T1D0) P Went very smoothly without file copy (file migration to Enstore) o CNAF (Castor2 to StoRM for TxD1) P Performed in May m SURLs migrated to SRM v2.2 end-points (October) P Needs as datasets in use for 2.5 years now (DC06) P No dependency on SRM v1 WLCG-MB report, February 083

TRANSCRIPT

WLCG LHCC mini-review LHCb Summary Outline m Activities in 2008: summary m Status of DIRAC m Activities in 2009: outlook m Resources in PhC2 LHCb Tier1s (re-)configuration m LFC mirror, ConditionsDB replication o DB replication using 3D from CERN to all Tier1s P In place for the whole year o LFC service for scalability and redundancy P Some problems at GridKa m Site SE migration (winter 08-09) o RAL (dCache to Castor2) P T0D1 migration went rather smoothly (FTS copy of files) P T1D0 migration extremely painful (staging tape by tape) d Took several months o PIC (Castor1 to dCache for T1D0) P Went very smoothly without file copy (file migration to Enstore) o CNAF (Castor2 to StoRM for TxD1) P Performed in May m SURLs migrated to SRM v2.2 end-points (October) P Needs as datasets in use for 2.5 years now (DC06) P No dependency on SRM v1 WLCG-MB report, February 083 Pit 8 to Tier0-Castor transfers m First weeks in February: continuous transfers at low rate m As of 18 Feb: nominal rate (70 MB/s) with ~50% duty cycle o A few longer stops for SW upgrades WLCG-MB report, February 084 Migration Tier1 Transfers Castor migration WLCG-MB report, February 085 May CCRC: reconstruction 41.2k reconstruction jobs submitted 27.6k jobs proceeded to done state Done/created ~67% Cosmics and first beam data m Cosmics used for detector commissioning o Of course very few detectors are hit! o Allows internal time alignment of subdetectors P Using readout of consecutive 25ns slots o Partially time alignment between detectors P Shifting timing by 12.5 ns and equalising population in consecutive bins o All subdetectors included in global runs as of end August m TED data o LHCb was first to see tracks coming from the injection line! o Single shots with ~2 muons/cm 2, but once every 48 s only! o First tracks in the VeLo o Allowed rough global detector time alignment (~2ns) m 10 th September o Only muons, calorimeters and for short time OT PhC8 LHCb Status to LCG-LHCC referees First data PhC9 LHCb Status to LCG-LHCC referees DIRAC3 put in production m Production activities o Started in July o Simulation, reconstruction, stripping P Includes file distribution strategy, failover mechanism P File access using local access protocol (rootd, rfio, (gsi)dcap, xrootd) P Commissioned alternative method: copy to local disk d Drawback: non-guaranteed space, less CPU efficiency, additional network traffic (possibly copied from remote site) o Failover using VOBOXes P File transfers (delegated to FTS) P LFC registration P Internal DIRAC operations (bookkeeping, job monitoring) m Analysis o Started in September o Ganga available for DIRAC3 in November o DIRAC2 de-commissioned on January 12th PhC10 LHCb DIRAC3 jobs PhC11 LHCb Status to LCG-LHCC referees Issues in 2008 m Data Management o Site configuration (non-scaling) o SRM v2.2 still not fully mature (e.g. pinning) o Many issues with StorageWare (mainly dCache) m Workload Management o Moved to gLite WMS, but still many issues with it (e.g. mix- up of identities). Better scaling behavior though than LCG-RB o LHCb moved to using generic pilot jobs (i.e. can execute workload from any user or production) P Not switching identity yet (gLexec / SCAS not available) P Not a show-stopper as not required by LHCb but by sites m Middleware deployment o LHCb distributes the client middleware P From distribution in the LCG-AA P Necessary to ensure bug fixes to be available P Allows multiple platform (OS, architecture, python version) PhC12 LHCb Computing Operations m Production manager o Schedules production work, sets up and checks workflows, reports to LHCb operations m Computing shifters o Computing Operations shifter (pool of ~12 shifters) P Covers 14h/day, 7 days / week P Computing Control room (2-R-014) o Data Quality shifter P Covers 8h/day, 7 days / week o Both are in the LHCb Computing Control room (2-R-014) m Daily DQ and Operations meetings o Week days (twice a week during shutdowns) m Grid Expert on-call o On duty for a week o Runs the operations meetings m Grid Team (~6 FTEs needed, ~2 missing) o Shared responsibilities (WMS, DMS, SAM, Bookkeeping) PhC13 LHCb Status to LCG-LHCC referees Plans for 2009 m Commissioning for data taking (FEST09) o See next slides m Simulation o Replacing DC06 datasets P Signal and background samples (~300 Mevts) P Minimum bias for L0 and HLT commissioning (~100 Mevts) P Used for CP-violation performance studies P Nominal LHC settings (7 TeV, 25 ns, cm -2 s -1 ) o Tuning stripping and HLT for 2010 P 4/5 TeV, 50 ns (no spillover), cm -1 s -1 P Benchmark channels for first physics studies B , s, B Dh, B s J/ , B K * P Large minimum bias samples (~ 1mn of LHC running) P Stripping performance required: ~ 50 Hz for benchmark channels P Tune HLT: efficiency vs retention, optimisation o Preparation for very first physics P 2 TeV, low luminosity P Large minimum bias sample (part used for FEST09) PhC14 LHCb FEST09 m Aim o Replace the non-existing 2008 beam data with MC o Points to be tested P L0 (Hardware trigger) strategy d Emulated in software P HLT strategy d First data (loose trigger) d Higher lumi/energy data (b-physics trigger) P Online detector monitoring d Based on event selection from HLT e.g. J/Psi events d Automatic detector problems detection P Data streaming d Physics stream (all triggers) and calibration stream (subset of triggers, typically 5 Hz) P Alignment and calibration loop d Trigger re-alignment d Run alignment processes d Validate new alignment (based on calibration stream) PhC15 LHCb Status to LCG-LHCC referees FEST09 preparation (2) m Online developments o Event injector P Read MC files P Emulate L0 trigger (conditions my be varied) P Creates multi-event packets (MEP as font-end does) P Send MEP to an HLT farm node o Event injector control system P Emulation of the standard Run Control P Simulates a regular run, but using event injector as source o Multiple online streams P Using HLT classification as criterion d Was not needed for 2008 run, hence was delayed o Readiness P Tests in December, operational in January P First FEST week: 26 January d Mainly online commissioning, limited data transfers P Next FEST week: 2 March PhC16 LHCb Status to LCG-LHCC referees FEST09 runs m FEST activity o Define running conditions (rate, HLT version + config) o Start runs from the Control System P Events are injected and follow the normal path o Files export to Tier0 and distribution to Tier1s o Automatic reconstruction jobs at CERN and Tier1s P Commission Data Quality green-light m Short test periods o Typically full week, to 1 day every week for tests o Depending on results, take a few weeks interval for fixing problems m Vary conditions o L0 parameters o Event rates o HLT parameters o Trigger calibration and alignment loop PhC17 LHCb Status to LCG-LHCC referees Resources (very preliminary) m Consider as a whole (new LHC schedule) o Real data P Split year in two parts: d s at low lumi LHC-phase1 d s at higher lumi ( ) LHC phase2 P Trigger rate independent on lumi and energy: 2 kHz o Simulation: events (nominal year) over 2 years m New assumptions for (re-)processing and analysis o More re-processings during LHC-phase1 o Add calibration checks (done at CERN) o Envision more analysis at CERN with first data P Increase from 25% (TDR) to 50% (phase1) and 35% (phase2) P Include SW development and testing (LXBATCH) o Adjust event sizes and CPU needs to current estimates P Important effort to reduce data size (packed format for rDST, DST, DST) P Use new HEP-SPEC06 benchmarking PhC18 LHCb Status to LCG-LHCC referees Resources (contd) m CERN usage o Tier0: P Real data recording, export to Tier1s P First pass reconstruction of ~85% of raw data P Reprocessing (in future foresee to use also the Online HLT farm) o CAF (Calibration and Alignment Facility) P Dedicated LXBATCH resources P Detector studies, alignment and calibration o CAF (CERN Analysis Facility) P Part of Grid distributed analysis facilities (estimate 40% in ) P Histograms and interactive analysis (lxplus, desk/lap-tops) m Tier1 usage o Reconstruction P First pass during data taking, reprocessing o Analysis facilities P Grid distributed analysis P Local storage for users data (LHCb_USER SRM space) PhC19 LHCb Conclusions m 2008 o CCRC very useful for LHCb (although irrelevant to be simultaneous due to low throughput) o DIRAC3 fully commissioned P Production in July P Analysis in November P As of now, called DIRAC o Last processing on DC06 P Analysis will continue in 2009 o Commission simulation and reconstruction for real data m o Large simulation requests for replacing DC06, preparing o FEST09: ~1 week a month and 1 day a week o Resource requirements being prepared for WLCG workshop in March and C-RRB in April PhC20 LHCb