grid computing status report jeff templon pdp group, nikhef nikhef scientific advisory committee 20...

13
Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Upload: silas-walsh

Post on 03-Jan-2016

215 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Grid Computing Status Report

Jeff Templon

PDP Group, NIKHEF

NIKHEF Scientific Advisory Committee

20 May 2005

Page 2: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 2

HEP Computing Model

Tier-0 : measurement center (CERN) Dedicated computers (L2/L3 trigger farms) Archival of raw data

Tier-1 : data centers Archival of 2nd copy of raw data Large-scale computing farms (e.g. reprocessing) Spread geographically Strong Support

Tier-2 : user facilities for data analysis / Monte Carlo

Page 3: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 3

Worldwide HEP Computing Needs

2006 2007 2008 2009 2010

ATLAS 1.2E+04 3.6E+04 5.9E+04 1.0E+05LHCb 2.6E+03 5.2E+03 8.6E+03 9.6E+03 1.2E+04ALICE 4.7E+03 9.3E+03 2.3E+04

ATLAS 9.7 27.7 42.8 72.5LHCb 1.0 2.0 3.3 4.0 4.7ALICE 2.0 4.1 10.2

ATLAS 5.2 16.8 29.7 48.4LHCb 1.0 2.1 3.4 7.1 11.6ALICE 2.0 3.9 9.8

world CPU needs (3,6 GHZ Xeon equiv CPU)

world fast storage (disk) needs (petabytes)

world permanent storage (tape) needs (petabytes)

Page 4: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 4

Amsterdam : Tier-1 for LHC

Three experiments : LHCb / ATLAS / ALICE

Overall scale determined by estimating funding in NL

Contribution to experiments scaled by NIKHEF presence:3:2:1

Resulting NIKHEF share of total Tier-1 needs: LHCb: 23%

ATLAS: 11,5%

ALICE: 5,75%

Page 5: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 5

Amsterdam Tier-1 Numbers

2006 2007 2008 2009 2010

CPU 204 1118 2921 5029 8203Disk 0.2 1.1 2.5 3.8 6.1Tape 0.1 0.7 1.8 3.5 5.7

Status: GOOD!Basic collaboration with SARA in place

Attitude adjustment needed (response time)

Appropriate funding line in NCF long-term draft plan

Just enough; concerns about ‘me-too’ (grids are popular)

Community-building (VL-E project)

Pull me-too people into same infrastructure

Page 6: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 6

Overall Status LHC Computing

LCG a successful service 14,000 CPUs and well-ordered operations, active community

Monte Carlo productions working well (next slide)

Data Management a problem Software never converged in EDG

May not be converging in EGEE (same team)

Risk losing HEP community on DM

Makes community-forming (generic middleware) difficult:I’ll just build my own, this one stinks

Page 7: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 7

Results of “Data Challenge ’04”

Monte Carlo tasks distributed to computers across world

Up to 3000 simultaneous “jobs” per experiment

2.2 million CPU-hours (250 years) used in one month

Total data volume > 25 TB

For LHCb: NIKHEF ~ 6% of global total

See it in actionBackup

Page 8: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 8

Transport of primary data to Tier-1s

Page 9: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

LCG Service Challenge II

“The Dutch Contribution”

                           

Page 10: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 10

Local Status Positioned well in LCG & EGEE

Present on ‘blessed’ Tier-1 list One of ‘best run’ sites One of first sites (#3 in EDG, compare #4 in WWW) Membership on:

Middleware Design Team (US collaboration here too) Project Technical Forum LCG Grid Applications Group (too bad, almost defunct) Middleware Security Group Etc etc etc

D. Groep chairs world-recognized EUGridPMA

K. Bos chairs LHC Grid Deployment Board

Page 11: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 11

Local Status #2

NIKHEF “grid site” Roughly 300 CPUs / 10 terabytes of storage

Several distinct components LCG / VL-E production LCG pre-production EGEE testing VL-E certification

Manpower: 8 staff, interviews this week for three more (project funding)

Page 12: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 12

PDP Group Activities

Middleware (3 FTE) Mostly “security” -- best bang for buck + local world expert

Operations (3 FTE) How does one operate a terascale / kilocomputer site? Knowledge transfer to SARA (they have support mandate) Contribute regularly to operational middleware

Applications (3 FTE) Strong ties to local HEP (ATLAS “Rome” production, LHCb

Physics Performance Report, D0 “SAMGrid”) Community forming: LOFAR & KNMI, looking for others

Page 13: Grid Computing Status Report Jeff Templon PDP Group, NIKHEF NIKHEF Scientific Advisory Committee 20 May 2005

Jeff Templon – NIKHEF SAC, 2005.05.20 - 13

Industrial Interest

GANG Hosted @ NIKHEF

IBM, LogicaCMG, Philips, HPC, UvA, SARA, NIKHEF … 16 industrial participants (24 total)