grid for cbm kilian schwarz, gsi. what is grid ? ● sharing of distributed resources within one...

Post on 18-Dec-2015

217 Views

Category:

Documents

1 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Grid for CBM

Kilian Schwarz, GSI

What is Grid ?

● Sharing of distributed resources within one Virtual Organisations !!!!

70

538

27

4603637

55

22 87

10

Europa: 267 Institute, 4603 UserSonstige: 208 Institute, 1632 User

LHC Wissenschaftler weltweit

Start of CBM Grid

● There are considerations to start a CBM Grid● Task: distributed MC production● Potential sites: 3 (Bergen, Dubna, GSI)● After positive experiences the Grid can be

enlarged to more sites and tasks, like distributed analysis

requirements* Globus-style X509 user certificates

issued for CBM by GermanGrid CA

http://www.gridka.de

* How to get a certificate ?

at GSI: > . globuslogin

> grid-cert-request –cn “<surname> <name>”

certificate request file and private key will be stored in $HOME/.globus

The request file has to be signed (openssl) by the CA responsible person and mailed to GermanGrid CA

The certificate will be mailed back via e-mail

GermanGrid CA

How to get a certificate in detail:

See http://wiki.gsi.de/Grid/DigitalCertificates

requirements: CBM VO Server (one per VO)

additional sites:

- Bergen, Dubna

additional users:

- to be added

Globus/LCG – creation of grid-mapfilenecessary for each site

● E.g. with gLite-security tools:

- adjust $GLITE_LOCATION/etc/glite-mkgridmap.conf

add: “group ldap://glite001.gsi.de:8389/o=cbm,dc=de,dc=de”

- Create grid-mapfile

$GLITE_LOCATION/sbin/glite-mkgridmap –output=/etc/grid-security/grid-mapfile

user creation on each site (support of CBM VO)

• Each site has to create cbm-user-IDs onto which the Grid-users will be mapped:

EGEE/LCG: a certain number of POOL accounts, e.g. cbmvo00 – cbmvo10

Globus & AliEn: one production user: via this userID the jobs will be submitted. E.g. cbmprod

CBM software environment

● To be able to send real CBM jobs to the Grid, the participating sites have to

* Install the CBM software and prepare the environment

* Or the job has to bring it’s own environment (static links)

Agreement on common Grid middleware

• basically, the possibilities are:• - Globus• - NorduGrid• - LCG-2• - AliEn• - gLite (EGEE)• - gLite (AliEn)

LHC Computing Grid ProjectFundamental Goal of the LCG

To help the experiments’ computing projects

Phase 1 – 2002-05prepare and deploy the environment for LHC computing

Phase 2 – 2006-08acquire, build and operate the LHC computing service

SC2 – Software & Computing CommitteeSC2 includes the four experiments, Tier 1 Regional CentresSC2 identifies common solutions and sets requirements for the project

PEB – Project Execution BoardPEB manages the implementation

organising projects, work packagescoordinating between the Regional Centres

EDG Middleware Architecture

Collective Services

Information &

Monitoring

Replica Manager

Grid Scheduler

Local Application Local Database

Underlying Grid Services

Computing Element Services

Authorization Authentication and Accounting

Replica Catalog

Storage Element Services

SQL Database Services

Fabric services

ConfigurationManagement

Node Installation &Management

Monitoringand

Fault Tolerance

Resource Management

Fabric StorageManagement

Grid

Fabric

Local Computing

Grid Grid Application Layer

Data Management

Job Management

Metadata Management

Service Index

APPLICATIONS

GLOBUSCondorG

(via VDT)

M / W

Dubna (JINR): LCG-2 site

Dubna (JINR): LCG-2 siteLCG-test mostly successful

JINR (LCG-2 site: job-submit)

Timeline

o After only 2 years of development, we have deployed a distributed computing environment which meets the needs of Alice experiment Simulation & Reconstruction Event mixing Analysis

o Using Open Source components (representing 99% of the code), internet standards (SOAP,XML, PKI…) and scripting language (perl) was the key element that alllowed quick prototyping and very fast development cycles

2001 2002 2003 2004 2005

First production (distributed simulation)

10% DC (analysis)

P. Buncic, CERN

Building AliEn

SvOutPlaceObjectSvOutPlaceObject

SvOutPlaceObject

SvOutPlaceObject

P. Saiz, CERN

SvOutPlaceObject

SvOutPlaceObject

SvOutPlaceObject

SvOutPlaceObject

AliEn Grid (ALICE VO):● 77 configured sites worldwide

DC Monitoring: http://alien.cern.ch● Monalisa: http://aliens3.cern.ch:8080

lxts05.gsi.de: AliEn client (PANDA VO)

JINR and Bergen: AliEn sites

JINR and Bergen: AliEn sites

Grids and Open Standards

Incr

ease

d fu

nctio

nalit

y,st

anda

rdiz

atio

n

Time

Customsolutions

Open GridServices

Arch

GGF: OGSI, …(+ OASIS, W3C)

Multiple implementations,including Globus Toolkit

Web services

Globus Toolkit

Defacto standardsGGF: GridFTP, GSI

X.509,LDAP,FTP, …

App-specificServices

Architecture Guiding Principles

● Lightweight (existing) services – Easily and quickly deployable– Use existing services where possible as

basis for re-engineering

● Interoperability– Allow for multiple implementations

● Resilience and Fault Tolerance

● Co-existence with deployed infrastructure– Run as an application (e.g. on LCG-2; Grid3)– Reduce requirements on site components

● Basically globus and SRM– Co-existence (and convergence) with LCG-2 and Grid3 are essential for the EGEE Grid

service

● Service oriented approach– WSRF still being standardized– No mature WSRF implementations exist to date, no clear picture about the impact of WSRF

hence: start with plain WS● WSRF compliance is not an immediate goal, but we follow the WSRF evolution● WS-I compliance is important

Approach

● Exploit experience and components from existing projects

– AliEn, VDT, EDG, LCG, and others● Design team works out architecture

and design– Architecture: https://edms.cern.ch/document/476451– Design: https://edms.cern.ch/document/487871/

● Components are initially deployed on a prototype infrastructure– Small scale (CERN & Univ. Wisconsin)– Get user feedback on service semantics and interfaces

● After internal integration and testing components are delivered to SA1 and deployed on the pre-production service

EDGVDT . . .

LCG

EGEE

. . .AliEn

gLite (AliEn)

* From now on used by ALICE for globally distributed analysis in connection with

• PROOF (at GSI: http://www-w2k.gsi.de/root/• PROOF at GSI )

gLite (EGEE)

* Will replace LCG-2.X in near? future, but nobody has real experience with it

summary (middlewares)● LCG-2: GSI and Dubna • - pro: large distribution, support• - contra: difficult to set up, no distributed analysis● AliEn: GSI, Dubna, Bergen

- pro: in production since 2001

- contra: unsecure future, no support

Globus 2: GSI, Dubna, Bergen?

- pro/contra: simple, but functioning (no RB, no FC, no support)

gLite/GT4: new on the market

- pro/contra: nobody has production experience (gLite)

lxg01-05.gsi.de● LCG test installation, visible in LCG –

preproduction testbed● Trying to port LCG to Debian Linux

top related