grid for cbm kilian schwarz, gsi. what is grid ? ● sharing of distributed resources within one...
Post on 18-Dec-2015
217 views
TRANSCRIPT
Grid for CBM
Kilian Schwarz, GSI
What is Grid ?
● Sharing of distributed resources within one Virtual Organisations !!!!
70
538
27
4603637
55
22 87
10
Europa: 267 Institute, 4603 UserSonstige: 208 Institute, 1632 User
LHC Wissenschaftler weltweit
Start of CBM Grid
● There are considerations to start a CBM Grid● Task: distributed MC production● Potential sites: 3 (Bergen, Dubna, GSI)● After positive experiences the Grid can be
enlarged to more sites and tasks, like distributed analysis
requirements* Globus-style X509 user certificates
issued for CBM by GermanGrid CA
http://www.gridka.de
* How to get a certificate ?
at GSI: > . globuslogin
> grid-cert-request –cn “<surname> <name>”
certificate request file and private key will be stored in $HOME/.globus
The request file has to be signed (openssl) by the CA responsible person and mailed to GermanGrid CA
The certificate will be mailed back via e-mail
GermanGrid CA
How to get a certificate in detail:
See http://wiki.gsi.de/Grid/DigitalCertificates
requirements: CBM VO Server (one per VO)
additional sites:
- Bergen, Dubna
additional users:
- to be added
Globus/LCG – creation of grid-mapfilenecessary for each site
● E.g. with gLite-security tools:
- adjust $GLITE_LOCATION/etc/glite-mkgridmap.conf
add: “group ldap://glite001.gsi.de:8389/o=cbm,dc=de,dc=de”
- Create grid-mapfile
$GLITE_LOCATION/sbin/glite-mkgridmap –output=/etc/grid-security/grid-mapfile
user creation on each site (support of CBM VO)
• Each site has to create cbm-user-IDs onto which the Grid-users will be mapped:
EGEE/LCG: a certain number of POOL accounts, e.g. cbmvo00 – cbmvo10
Globus & AliEn: one production user: via this userID the jobs will be submitted. E.g. cbmprod
CBM software environment
● To be able to send real CBM jobs to the Grid, the participating sites have to
* Install the CBM software and prepare the environment
* Or the job has to bring it’s own environment (static links)
Agreement on common Grid middleware
• basically, the possibilities are:• - Globus• - NorduGrid• - LCG-2• - AliEn• - gLite (EGEE)• - gLite (AliEn)
LHC Computing Grid ProjectFundamental Goal of the LCG
To help the experiments’ computing projects
Phase 1 – 2002-05prepare and deploy the environment for LHC computing
Phase 2 – 2006-08acquire, build and operate the LHC computing service
SC2 – Software & Computing CommitteeSC2 includes the four experiments, Tier 1 Regional CentresSC2 identifies common solutions and sets requirements for the project
PEB – Project Execution BoardPEB manages the implementation
organising projects, work packagescoordinating between the Regional Centres
EDG Middleware Architecture
Collective Services
Information &
Monitoring
Replica Manager
Grid Scheduler
Local Application Local Database
Underlying Grid Services
Computing Element Services
Authorization Authentication and Accounting
Replica Catalog
Storage Element Services
SQL Database Services
Fabric services
ConfigurationManagement
Node Installation &Management
Monitoringand
Fault Tolerance
Resource Management
Fabric StorageManagement
Grid
Fabric
Local Computing
Grid Grid Application Layer
Data Management
Job Management
Metadata Management
Service Index
APPLICATIONS
GLOBUSCondorG
(via VDT)
M / W
Dubna (JINR): LCG-2 site
Dubna (JINR): LCG-2 siteLCG-test mostly successful
JINR (LCG-2 site: job-submit)
Timeline
o After only 2 years of development, we have deployed a distributed computing environment which meets the needs of Alice experiment Simulation & Reconstruction Event mixing Analysis
o Using Open Source components (representing 99% of the code), internet standards (SOAP,XML, PKI…) and scripting language (perl) was the key element that alllowed quick prototyping and very fast development cycles
2001 2002 2003 2004 2005
First production (distributed simulation)
10% DC (analysis)
P. Buncic, CERN
Building AliEn
SvOutPlaceObjectSvOutPlaceObject
SvOutPlaceObject
SvOutPlaceObject
P. Saiz, CERN
SvOutPlaceObject
SvOutPlaceObject
SvOutPlaceObject
SvOutPlaceObject
AliEn Grid (ALICE VO):● 77 configured sites worldwide
DC Monitoring: http://alien.cern.ch● Monalisa: http://aliens3.cern.ch:8080
lxts05.gsi.de: AliEn client (PANDA VO)
JINR and Bergen: AliEn sites
JINR and Bergen: AliEn sites
Grids and Open Standards
Incr
ease
d fu
nctio
nalit
y,st
anda
rdiz
atio
n
Time
Customsolutions
Open GridServices
Arch
GGF: OGSI, …(+ OASIS, W3C)
Multiple implementations,including Globus Toolkit
Web services
Globus Toolkit
Defacto standardsGGF: GridFTP, GSI
X.509,LDAP,FTP, …
App-specificServices
Architecture Guiding Principles
● Lightweight (existing) services – Easily and quickly deployable– Use existing services where possible as
basis for re-engineering
● Interoperability– Allow for multiple implementations
● Resilience and Fault Tolerance
● Co-existence with deployed infrastructure– Run as an application (e.g. on LCG-2; Grid3)– Reduce requirements on site components
● Basically globus and SRM– Co-existence (and convergence) with LCG-2 and Grid3 are essential for the EGEE Grid
service
● Service oriented approach– WSRF still being standardized– No mature WSRF implementations exist to date, no clear picture about the impact of WSRF
hence: start with plain WS● WSRF compliance is not an immediate goal, but we follow the WSRF evolution● WS-I compliance is important
Approach
● Exploit experience and components from existing projects
– AliEn, VDT, EDG, LCG, and others● Design team works out architecture
and design– Architecture: https://edms.cern.ch/document/476451– Design: https://edms.cern.ch/document/487871/
● Components are initially deployed on a prototype infrastructure– Small scale (CERN & Univ. Wisconsin)– Get user feedback on service semantics and interfaces
● After internal integration and testing components are delivered to SA1 and deployed on the pre-production service
EDGVDT . . .
LCG
EGEE
. . .AliEn
gLite (AliEn)
* From now on used by ALICE for globally distributed analysis in connection with
• PROOF (at GSI: http://www-w2k.gsi.de/root/• PROOF at GSI )
gLite (EGEE)
* Will replace LCG-2.X in near? future, but nobody has real experience with it
summary (middlewares)● LCG-2: GSI and Dubna • - pro: large distribution, support• - contra: difficult to set up, no distributed analysis● AliEn: GSI, Dubna, Bergen
- pro: in production since 2001
- contra: unsecure future, no support
Globus 2: GSI, Dubna, Bergen?
- pro/contra: simple, but functioning (no RB, no FC, no support)
gLite/GT4: new on the market
- pro/contra: nobody has production experience (gLite)
lxg01-05.gsi.de● LCG test installation, visible in LCG –
preproduction testbed● Trying to port LCG to Debian Linux