atlas tdaq system administration: evolution and...

22
ATLAS TDAQ System Administration: evolution and re-design CHEP 2015 Christopher Jon Lee University of Johannesburg, South Africa CERN for and on behalf of the ATLAS TDAQ SysAdmin team

Upload: others

Post on 02-Jun-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

ATLAS TDAQ System Administration: evolution and re-design

CHEP 2015

Christopher Jon Lee University of Johannesburg, South Africa

CERN

for and on behalf of the

ATLAS TDAQ SysAdmin team

Page 2: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Overview

CHEP April 2015 2

After 3 years of LHC beam (Run1), 2 years of upgrades (LS1), Restart has begun…

NAS CFS LFS

LFS PUPPET SVN

CLIENTS

Configuration Data

*See appendix for acronyms

Page 3: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Overview

CHEP April 2015 3

After 3 years of LHC beam (Run1), 2 years of upgrades (LS1), Restart has begun…

Config

Monitoring

Network Services

LDAP

DHCP DNS

NAS & CFS ATLAS / SATELLITE CONTROL ROOMS

DCS

Detector Hosts

HLT / DAQ Farm: 1 LFS / rack of clients

SFO: Sub Farm Output

WWW

Gateways

*See appendix for acronyms

Page 4: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

OS Upgrade Scientific Linux CERN

Ø  Only supported Linux OS in use Ø  Full support direct from on-site experts Ø  All Linux machines are now running SLC6

ü  will remain the OS (Major) version for Run 2 Windows OS

Ø  Used by Detector Control System (DCS) for one specific application Ø  SLC6 hosts Windows VM Ø  Windows VM is managed by DCS

²  During beam, NO changes are made to the running system

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 4

Page 5: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Local boot vs Net boot u  664 Local boot – Servers, DCS, TDAQ/SysAdmin Infrastructure u  2392 Net boot – HLT Farm, Read Out Systems, Single Board Computers, etc u  Local boot (Standard installation with boot from disk)

Ø  provisioning by PXE + Kickstart + Ø  DHCP + PXE provided by an LFS from Configuration Database (See slides below)

Ø  template-based kickstart files u  “Net Boot” via PXE

Ø  the more components one has in a system, the greater the risk of failure, So... reduce any components that are not “needed”

Ø  in ATLAS, extensive use of PCs with no operating system on disk u  Each reboot, is essentially a fresh clean OS

Advantages: Ø  ease of maintenance Ø  reproducibility on a large scale Ø  reduced HW replacement times

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 5

Disadvantages: Ø  requires ad-hoc development and support Ø  not suitable for running servers Ø  less flexible

Page 6: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Netbooted – Redesign for SLC6 u  A completely new netbooting system compared to SLC 4 & 5

Ø  based on NFSroot and customized to our needs §  only R/W areas are kept in RAM (e.g.: /etc, /var, …) §  “bind-mounts” overlaid on R/O NFS mount of / from the LFS §  gives the users more free RAM for running their apps

Ø  Image created by in a chrooted environment Ø  NO “Golden Image”

Ø  always able to rebuild from versioned config u  Support for old hardware

Ø  32 bit non-PAE kernel provided and maintained by CERN IT (on a best effort basis)

Ø  ELF image for non-PXE clients requires private patch of mknbi package

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 6

Page 7: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Configuration Management Systems u  With such a large cluster of machines CMS’s are the only way

to sanely control what happens on machines in a large farm u  Quattor was retired as it was becoming obsolescent u  chosen over Chef and CF Engine

Ø  previous experience by existing Sys Admins Ø  during LS1 CERN IT also adopted Ø  WLCG applications mostly puppetised

u  All systems are Puppet controlled Ø  local boot : Daemon run, every 30 min Ø  net boot: puppet --apply via hourly cron job Ø  no need to reboot in order to apply

simple configuration changes u  Very similar system between the two

Ø  code re-usability Ø  easier to maintain

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 7

Page 8: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

version2 u  ConfDB is our core Configuration DataBase

Ø  PHP based web UI, Python for utilities and REST API Ø  configuration “state” database of all systems

§  DHCP details §  checks

Ø  entire system other than ConfDB is maintained by “code” Ø  interface between CERN IT Databases and Ø  web UI provides various tools to ease cluster management – ssh, IPMI etc

u  Included an OS release system for SLC6+ Ø  different release versions can run on different machines Ø  useful for testing new versions and/or revert changes in case of problems

u  More functionality added to REST API, used by puppet and other tools e.g. Ø  network interface configuration (including bonding) Ø  getting machine status in and configure it accordingly

(e.g. TDAQ or Sim@P1*) CHEP April 2015 C.J.Lee – CERN, UJ: ATLAS TDAQ SysAdmin 8

§  operational status §  boot type / parameters

* See slide 9

Page 9: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Virtualisation u  Mostly CORE and TEST systems, NOT for DAQ/HLT u  No cloud-like approach

Ø  no shared storage, privileged simplicity u  Instead of a single redundant system, rely on multiplicity of systems

u  Currently running as VM’s Ø  gateways Ø  DCS Windows services Ø  Ø  public Nodes

u  For new VH hardware, tested and really happy with the results of FlashCache Ø  huge improvement on disk IO vs cost

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 9

Ø  domain controllers Ø  development web servers Ø  Ø  LDAP

ü  50 VMs in Point 1 ü  ~100 DCS

ü  51 VMs in GPN ü  ~2700 Sim@P1 *

* See next slide

Page 10: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Simulation @ P1 u  Sim@P1 is the opportunistic use of the existing TDAQ HLT farm as a grid cluster

Ø  allow non-utilised resources to be exploited for ATLAS Prod jobs u  Virtual machines on top of the HLT machines, acting as computing nodes interconnected through a

Virtual LAN on the data network Ø  VMs and VLANs isolate the offline computing nodes network

§  no interferences with ATCN & DCS §  security

u  Communicate with the outside world (GPN) via a logically separated link to CERN IT Ø  ACLs are in place for allowing only traffic towards the needed CERN services (Castor / EOS, Condor, etc.)

u  More than 1300 nodes of the HLT farm are now able to run Monte Carlo, high CPU intensive jobs Ø  EVGEN and SIMULATION only, by design

u  Produced 1.7 billion Monte Carlo events since 1st Jan 2014 Ø  limited by downtimes during LS1 (cooling, power interventions etc)

u  Switching between states is controlled by ATLAS control room u  For more information, see our poster on the topic

Ø  “Design, Results, Evolution and Status of ATLAS simulation in Point 1”. Poster by Franco Brasolin

CHEP April 2015 C.J.Lee – CERN, UJ: ATLAS TDAQ SysAdmin 10

Page 11: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

CHEP April 2015 C.J.Lee – CERN, UJ: ATLAS TDAQ SysAdmin 11

u  Ganglia as collector for performance/health metrics, high scalability with rrdcache u  Ganglia-web provides advanced user interface to historical data in RRDs u  Icinga replaces Nagios

Ø  provides active checks and alerting Ø  can use Ganglia data Ø  can reuse Nagios plugins and much of Nagios configuration

HW monitoring via IPMI u  Complete rewrite during LS1, work still in progress u  New version based on OpenIPMI, previous based on IPMItool

Ø  unique sensor ID’s vs (unstable) sensor names Ø  better performance with SDR caching

u  Local readout fed to Ganglia u  ICINGA monitors SEL, Sensor OK state,

specific values via Ganglia u  IPMI varies with vendor, type, version…

always catching up

&

!!!!! CACHE!

!!!!!

!!!!!!!!!!!!!!!!!!!!!!!!!!!!PC!

BMC!

gmetric!

gmond!

SNMP! snmp7ext!

Ping,!service,!etc!

Pull Push

Page 12: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Monitoring & logging implementation

u  One VM for core systems: ~570 nodes, ~5000 checks

u  One PC for farm systems: ~2200 nodes, ~31000 checks

u  Users receiving status notifications u  Testing Icinga 2 for distributed scheduling,

configuration and performance

CHEP April 2015 12

System Logs management u  Rsyslog on all machines, also as collector for remote (replaces syslog-ng) u  Remote logging to LFS or central syslog servers

Ø  net boot: 2 day retention (local), 30 day on LFS Ø  local boot: 30 day retention Ø  exposed systems: 12 weeks

u  Remote logging for security critical servers to CERN IT u  Investigating central collection & analysis tools: Splunk, ELSA, logstash + elastic search

Page 13: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Summary u  LS1 was anything but a “shutdown” for our Team u  We have streamlined and improved the Point1 system

Ø  many more tasks are now fully automated, with very little human intervention needed u  Puppet, cleaner and easier to maintain u  Monitoring much more comprehensive than before

Ø  provides many more checks, still rapidly evolving u  Still Investigating Open LDAP (2.4.39) issues with opening/closing connections. u  Hoping Run 2 will provide us some “quiet time” to clean up

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 13

Page 14: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Glossary

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 14

ATCN: Atlas Technical and Control Network CFS: Central File Server DCS: Detector Control Systems GPN: General Public Network HLT: High Level Trigger LFS: Local File Server LHC: Large Hadron Collider LS1: Long Shutdown 1 NAS: Network Attached Storage PXE: Preboot eXecution Environment ROS: Read Out System SBC: Single Board Computer SLC: Scientific Linux: Cern edition SVN: Subversion TDAQ: Trigger and Data Acquisition WLCG: Worldwide LHC Computing Grid

Page 15: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

BACKUPS / SPARE

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 15

Page 16: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Introduction LHC & ATLAS

u  Large Hadron Collider, an accelerator, ~100m underground u  27 km in circumference u  Protons are accelerated in opposite directions at 4 TeV u  Smashing together in the center of ATLAS, one of 7 experiments u  600 million collisions per second u  Data about these collisions are recorded by the

Trigger and Data Acquisition system

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 16

Page 17: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

“Private networks” now managed by IT u  Private networks for us:

Ø  isolated Networks within the ATCN §  high voltage power supplies

§  ATCA / VME crates u  Use of unregistered private networks is against CERN IT security policies u  Unmanaged switches are not supported anymore ( = no spares) u  Integration with CERN IT

Ø  standardised system Ø  no RFC1918 networks Ø  all devices on the network registered and traceable Ø  increased security Ø  increased management Ø  4 hour piquet-like support from CERN IT

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 17

Page 18: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

FlashCache tests

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 18

280

436

225

338

181 279 287

832

244

659

203

503

0 100 200 300 400 500 600 700 800 900

Random read Random write

Normal - RAID5 - Random R/W

SATA 32

SATA 64

SATA 128

SAS 32

SAS 64

SAS 128

3064

7273

2080

3996

1344 1961

3165

7227

2100

3805

1364 2013

0

1000

2000

3000

4000

5000

6000

7000

8000

Random read Random write

Flachcache -Run 1 - Random R/W

SATA 32

SATA 64

SATA 128

SAS 32

SAS 64

SAS 128

Page 19: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Sim@p1 -> TDAQ Mode u  Switching from control room shifter through a WEB gui when a LHC beam

stop longer than 24h is foreseen u  Fast and automated:

Ø  from sim@p1 to TDAQ ~ 12 min Ø  from TDAQ to sim@p1 ~ 1h Ø  emergency switch from sim@p1 to TDAQ: 100 s

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 19

Page 20: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Monitoring system upgrade Nagios v2 + Custom UI

u  Old production system: stable, but complex u  Multiple standalone Nagios servers u  Central configuration from ConfDB u  Central storage in MySQL/cluster and RRDs u  High I/O load on MySQL and NetApp can become a bottleneck u  Custom WebUI requires maintenance u  Nagios v2 is obsolete

Nagios v3 + Custom UI u  With the LFS SLC6 migration complete, this was put into production u  One single MySQL server replaced the 4 machines cluster used for storing

Nagios data Ø  better performance Ø  easier maintenance

CHEP April 2015 C.J.Lee – CERN, UJ: ATLAS TDAQ SysAdmin 20

Page 21: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Atlas Control Room u  Completion of the plan for PCoverIP migration

(Desk remote technology over TCP/IP) Ø  KVM (keyboard, video, mouse) from SDX1 to ACR over network Ø  each machine has 1 or 2 PCoverIP cards (depending on number of screens) Ø  each desk has 1 terminal client

u  A joint collaboration between OPM and TDAQ NetAdmins+SysAdmins u  Full redundancy

Ø  2 switches in SDX1 and 2 switches in SCX1. Ø  Cards and Terminals have dual connections

u  Less clutter Ø  2 optical fibers between SDX1 and SCX1, providing two independent

connections, replaced ~100 copper cables u  Delayed updating existing systems as current available market hardware provides

no redundancy and no major improvements

CHEP April 2015 C.J.Lee – CERN, UJ: ATLAS TDAQ SysAdmin 21

Page 22: ATLAS TDAQ System Administration: evolution and re-designcds.cern.ch/record/2005740/files/ATL-DAQ-SLIDE-2015-126.pdf · $ Testing Icinga 2 for distributed scheduling, configuration

Satellite Control Room u  Experts base of operations

Ø  provides similar workspaces as in ACR u  Advanced debug tools allowed u  extensive use of IMACS and Mac MINI running SLC6

Ø  allowed for long lasting, system that can be replaced

CHEP April 2015 C.J.Lee - UJ, CERN - ATLAS TDAQ SysAdmin 22

CRD (KDE) u  Control Room Desktop u  Provides tools required per “seat” in the ACR u  No direct terminal access

Ø  controlled and authenticated access to terminal windows u  This implementation needed a version of KDE not available in SLC6