praktickÝ Úvod do superpoČÍtaČe anselm infrastruktura, přístup a podpora uživatelů david...

Download PRAKTICKÝ ÚVOD DO SUPERPOČÍTAČE ANSELM Infrastruktura, přístup a podpora uživatelů David Hrbáč 2013-09-27

If you can't read please download the document

Upload: marlon-frye

Post on 15-Dec-2015

217 views

Category:

Documents


0 download

TRANSCRIPT

  • Slide 1
  • PRAKTICK VOD DO SUPERPOTAE ANSELM Infrastruktura, pstup a podpora uivatel David Hrb 2013-09-27
  • Slide 2
  • Intro What is the supercomputer Infrastructure Access to cluster Support Log-in http://www.it4i.cz/files/hrbac.pdf http://www.it4i.cz/files/hrbac.pptx
  • Slide 3
  • Why Anselm 6000 name suggestions The very first coal mine at the region The very first to have a steam engine Anselm of Canterbury
  • Slide 4
  • Early days
  • Slide 5
  • Future - Hal
  • Slide 6
  • What is a supercomputer Bunch of computers Having a lot of CPU power Having a lot of RAM Local storage Shared storage High-speed interconnected Message Process Interface
  • Slide 7
  • Supercomputer
  • Slide 8
  • Supercomputer ?!?
  • Slide 9
  • Slide 10
  • Anselm HW 209 compute nodes 3344 cores 15TB RAM 300TB /home 135TB /scratch Bull Extreme Computing Linux (RHEL clone)
  • Slide 11
  • Type of Nodes 180 compute nodes 23 GPU accelerated nodes 4 MIC accelerated nodes 2 fat nodes
  • Slide 12
  • General Node 180 nodes 2880 cores in total two Intel Sandy Bridge E5-2665, 8-core, 2.4GHz processors per node 64 GB of physical memory per node one 500GB SATA 2,5 7,2 krpm HDD per node bullx B510 blade servers cn[1-180]
  • Slide 13
  • GPU Accelerated Nodes 23 nodes 368 cores in total two Intel Sandy Bridge E5-2470, 8-core, 2.3GHz processors per node 96 GB of physical memory per node one 500GB SATA 2,5 7,2 krpm HDD per node GPU accelerator 1x NVIDIA Tesla Kepler K20 per node bullx B515 blade servers cn[181-203]
  • Slide 14
  • MIC Accelerated Nodes Intel Many Integrated Core Architecture 4 nodes 64 cores in total two Intel Sandy Bridge E5-2470, 8-core, 2.3GHz processors per node 96 GB of physical memory per node one 500GB SATA 2,5 7,2 krpm HDD per node MIC accelerator 1x Intel Phi 5110P per node bullx B515 blade servers cn[204-207]
  • Slide 15
  • Fat Node 2 nodes 32 cores in total 2 Intel Sandy Bridge E5-2665, 8-core, 2.4GHz processors per node 512 GB of physical memory per node two 300GB SAS 3,515krpm HDD (RAID1) per node two 100GB SLC SSD per node bullx R423-E3 servers cn[208-209]
  • Slide 16
  • Slide 17
  • Storage 300TB /home 135TB /scratch Infiniband 40 Gb/s Native 3600 MB/s Over TCP 1700MB/s Ethernet 114MB/s LustreFS
  • Slide 18
  • Lustre File System Clustered OSS object storage server MDS meta-data server Limits in petabytes Parallel - striped
  • Slide 19
  • Slide 20
  • Slide 21
  • Slide 22
  • Slide 23
  • Stripes Stripe count Parallel access Mind the script processes Stripe per gigabyte lfs setstripe|getstripe
  • Slide 24
  • Slide 25
  • Quotas /home 250GB /scratch no quota lfs quota u hrb33 /home
  • Slide 26
  • Slide 27
  • Access to Anselm Internal Access Call - 4x a year 3 rd round Open Access Call 2x a year 2 nd round
  • Slide 28
  • Proposals Proposals undergoing evaluation Scientific Technical Economical Primary Investigator List of co-operators
  • Slide 29
  • Login Credentials Personal certificate Signed request Credentials encrypted Login Password Ssh keys Password to the key
  • Slide 30
  • Credentials lifetime Active project(s) or affiliation to IT4Innovations Deleted 1 year after the last project Announcement 3 months before the removal 1 month before the removal 1 week before the removal
  • Slide 31
  • Support Bug tracking and trouble ticketing system Documentation IT4I internal command line tools IT4I web applications IT4I android application End-user courses
  • Slide 32
  • Main Mean Request Tracker [email protected]
  • Slide 33
  • Slide 34
  • Slide 35
  • Documentation https://support.it4i.cz/docs/anselm-cluster- documentation/ https://support.it4i.cz/docs/anselm-cluster- documentation/ Still evolving Changes almost every day
  • Slide 36
  • Slide 37
  • IT4I internal command line tools It4free Rspbs Licenses allocation Internal in-house scripts Automation to handle the credentials Cluster automation PBS accounting
  • Slide 38
  • IT4I web applications Internal information system Project management Project accounting User management Cluster monitoring
  • Slide 39
  • Slide 40
  • Slide 41
  • IT4I android application Internal tool Considering the release to end-users Features News Graphs Feature requests Accounting Support Nodes allocation Jobs status
  • Slide 42
  • Slide 43
  • Log-in to Anselm Finally! Ssh protocol Via anselm.it4i.cz login1.anselm.it4i.cz login2.anselm.it4i.cz
  • Slide 44
  • Slide 45
  • Slide 46
  • Slide 47
  • Slide 48
  • Slide 49
  • Slide 50
  • Slide 51
  • Slide 52
  • Slide 53
  • VNC ssh anselm L 5961:localhost:5961 Remmina Vncviewer 127.0.0.1:5961
  • Slide 54
  • Links https://support.it4i.cz/docs/anselm-cluster- documentation/ https://support.it4i.cz/docs/anselm-cluster- documentation/ https://support.it4i.cz/ https://www.it4i.cz/
  • Slide 55
  • Questions Thank you.