Navigation

HPC System Utilization

User projects

From time to time RRZE asks its HPC customers to provide a short report of the work they are doing on the HPC systems. You can find an extensive list of ongoing and past user projects on our User Project Page.

HPC users and usage

pie chart with HPC usage

weighted HPC usage (year 2017)

RRZE operates three big parallel computers (LiMa – shutdown end of 2018, Emmy, Meggie), a throughput cluster (Woody/TinyEth) and three specialized systems (TinyFAT, TinyGPU, Windows HPC – shutdown in June 2018). In total there are more than 1’800 nodes with 30’000 cores and almost 100 TB of (distributed) main memory as well as more than 2.0 PB of disk storage in five storage systems plus a transparently managed tape library for offline-storage.

In the year 2017 more than 550 accounts from almost 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total about 200 million core hours have been delivered to 1.7 million jobs.

The year 2018 was marked by a significant extension of storage capacity through a “shareholder NFS server”. TinyGPU has been extended by several user groups to include up-to-date (consumer) GPUs. But we also had to shutdown two systems (LiMa and Windows HPC). Both were up and running for more than eight years. Over the years, LiMa served more than one thousand users and delivered almost 300 million core hours to 2.6 million jobs.

In the year 2018 almost 600 accounts from more than 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total more than 200 million core hours have been delivered to almost 1.6 million jobs.

Roofline diagram of all tracked jobs on Meggie in the year 2019.

In the year 2019 more about 530 accounts from 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes again scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total about 215 million core hours have been delivered to more than 1.2 million jobs.

Only “minor” hardware extensions occurred during the year 2019: 12 additional nodes with 48 NVidia RZX2080Ti GPUs for TinyGPU (+51%; all financed by individual user groups), yet an other “shareholder NFS server” (+450 TB), and finally an addition of 112 nodes to the Woody throughput cluster (+64%) financed by the ending excellence cluster EAM and an upcoming CRC. The grand proposal for FAU’s next big parallel cluster (4.5 mio EUR) is under review by DFG and we hope to get the approval in early 2020.

In the year 2020 our DFG application was approved and we succeeded with our NHR application. The tender for a big procurement (8 Mio EUR for a parallel computer and GPGPU cluster) was started in autumn. Several groups again financed significant extensions of TinyGPU and TinyFAT. These new nodes also mark the transition from Ubuntu 18.04 to Ubuntu 20.04 and from Torque/maui to Slurm as batch system. Moreover, the HPC storage system for $HOME and $VAULT has been replaced during summer 2020 – significantly increasing the online disk capacity to several PetaBytes.

The number of active accounts and groups further increased to 738 and 83, respectively. This includes again scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total about 213 million core hours have been delivered to more than 1.6 million jobs.

 

Graphs of the current utilization of RRZE’s HPC systems can be found at http://www.statistiken.rrze.de/hpc/auslastung.shtml.