Navigation

HPC System Utilization

User projects

From time to time RRZE asks its HPC customers to provide a short report of the work they are doing on the HPC systems. You can find an extensive list of ongoing and past user projects on our User Project Page.

HPC users and usage

pie chart with HPC usage

weighted HPC usage (year 2017)

RRZE operates three big parallel computers (LiMa – shutdown end of 2018, Emmy, Meggie), a throughput cluster (Woody/TinyEth) and three specialized systems (TinyFAT, TinyGPU, Windows HPC – shutdown in June 2018). In total there are more than 1’800 nodes with 30’000 cores and almost 100 TB of (distributed) main memory as well as more than 2.0 PB of disk storage in five storage systems plus a transparently managed tape library for offline-storage.

In the year 2017 more than 550 accounts from almost 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total about 200 million core hours have been delivered to 1.7 million jobs.

The year 2018 was marked by a significant extension of storage capacity through a “shareholder NFS server”. TinyGPU has been extended by several user groups to include up-to-date (consumer) GPUs. But we also had to shutdown two systems (LiMa and Windows HPC). Both were up and running for more than eight years. Over the years, LiMa served more than one thousand users and delivered almost 300 million core hours to 2.6 million jobs.

In the year 2018 almost 600 accounts from more than 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total more than 200 million core hours have been delivered to almost 1.6 million jobs.

Roofline diagram of all tracked jobs on Meggie in the year 2019.

In the year 2019 more about 530 accounts from 70 groups have been active (i.e. consumed CPU cycles) on RRZE’s HPC systems. This includes again scientists from all five faculties of the University, students doing labs as part of their studies or for their final bachelor or master thesis, as well as a few users from regional universities and colleges or external collaborators. In total about 215 million core hours have been delivered to more than 1.2 million jobs.

Only “minor” hardware extensions occurred during the year 2019: 12 additional nodes with 48 NVidia RZX2080Ti GPUs for TinyGPU (+51%; all financed by individual user groups), yet an other “shareholder NFS server” (+450 TB), and finally an addition of 112 nodes to the Woody throughput cluster (+64%) financed by the ending excellence cluster EAM and an upcoming CRC. The grand proposal for FAU’s next big parallel cluster (4.5 mio EUR) is under review by DFG and we hope to get the approval in early 2020.

Graphs of the current utilization of RRZE’s HPC systems can be found at http://www.statistiken.rrze.de/hpc/auslastung.shtml.