MEMORY INTENSIVE
SERVICE (DURHAM)

MEMORY INTENSIVE at durhaM

System names: COSMA7, COSMA8

The workflow prototype underlying the design of the Memory Intensive Service (COSMA 7 and 8) is large-scale cosmological simulation of the formation and evolution of galaxies in the universe. Such simulations are extremely challenging due to the complex interplay between diverse physical processes operating on length-scales ranging from 10⁷ km to 10²² km, and timescales ranging from hours to tens of billions of years. To enable this calculational complexity, COSMA 8 has 1TB of memory on each node – by mid-2023 the total memory footprint will be an impressive 528 TB — ensuring that each node holds enough data to keep its 128 AMD cores busy. Even so, the nodes need to communicate efficiently, and a non-blocking InfiniBand network is essential. 

However, it’s not enough simply to carry out lots of calculations – we also need to save the current state of the simulation by copying every detail from the node memory into the storage system. While previous checkpoints took over 12 hours, COSMA 8’s large volume of flash-based storage makes it possible to perform a complete 528TB checkpoint in under 10 minutes! This time-saving by itself makes large-scale calculations roughly 10% more efficient.

Finally, we need somewhere to keep the output data – COSMA has 14PB of spinning disk storage, and an additional 26PB of tape storage for those data not actively in use.

The careful design matching the needs of simulation workflows, combined with expert support and a flexible allocation model, makes it possible to reserve up to 80% of the COSMA system for a single calculation making it the capability system of choice for cosmology. Due to its greater efficiency and flexibility, it has been used in preference to European systems with four times as many cores.

COSMA7

COSMA7 COMPRISES
2x 1.5TB AND 1 x 768GB LOGIN NODES WITH INTEL XEON 5120 SKYLAKE PROCESSORS, 2.2GHz, 28 CORES
452 COMPUTE NODES, EACH WITH 460TB OF RAM AND 2 x X5120 2.2GHz PER NODE, OFFERING A TOTAL OF 12656 CORES
HALF OF THE SYSTEM IS CONNECTED VIA MELLANOX EDR 11GBIT/S IN A 2:1 BLOCKING FAT TREE CONFIGURATION, WHILE THE OTHER HALF IS CONNECTED WITH A ROCKPORT 100GBIT/S 6D TORUS NETWORK FABIC SPACE AND 3.1PB OF DATA SPACE ON LUSTRE

DATA MANAGEMENT PLAN

COSMA8

COSMA8 COMPRISES
528 COMPUPTE NODES. (360 WITH 2x AMD ROME PROCESSORS AND 168 WITH AMD MILAN 7763 PROCESSORS), 1TB RAM AND NON-BLOOCKING HDR200 INFINIBAND NETWORK
2x 2TB LOGIN NODES WITH 64 CORES (DUAL AMD 7h12 ROME 7542 PROCESSORS)
TWO FAT NODES WITH 4TB RAM AND 128 CORES
GPU NODES WITH NVIDIA A100, V100 AND AMD MI200 AND MI100 GPUs – 7 NODES WITH 18 GPUs13PB BULK LUSTRE STORAGE
1.2 FAST SCRATCH STORAGE (~350GBytes/s)
HIGH-CACHE NODE (768MB CACHE MILAN-X) WITH 127 CORES AND 1TB RAM
26PB TAPE ARCHIVAL FACILITY
 
THE BENCHMARK CODES USED DURING THE DESIGN PHASE OF COSMA8 WERE:
SWIFT – THE UK-DEVELOPED COSMOLOGICAL SIMULATION CODE, WHICH USES TASK-BASED PARALLELISM TO MAXIMISE EFFICIENCY. 
BAM – USED TO SIMULATE THE GRAVITATIONAL WAVES SIGNALS GENERATED BY THE MERGERS OF BLACK HOLES AND NEUTRON STARS. A PREDECESSOR OF COSMA8 WAS USED TO CARRY OUT SIMULATIONS THAT SUPPORTED THE DISCOVER OF GRAVITATIONAL WAVES IN 2016.
 

SITE SPECIFIC USER GUIDE

Our site spcific user guide, hosted by Durham University, contains a full user guide for COSMA8

SCIENCE ON MEMORY INTENSIVE SERVICE (DURHAM)