Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...



Booster

Data Centric

Model

Atos BullSequana XH21355 "Da Vinci" blade

Atos BullSequana X2610 compute blade

Racks

150

Nodes

3456

1536

Processors

32 cores Intel Ice Lake 

Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz

56 cores Intel Sapphire Rapids

Accelerators

4 x NVIDIA Ampere GPUs/node, 64GB HBM2

-

Cores

32 cores/node

56 cores/node

RAM

512 (8x64) GB DDR4 3200 MHz

(16 x 32) GB DDR5 4800 MHz

Peak Performance

about 309 Pflop/s

9 Pflops/s

Internal Network

NVIDIA Mellanox HDR DragonFly++ 200Gb/s

2 x NVIDIA HDR 2×100 Gb/s cards                                                            1x Nvidia HDR100 100 Gb/s card                                                                       

Disk Space

106PB Large capacity storage

5.4 PB of High performance storage







The β access phase

Leonardo is currently under testing by the vendor team, CINECA staff, and authorized external users. The environment is not finalized yet on storage, system configuration, and software stack.  

Software environment and test jobs

The available software environment is based on spack and modules, and needs to be activated. Some vendor installations are also available and presented in an lmod environment on the login node, but we warmly encourage the beta testers to use the spack environment to provide a valuable feedback on the software stack provided by CINECA. 

The production environment will be based on the slurm scheduler, already in place on the cluster but in a very preliminary configuration. The only available partition is "prod" (#SBATCH --partition=prod).  Please refer to the general online guide to slurm and on task/thread bindings, and please pay attention to the setting of the SRUN_CPUS_PER_TASK for hybrid applications dispatched with "srun". In this preliminary configuration, please explicit the request of the correct pmix plugin when launching your parallel applications with "srun":

srun --mpi=pmix_v3 <options>  <exe>

No pmi settings are needed if you launch with "mpirun".