The computing cluster Castor is accessible via ssh from the frontend as  The operating system on the login nodes is Red Hat Enterprise Linux Server 6.6: direct access to the compute nodes is not allowed, to run jobs you need to use the SLURM batch queueing system (see RUNNING JOBS).



IBM iDataPlex

32 Compute Nodes

2 x Intel® Xeon® CPU X5650 @ 2.60GHz (12 cores, 24 to 96GB RAM) + 2 GPU Nvidia Tesla M2090

2 Login Nodes

2 x Intel® Xeon® CPU E5620 @ 2.40GHz (8 cores, 48GB RAM)

Theoretical Peak Performance


Memory Capacity per node

24,48,96 GB (DDR3-1300)

Memory Bandwidth per node

41.6 GB/s

Total System Memory

1.5 TB DDR3

Interconnect Configuration

1 high-speed interconnect based on InfiniBand FDR (used for both MPI traffic and high-speed storage traffic)

Scratch capacity

/scratch/castor 640TB