- Current System Load - CPU, GPU
- Service Alerts
- Maintenance Sessions
- Previous Service Alerts
- System Status Mailings
- FAQ
- Usage statistics
Current System Load - CPU
The plot below shows the status of nodes on the current ARCHER2 Full System service. A description of each of the status types is provided below the plot.
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Note: the long running reservation visible in the plot corresponds to the short QoS which is used to support small, short jobs with fast turnaround time.
Current System Load - GPU
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Service Alerts
The ARCHER2 documentation also covers some Known Issues which users may encounter when using the system.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Ongoing | Service Alert | 2025-06-30 13:36 | Compute nodes | Increased queue times and reduced node availability | A number of compute nodes will be shut down to ensure cooling is adequate as higher temperatures are forecast in the Edinburgh area. The short queue will not be available at the moment. We apologise for the inconvenience caused by longer queue times. | |
Ongoing | Issue | 2025-06-25 13:00 | Ongoing issues with /work1 file system | An issue with the /work1 file system is currently being investigated. This is only impacting some jobs and examples of errors seen include Cannot read/write checkpoint; corrupt file, or maybe you are out of disk space. | Under investigation |
Maintenance Sessions
This section lists recent and upcoming maintenance sessions. A full list of past maintenance sessions is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Planned | Full | 2025-08-29 12:00 | 2025-09-15 12:00 | Full ARCHER2 System | Users will not be able to connect to ARCHER2 and will not be able to access data on any of the ARCHER2 file systems. The system will be drained of jobs ahead of the power outage and jobs will not run during this period. Any queued jobs will remain in the queue during the outage and jobs will start once the service is returned. SAFE and the ARCHER2 website will be available. | Due to a significant Health and Safety risk, associated with our power supply to the site, action is required at the Advanced Computing Facility (ACF). There will be a full power outage to the site during this period. Specialised external contractors will be working on a 24/7 basis for the outage period replacing switchgear. |
Previous Service Alerts
This section lists the five most recent resolved service alerts from the past 30 days. A full list of historical resolved service alerts is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Resolved | Issue | 2025-06-26 13:00 | 2025-06-26 13:40 | Slurm controller restart between 13:00 and 14:00 | The slurm controller will be restarted at some point between 13:00 and 14:00 today, 26/06/25 and will take approximately 10 minutes to complete. Whilst this is happening users will be unable to submit jobs or query job status. | In order to try and resolve an ongoing issue |
Resolved | Service Alert | 2025-06-25 08:00 | 2025-06-25 17:10 | Compute nodes | Increased queue times and reduced node availability. Possible intermittent issues with file system or internode communication due to change in interconnect topology while cabinets and switches are unavailable. | Pump replacement on 3 cabinets which will be removed from service while pump replacement takes place. |
Resolved | Service Alert | 2025-06-20 11:00 | 2025-06-23 08:30 | 40% - 50% Compute nodes | Increased queue times and reduced node availability | Update An issue was identified which caused the temperature to remain high. This has now been resolved and nodes are being returned to serivce. We have powered down 40% of compute nodes to ensure cooling is adequate due to unusually high temperatures in the Edinburgh area. We have placed a further 10% of compute nodes into maintenance mode. The maintenance will be lifted at 2200 on Saturday 21st June if cooling is at a good level. A further assessment will take place at 0900 on Sunday 22nd June and if possible nodes will be returned to service. We apologise for the inconvenience caused by longer queue times. |
Resolved | Service Alert | 2025-06-05 10:00 | 2025-06-05 10:45 | ARCHER2 Slurm scheduler | ARCHER2 Slurm Controller will be restarted. Running jobs will continue to run, but Slurm commands will be unavailable for a few minutes. | Update to the Slurm configuration. |
System Status mailings
If you would like to receive email notifications about system issues and outages, please subscribe to the System Status Notifications mailing list via SAFE
FAQ
Usage statistics
This section contains data on ARCHER2 usage for May 2025. Access to historical usage data is available at the end of the section.
Usage by job size and length
Queue length data
The colour indicates scheduling coefficient which is computed as [run time] divided by [run time + queue time]. A scheduling coefficient of 1 indicates that there was zero time queuing, a scheduling coefficient of 0.5 means that the job spent as long queuing as it did running.
Software usage data
Plot and table of % use and job step size statistics for different software on ARCHER2 for May 2025. This data is also available as a CSV file.
This table shows job step size statistics in cores weighted by usage, total number of job steps and percent usage broken down by different software for May 2025.
Software | Min | Q1 | Median | Q3 | Max | Jobs | Nodeh | PercentUse | Users | Projects |
---|---|---|---|---|---|---|---|---|---|---|
Overall | 1 | 576.0 | 2048.0 | 6500.0 | 524288 | 1284694 | 4449322.2 | 100.0 | 887 | 129 |
Unknown | 1 | 600.0 | 2816.0 | 6000.0 | 99104 | 556969 | 822754.8 | 18.5 | 411 | 100 |
VASP | 1 | 512.0 | 1024.0 | 1280.0 | 10240 | 70374 | 772565.3 | 17.4 | 160 | 17 |
Nektar++ | 1 | 6400.0 | 8192.0 | 20480.0 | 64000 | 1074 | 312503.3 | 7.0 | 14 | 4 |
No srun | 1 | 1536.0 | 6400.0 | 23040.0 | 524288 | 49631 | 310376.6 | 7.0 | 656 | 100 |
SENGA | 10 | 5120.0 | 33500.0 | 33500.0 | 33500 | 218 | 222929.4 | 5.0 | 7 | 3 |
GROMACS | 1 | 1024.0 | 2048.0 | 2560.0 | 12288 | 11797 | 217634.5 | 4.9 | 44 | 6 |
Met Office UM | 1 | 1024.0 | 1296.0 | 6165.0 | 12544 | 28063 | 202134.9 | 4.5 | 42 | 5 |
LAMMPS | 1 | 128.0 | 1536.0 | 3840.0 | 131072 | 66700 | 190865.3 | 4.3 | 51 | 16 |
CP2K | 1 | 256.0 | 512.0 | 1024.0 | 3072 | 63673 | 175080.9 | 3.9 | 47 | 13 |
OpenFOAM | 1 | 512.0 | 2048.0 | 5120.0 | 93750 | 3462 | 160177.4 | 3.6 | 52 | 18 |
Python | 1 | 4096.0 | 4096.0 | 9216.0 | 32768 | 103279 | 134839.7 | 3.0 | 64 | 28 |
FHI aims | 1 | 512.0 | 1024.0 | 5375.0 | 6875 | 58548 | 107131.5 | 2.4 | 26 | 6 |
CASTEP | 16 | 152.0 | 512.0 | 1440.0 | 4096 | 8922 | 84070.1 | 1.9 | 42 | 7 |
OpenSBLI | 128 | 25600.0 | 25600.0 | 64000.0 | 131072 | 129 | 73212.3 | 1.6 | 4 | 3 |
Xcompact3d | 128 | 2048.0 | 32768.0 | 32768.0 | 32768 | 1194 | 58040.1 | 1.3 | 8 | 3 |
Quantum Espresso | 4 | 256.0 | 512.0 | 512.0 | 6144 | 15252 | 56050.1 | 1.3 | 26 | 10 |
ChemShell | 1 | 1024.0 | 5760.0 | 12800.0 | 12800 | 1042 | 55008.7 | 1.2 | 10 | 4 |
VAMPIRE | 24 | 16384.0 | 65536.0 | 65536.0 | 65536 | 458 | 48036.0 | 1.1 | 6 | 2 |
Code_Saturne | 1 | 4096.0 | 524288.0 | 524288.0 | 524288 | 82 | 44560.7 | 1.0 | 6 | 3 |
CRYSTAL | 1 | 1024.0 | 131072.0 | 131072.0 | 131072 | 185 | 39920.7 | 0.9 | 3 | 3 |
iIMB | 128 | 2048.0 | 2304.0 | 6400.0 | 6400 | 90 | 31838.8 | 0.7 | 2 | 2 |
CASINO | 128 | 1024.0 | 1024.0 | 2048.0 | 2048 | 347 | 26188.9 | 0.6 | 2 | 2 |
ONETEP | 1 | 128.0 | 128.0 | 150.0 | 2112 | 1772 | 25370.6 | 0.6 | 6 | 1 |
Smilei | 1 | 512.0 | 512.0 | 512.0 | 768 | 201 | 24453.4 | 0.5 | 4 | 1 |
MITgcm | 18 | 112.0 | 126.0 | 384.0 | 3072 | 20651 | 23269.6 | 0.5 | 13 | 3 |
SIESTA | 1 | 2304.0 | 2304.0 | 2304.0 | 2304 | 353 | 22323.3 | 0.5 | 2 | 1 |
NAMD | 4 | 512.0 | 640.0 | 768.0 | 4096 | 6879 | 20778.2 | 0.5 | 7 | 4 |
EDAMAME | 64 | 6859.0 | 6859.0 | 6859.0 | 6859 | 103 | 20704.7 | 0.5 | 2 | 1 |
SU2 | 1280 | 6400.0 | 6400.0 | 6400.0 | 12800 | 25 | 16477.0 | 0.4 | 1 | 1 |
Nek5000 | 4 | 2048.0 | 4096.0 | 8192.0 | 32768 | 127 | 16157.1 | 0.4 | 3 | 2 |
NEMO | 1 | 340.0 | 480.0 | 1320.0 | 8192 | 4007 | 14815.2 | 0.3 | 22 | 4 |
WRF | 1 | 384.0 | 384.0 | 384.0 | 384 | 159 | 14531.5 | 0.3 | 4 | 2 |
CESM | 1 | 3840.0 | 6656.0 | 6656.0 | 6784 | 344 | 12824.3 | 0.3 | 6 | 1 |
BOUT++ | 1344 | 1344.0 | 1344.0 | 1344.0 | 1344 | 52 | 11524.9 | 0.3 | 1 | 1 |
PRECISE | 8 | 1920.0 | 1920.0 | 1920.0 | 1920 | 76 | 10387.6 | 0.2 | 2 | 2 |
PeleLMeX | 128 | 512.0 | 512.0 | 512.0 | 1536 | 140 | 10345.3 | 0.2 | 2 | 1 |
GS2 | 256 | 1280.0 | 1280.0 | 1280.0 | 1664 | 8405 | 9681.6 | 0.2 | 3 | 1 |
EPOCH | 128 | 512.0 | 2304.0 | 2304.0 | 3840 | 409 | 9131.2 | 0.2 | 9 | 1 |
HYDRA | 1 | 3840.0 | 3840.0 | 3840.0 | 4480 | 127 | 8252.1 | 0.2 | 8 | 6 |
GENE | 1 | 8192.0 | 8192.0 | 8192.0 | 8192 | 118 | 7510.5 | 0.2 | 10 | 3 |
a.out | 1 | 256.0 | 1280.0 | 4096.0 | 8192 | 1998 | 6657.1 | 0.1 | 7 | 4 |
RMT | 640 | 640.0 | 1152.0 | 1280.0 | 5120 | 193 | 3783.2 | 0.1 | 3 | 1 |
Hydro3D | 9 | 468.0 | 1440.0 | 1830.0 | 32640 | 159 | 2778.7 | 0.1 | 3 | 2 |
NWChem | 1 | 128.0 | 128.0 | 1024.0 | 1024 | 196220 | 2478.3 | 0.1 | 9 | 5 |
TPLS | 1 | 4096.0 | 4096.0 | 4096.0 | 4096 | 91 | 2291.6 | 0.1 | 2 | 1 |
Amber | 8 | 640.0 | 768.0 | 768.0 | 1536 | 43 | 1898.4 | 0.0 | 3 | 2 |
PDNS3D | 1024 | 1024.0 | 1024.0 | 1024.0 | 1024 | 24 | 1633.6 | 0.0 | 1 | 1 |
HemeLB | 256 | 1024.0 | 1024.0 | 1024.0 | 1024 | 26 | 869.3 | 0.0 | 2 | 1 |
FDS | 248 | 384.0 | 384.0 | 384.0 | 384 | 14 | 864.2 | 0.0 | 1 | 1 |
SBLI | 128 | 1024.0 | 1024.0 | 1024.0 | 2048 | 30 | 512.3 | 0.0 | 1 | 1 |
ptau3d | 8 | 160.0 | 400.0 | 400.0 | 400 | 42 | 462.4 | 0.0 | 3 | 1 |
CPMD | 128 | 128.0 | 128.0 | 128.0 | 128 | 24 | 195.3 | 0.0 | 1 | 1 |
ABINIT | 128 | 128.0 | 128.0 | 128.0 | 128 | 32 | 168.8 | 0.0 | 1 | 1 |
FVCOM | 16 | 128.0 | 384.0 | 384.0 | 384 | 35 | 136.4 | 0.0 | 3 | 2 |
DL_POLY | 16 | 32.0 | 32.0 | 32.0 | 32 | 55 | 45.3 | 0.0 | 1 | 1 |
OSIRIS | 512 | 512.0 | 512.0 | 512.0 | 512 | 1 | 41.7 | 0.0 | 1 | 1 |
Arm Forge | 1 | 96.0 | 192.0 | 496.0 | 1280 | 197 | 40.2 | 0.0 | 15 | 10 |
ludwig | 8 | 512.0 | 512.0 | 512.0 | 512 | 63 | 3.9 | 0.0 | 2 | 2 |
DL_MESO | 128 | 128.0 | 128.0 | 128.0 | 128 | 1 | 2.7 | 0.0 | 1 | 1 |
AxiSEM3D | 128 | 128.0 | 128.0 | 128.0 | 128 | 9 | 0.8 | 0.0 | 1 | 1 |