- Current System Load - CPU, GPU
- Service Alerts
- Maintenance Sessions
- Previous Service Alerts
- System Status Mailings
- FAQ
- Usage statistics
Current System Load - CPU
The plot below shows the status of nodes on the current ARCHER2 Full System service. A description of each of the status types is provided below the plot.
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Note: the long running reservation visible in the plot corresponds to the short QoS which is used to support small, short jobs with fast turnaround time.
Current System Load - GPU
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Service Alerts
The ARCHER2 documentation also covers some Known Issues which users may encounter when using the system.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Ongoing | Issue | 2025-06-25 13:00 | Ongoing issues with /work1 file system | An issue with the /work1 file system is currently being investigated. This is only impacting some jobs and examples of errors seen include Cannot read/write checkpoint; corrupt file, or maybe you are out of disk space. | Under investigation |
Maintenance Sessions
This section lists recent and upcoming maintenance sessions. A full list of past maintenance sessions is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Planned | Full | 2025-08-29 12:00 | 2025-09-15 12:00 | Full ARCHER2 System | Users will not be able to connect to ARCHER2 and will not be able to access data on any of the ARCHER2 file systems. The system will be drained of jobs ahead of the power outage and jobs will not run during this period. Any queued jobs will remain in the queue during the outage and jobs will start once the service is returned. SAFE and the ARCHER2 website will be available. | Due to a significant Health and Safety risk, associated with our power supply to the site, action is required at the Advanced Computing Facility (ACF). There will be a full power outage to the site during this period. Specialised external contractors will be working on a 24/7 basis for the outage period replacing switchgear. |
Previous Service Alerts
This section lists the five most recent resolved service alerts from the past 30 days. A full list of historical resolved service alerts is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Resolved | Service Alert | 2025-07-11 10:00 | 2025-07-14 08:00 | Compute nodes | Increased queue times and reduced node availability | Update 2025-07014 All nodes have been returned to service. We are planning to remove a number of compute nodes to ensure cooling is adequate as higher temperatures are forecast in the Edinburgh area. Further details will be provided as they are available. We apologise for the inconvenience caused by longer queue times. |
Resolved | Service Alert | 2025-07-09 15:30 | 2025-07-11 08:40 | All parallel jobs launched using srun | All parallel jobs launched using `srun` will have their IO profile captured by the Darshan IO profiling tool. In rare cases this may cause jobs to fail or impact performance. Users can disable Darshan by adding the line `module remove darshan` before they use `srun` in their job submission scripts. | Capturing data on the IO use on ARCHER2 to improve the service. |
Resolved | Service Alert | 2025-06-30 13:30 | 2025-06-30 21:00 | Compute nodes | Increased queue times and reduced node availability | Update 2100 All nodes returned to service. Temperatures within cooling loop are good and the forecast will be fine over night. Update 1800 - 500 nodes remain out of service. 245 will be released at 2000 BST and 255 ndoes will be released at 2100 BST. A number of compute nodes will be removed from service to ensure cooling is adequate as higher temperatures are forecast in the Edinburgh area. The short queue is available. We apologise for the inconvenience caused by longer queue times. |
Resolved | Issue | 2025-06-26 13:00 | 2025-06-26 13:40 | Slurm controller restart between 13:00 and 14:00 | The slurm controller will be restarted at some point between 13:00 and 14:00 today, 26/06/25 and will take approximately 10 minutes to complete. Whilst this is happening users will be unable to submit jobs or query job status. | In order to try and resolve an ongoing issue |
Resolved | Service Alert | 2025-06-25 08:00 | 2025-06-25 17:10 | Compute nodes | Increased queue times and reduced node availability. Possible intermittent issues with file system or internode communication due to change in interconnect topology while cabinets and switches are unavailable. | Pump replacement on 3 cabinets which will be removed from service while pump replacement takes place. |
System Status mailings
If you would like to receive email notifications about system issues and outages, please subscribe to the System Status Notifications mailing list via SAFE
FAQ
Usage statistics
This section contains data on ARCHER2 usage for Jun 2025. Access to historical usage data is available at the end of the section.
Usage by job size and length
Queue length data
The colour indicates scheduling coefficient which is computed as [run time] divided by [run time + queue time]. A scheduling coefficient of 1 indicates that there was zero time queuing, a scheduling coefficient of 0.5 means that the job spent as long queuing as it did running.
Software usage data
Plot and table of % use and job step size statistics for different software on ARCHER2 for Jun 2025. This data is also available as a CSV file.
This table shows job step size statistics in cores weighted by usage, total number of job steps and percent usage broken down by different software for Jun 2025.
Software | Min | Q1 | Median | Q3 | Max | Jobs | Nodeh | PercentUse | Users | Projects |
---|---|---|---|---|---|---|---|---|---|---|
Overall | 1 | 512.0 | 1280.0 | 6400.0 | 131072 | 2364433 | 3896368.9 | 100.0 | 858 | 117 |
VASP | 1 | 512.0 | 768.0 | 1024.0 | 38880 | 1481290 | 769848.2 | 19.8 | 151 | 16 |
Unknown | 1 | 320.0 | 1024.0 | 6000.0 | 131072 | 291897 | 673945.1 | 17.3 | 406 | 90 |
No srun | 1 | 1024.0 | 3840.0 | 25600.0 | 128000 | 60421 | 419461.3 | 10.8 | 679 | 96 |
LAMMPS | 1 | 128.0 | 1408.0 | 3840.0 | 131072 | 60903 | 227924.9 | 5.8 | 48 | 17 |
Met Office UM | 1 | 576.0 | 1296.0 | 6840.0 | 12544 | 25228 | 222653.3 | 5.7 | 41 | 5 |
Nektar++ | 1 | 5120.0 | 8192.0 | 12800.0 | 131072 | 783 | 207827.2 | 5.3 | 13 | 4 |
SENGA | 10 | 8192.0 | 33500.0 | 33500.0 | 33500 | 67 | 187453.7 | 4.8 | 6 | 3 |
CP2K | 2 | 512.0 | 1024.0 | 1024.0 | 16384 | 73016 | 174349.1 | 4.5 | 56 | 14 |
GROMACS | 1 | 256.0 | 512.0 | 2048.0 | 12288 | 47129 | 170524.4 | 4.4 | 45 | 5 |
OpenFOAM | 1 | 640.0 | 1536.0 | 4096.0 | 25600 | 2289 | 93389.2 | 2.4 | 55 | 17 |
OpenSBLI | 4096 | 64000.0 | 64000.0 | 64000.0 | 131072 | 32 | 83719.4 | 2.1 | 3 | 2 |
FHI aims | 1 | 256.0 | 512.0 | 1024.0 | 3200 | 76587 | 83184.8 | 2.1 | 26 | 7 |
Python | 1 | 2000.0 | 4096.0 | 9216.0 | 65536 | 168948 | 77423.6 | 2.0 | 57 | 25 |
CASTEP | 4 | 256.0 | 512.0 | 2000.0 | 4200 | 9689 | 52826.8 | 1.4 | 36 | 6 |
Hydro3D | 48 | 36040.0 | 36040.0 | 73440.0 | 79560 | 87 | 40622.9 | 1.0 | 4 | 3 |
ChemShell | 1 | 256.0 | 1024.0 | 10368.0 | 12800 | 1934 | 38575.6 | 1.0 | 15 | 5 |
GENE | 1 | 8192.0 | 10240.0 | 10240.0 | 10240 | 176 | 34099.7 | 0.9 | 7 | 4 |
iIMB | 128 | 2304.0 | 2304.0 | 6400.0 | 9216 | 85 | 31832.0 | 0.8 | 2 | 2 |
Nek5000 | 2048 | 32768.0 | 65536.0 | 65536.0 | 65536 | 26 | 31171.4 | 0.8 | 3 | 2 |
EPOCH | 128 | 2304.0 | 2304.0 | 2560.0 | 4096 | 716 | 27612.4 | 0.7 | 8 | 1 |
NEMO | 1 | 480.0 | 1808.0 | 4352.0 | 8192 | 2171 | 26351.5 | 0.7 | 21 | 3 |
Quantum Espresso | 4 | 384.0 | 512.0 | 512.0 | 4864 | 29303 | 20127.4 | 0.5 | 22 | 8 |
EDAMAME | 1331 | 1331.0 | 6859.0 | 6859.0 | 6859 | 100 | 20026.8 | 0.5 | 2 | 1 |
MITgcm | 80 | 112.0 | 126.0 | 240.0 | 384 | 22152 | 17913.6 | 0.5 | 12 | 4 |
ONETEP | 1 | 128.0 | 128.0 | 128.0 | 512 | 1579 | 16328.2 | 0.4 | 8 | 3 |
WRF | 1 | 384.0 | 384.0 | 384.0 | 483 | 321 | 15873.2 | 0.4 | 5 | 2 |
Code_Saturne | 128 | 1024.0 | 3072.0 | 4096.0 | 8192 | 100 | 15237.4 | 0.4 | 6 | 2 |
Xcompact3d | 256 | 3840.0 | 3840.0 | 32768.0 | 32768 | 107 | 14635.5 | 0.4 | 4 | 1 |
CRYSTAL | 128 | 512.0 | 768.0 | 2048.0 | 2048 | 289 | 13898.3 | 0.4 | 4 | 2 |
PeleLMeX | 4 | 512.0 | 512.0 | 512.0 | 2048 | 286 | 12561.8 | 0.3 | 4 | 1 |
NAMD | 4 | 512.0 | 768.0 | 1536.0 | 2048 | 638 | 12555.6 | 0.3 | 7 | 5 |
CESM | 1 | 1536.0 | 1920.0 | 6400.0 | 6656 | 200 | 11411.2 | 0.3 | 5 | 1 |
CASINO | 16 | 1024.0 | 1024.0 | 1024.0 | 19200 | 107 | 9875.7 | 0.3 | 2 | 2 |
a.out | 1 | 1152.0 | 8192.0 | 12288.0 | 16384 | 721 | 8715.1 | 0.2 | 11 | 7 |
HYDRA | 1 | 3840.0 | 3840.0 | 4480.0 | 4480 | 67 | 6578.5 | 0.2 | 9 | 7 |
VAMPIRE | 128 | 1024.0 | 1024.0 | 8192.0 | 8192 | 304 | 6015.8 | 0.2 | 5 | 2 |
GS2 | 320 | 1280.0 | 1280.0 | 2784.0 | 4096 | 2347 | 3980.7 | 0.1 | 4 | 2 |
RMT | 128 | 1280.0 | 1408.0 | 2560.0 | 2560 | 70 | 2458.1 | 0.1 | 3 | 1 |
TPLS | 64 | 1024.0 | 2048.0 | 2048.0 | 2048 | 124 | 2229.4 | 0.1 | 3 | 1 |
ptau3d | 8 | 160.0 | 400.0 | 400.0 | 512 | 241 | 1964.6 | 0.1 | 4 | 2 |
BOUT++ | 1344 | 1344.0 | 1344.0 | 1344.0 | 1344 | 7 | 1549.0 | 0.0 | 1 | 1 |
SIESTA | 1 | 2304.0 | 2304.0 | 2304.0 | 2304 | 230 | 1456.7 | 0.0 | 4 | 3 |
NWChem | 16 | 128.0 | 512.0 | 512.0 | 1024 | 619 | 1212.8 | 0.0 | 12 | 7 |
Amber | 64 | 96.0 | 128.0 | 128.0 | 768 | 44 | 1046.8 | 0.0 | 2 | 1 |
PDNS3D | 1024 | 1024.0 | 1024.0 | 1024.0 | 1024 | 42 | 941.7 | 0.0 | 2 | 1 |
HemeLB | 4 | 768.0 | 1280.0 | 2048.0 | 4096 | 147 | 707.9 | 0.0 | 3 | 3 |
Smilei | 8 | 128.0 | 256.0 | 256.0 | 512 | 65 | 483.8 | 0.0 | 3 | 1 |
DL_POLY | 1 | 512.0 | 512.0 | 512.0 | 512 | 23 | 449.7 | 0.0 | 2 | 1 |
SPECFEM3D | 4 | 600.0 | 600.0 | 600.0 | 600 | 22 | 387.5 | 0.0 | 1 | 1 |
SBLI | 1024 | 1024.0 | 1024.0 | 1024.0 | 1024 | 40 | 333.1 | 0.0 | 1 | 1 |
PRECISE | 8 | 1920.0 | 1920.0 | 1920.0 | 1920 | 22 | 210.1 | 0.0 | 2 | 2 |
SU2 | 128 | 1280.0 | 1280.0 | 1280.0 | 1280 | 23 | 193.3 | 0.0 | 1 | 1 |
Arm Forge | 1 | 128.0 | 256.0 | 512.0 | 2048 | 481 | 175.7 | 0.0 | 6 | 5 |
ABINIT | 64 | 64.0 | 64.0 | 64.0 | 128 | 12 | 11.5 | 0.0 | 2 | 2 |
ludwig | 8 | 128.0 | 256.0 | 256.0 | 256 | 48 | 10.3 | 0.0 | 1 | 1 |
DL_MESO | 64 | 64.0 | 64.0 | 64.0 | 64 | 4 | 6.0 | 0.0 | 1 | 1 |
AxiSEM3D | 4 | 128.0 | 128.0 | 128.0 | 128 | 43 | 5.7 | 0.0 | 2 | 1 |
FVCOM | 16 | 16.0 | 16.0 | 16.0 | 16 | 20 | 3.0 | 0.0 | 1 | 1 |
CloverLeaf | 128 | 128.0 | 256.0 | 512.0 | 1024 | 11 | 0.9 | 0.0 | 1 | 2 |