- Current System Load - Full System
- Service Alerts
- Maintenance Sessions
- Previous Service Alerts
- System Status Mailings
- FAQ
- Usage statistics
Current System Load - CPU
The plot below shows the status of nodes on the current ARCHER2 Full System service. A description of each of the status types is provided below the plot.
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Note: the long running reservation visible in the plot corresponds to the short QoS which is used to support small, short jobs with fast turnaround time.
Current System Load - GPU
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Service Alerts
The ARCHER2 documentation also covers some Known Issues which users may encounter when using the system.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Ongoing | Service Alert | 2025-02-03 08:00 | 2025-02-03 18:00 | RDFaaS (file systems /epsrc and /general) | Users will not be able to access files within /epsrc and /general | Essential work on the RDFaaS hardware |
Ongoing | Service Alert | 2025-01-24 10:00 | 2025-01-24 17:00 | Whole ARCHER2 service | Service at higher risk of disruption than usual. If issues arise, service may take longer to restore. | Red weather warning for high winds in Edinburgh area lead to travel restrictions and higher than usual risk of power/building damage issues. |
Ongoing | Service Alert | 2025-01-24 12:45 | 2025-01-27 10:00 | ARCHER2 compute nodes | All ARCHER2 compute nodes are unavailable. Jobs running at time of incident will have failed. No new jobs will start and Short QoS is unavailable. | Extreme weather in Edinburgh area has interrupted power supply at ACF data centre. Compute nodes will remain unavailable until extreme weather has passed and staff are able to restore them to service. Next update will be on Monday morning when staff can visit the site. |
Maintenance Sessions
This section lists recent and upcoming maintenance sessions. A full list of past maintenance sessions is available.
No scheduled or recent maintenance sessions
Previous Service Alerts
This section lists the five most recent resolved service alerts from the past 30 days. A full list of historical resolved service alerts is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Resolved | Service Alert | 2025-01-22 09:15 | 2025-01-22 13:15 | All parallel jobs launched using srun | All parallel jobs launched using `srun` will have their IO profile captured by the Darshan IO profiling tool. In rare cases this may cause jobs to fail or impact performance. Users can disable Darshan by adding the line `module remove darshan` before they use `srun` in their job submission scripts. | Capturing data on the IO use on ARCHER2 to improve the service. |
Resolved | Service Alert | 2025-01-22 14:45 | 2025-01-22 18:00 | ARCHER2 login, compute and data analysis nodes | No login access, jobs running at time of failure will have failed | Failure on solid state Lustre file system |
Resolved | Service Alert | 2025-01-15 11:45 | 2025-01-16 09:20 | ARCHER2 login, compute and data analysis nodes | No login access, jobs running at time of failure will have failed | Failure on solid state Lustre file system |
Resolved | Service Alert | 2025-01-16 13:00 | 2025-01-16 14:00 | ARCHER2 software modules | Changes to module loaction, remove old versions, update default versions | Periodic software update |
Resolved | Service Alert | 2025-01-09 12:30 | 2025-01-10 08:10 | a2fs-work2 Lustre file system | A hardware issue has been identified on part of the ARCHER2 work2 filesystem. No jobs allowed to start for projects hosted on this file system. | Hardware issue with work2 filesystem |
System Status mailings
If you would like to receive email notifications about system issues and outages, please subscribe to the System Status Notifications mailing list via SAFE
FAQ
Usage statistics
This section contains data on ARCHER2 usage for Dec 2024. Access to historical usage data is available at the end of the section.
Usage by job size and length
Queue length data
The colour indicates scheduling coefficient which is computed as [run time] divided by [run time + queue time]. A scheduling coefficient of 1 indicates that there was zero time queuing, a scheduling coefficient of 0.5 means that the job spent as long queuing as it did running.
Software usage data
Plot and table of % use and job step size statistics for different software on ARCHER2 for Dec 2024. This data is also available as a CSV file.
This table shows job step size statistics in cores weighted by usage, total number of job steps and percent usage broken down by different software for Dec 2024.
Software | Min | Q1 | Median | Q3 | Max | Jobs | Nodeh | PercentUse | Users | Projects |
---|---|---|---|---|---|---|---|---|---|---|
Overall | 1 | 512.0 | 2048.0 | 6400.0 | 262144 | 1345274 | 2822560.1 | 100.0 | 831 | 121 |
Unknown | 1 | 512.0 | 4096.0 | 8192.0 | 67500 | 358193 | 590178.1 | 20.9 | 378 | 89 |
VASP | 1 | 512.0 | 1260.0 | 2304.0 | 25600 | 66595 | 379432.6 | 13.4 | 126 | 15 |
Met Office UM | 26 | 1024.0 | 1024.0 | 6165.0 | 12544 | 28940 | 296292.6 | 10.5 | 33 | 2 |
No srun | 1 | 1024.0 | 2560.0 | 10496.0 | 131072 | 45045 | 212399.3 | 7.5 | 574 | 87 |
OpenFOAM | 1 | 512.0 | 2048.0 | 12800.0 | 67500 | 5415 | 145116.7 | 5.1 | 49 | 17 |
GROMACS | 1 | 640.0 | 1536.0 | 2560.0 | 12800 | 4500 | 139827.7 | 5.0 | 42 | 4 |
Python | 1 | 9216.0 | 16384.0 | 25600.0 | 36864 | 504301 | 121471.4 | 4.3 | 61 | 23 |
Nektar++ | 5 | 5120.0 | 6400.0 | 8960.0 | 16640 | 614 | 114791.6 | 4.1 | 11 | 3 |
CP2K | 1 | 512.0 | 512.0 | 1024.0 | 64000 | 22314 | 111590.5 | 4.0 | 44 | 11 |
Code_Saturne | 128 | 2048.0 | 32768.0 | 32768.0 | 32768 | 190 | 86312.3 | 3.1 | 6 | 3 |
LAMMPS | 1 | 256.0 | 640.0 | 3840.0 | 262144 | 7235 | 77861.7 | 2.8 | 53 | 16 |
GENE | 1 | 4096.0 | 8192.0 | 10240.0 | 10240 | 284 | 75298.4 | 2.7 | 10 | 2 |
CASTEP | 1 | 128.0 | 384.0 | 400.0 | 4096 | 174583 | 49892.8 | 1.8 | 38 | 8 |
iIMB | 128 | 3072.0 | 6400.0 | 6400.0 | 10752 | 136 | 42752.5 | 1.5 | 3 | 2 |
SENGA | 1500 | 5120.0 | 5120.0 | 5120.0 | 36848 | 40 | 40320.2 | 1.4 | 5 | 4 |
Quantum Espresso | 1 | 256.0 | 512.0 | 1152.0 | 2048 | 15460 | 38961.5 | 1.4 | 17 | 6 |
PeleLMeX | 128 | 512.0 | 512.0 | 2048.0 | 4096 | 875 | 27783.4 | 1.0 | 4 | 1 |
VAMPIRE | 1 | 2048.0 | 2048.0 | 2048.0 | 2048 | 2272 | 27647.5 | 1.0 | 5 | 3 |
FHI aims | 8 | 384.0 | 768.0 | 1792.0 | 2816 | 2158 | 26454.6 | 0.9 | 19 | 3 |
ChemShell | 1 | 1024.0 | 1024.0 | 1024.0 | 5376 | 884 | 24522.8 | 0.9 | 16 | 4 |
RMT | 256 | 2432.0 | 2432.0 | 2432.0 | 2432 | 361 | 24461.9 | 0.9 | 5 | 1 |
NEMO | 1 | 768.0 | 1890.0 | 2912.0 | 10080 | 8920 | 21218.5 | 0.8 | 21 | 3 |
MITgcm | 28 | 112.0 | 126.0 | 240.0 | 512 | 20040 | 15771.1 | 0.6 | 11 | 3 |
a.out | 1 | 4096.0 | 4096.0 | 5120.0 | 5120 | 366 | 12241.3 | 0.4 | 7 | 6 |
CASINO | 640 | 2560.0 | 5120.0 | 5120.0 | 5120 | 46 | 12106.3 | 0.4 | 1 | 1 |
Nek5000 | 5 | 1280.0 | 1664.0 | 2176.0 | 2560 | 175 | 10755.2 | 0.4 | 1 | 1 |
WRF | 50 | 384.0 | 384.0 | 384.0 | 384 | 155 | 10629.7 | 0.4 | 4 | 2 |
EDAMAME | 1331 | 1331.0 | 1331.0 | 1331.0 | 1331 | 115 | 9873.0 | 0.3 | 2 | 1 |
Xcompact3d | 128 | 512.0 | 1024.0 | 6144.0 | 6144 | 259 | 9841.8 | 0.3 | 7 | 4 |
HYDRA | 1 | 3840.0 | 3840.0 | 3840.0 | 8192 | 271 | 7307.6 | 0.3 | 9 | 5 |
EPOCH | 512 | 1024.0 | 2048.0 | 2048.0 | 4096 | 426 | 7289.4 | 0.3 | 5 | 1 |
ONETEP | 8 | 64.0 | 64.0 | 150.0 | 384 | 66721 | 6287.0 | 0.2 | 7 | 2 |
OpenSBLI | 384 | 8192.0 | 8192.0 | 8192.0 | 131072 | 57 | 5836.9 | 0.2 | 3 | 2 |
CRYSTAL | 128 | 128.0 | 128.0 | 768.0 | 768 | 276 | 4718.3 | 0.2 | 3 | 2 |
BOUT++ | 768 | 768.0 | 1344.0 | 1344.0 | 1344 | 24 | 4506.4 | 0.2 | 1 | 1 |
NWChem | 4 | 128.0 | 128.0 | 128.0 | 384 | 4699 | 4342.6 | 0.2 | 8 | 4 |
3DNS | 6930 | 17680.0 | 17680.0 | 17680.0 | 17680 | 2 | 3863.6 | 0.1 | 1 | 1 |
CESM | 1 | 768.0 | 768.0 | 1280.0 | 1280 | 593 | 3604.0 | 0.1 | 9 | 2 |
GS2 | 256 | 1536.0 | 2048.0 | 2048.0 | 2048 | 28 | 3594.4 | 0.1 | 3 | 2 |
ptau3d | 32 | 800.0 | 800.0 | 800.0 | 800 | 12 | 3517.0 | 0.1 | 1 | 1 |
Hydro3D | 9800 | 9800.0 | 9800.0 | 9800.0 | 9800 | 6 | 2437.6 | 0.1 | 1 | 1 |
SU2 | 128 | 512.0 | 512.0 | 3200.0 | 8320 | 221 | 2182.1 | 0.1 | 2 | 1 |
NAMD | 4 | 64.0 | 64.0 | 64.0 | 512 | 475 | 1438.4 | 0.1 | 5 | 5 |
OSIRIS | 128 | 256.0 | 256.0 | 256.0 | 256 | 13 | 1187.8 | 0.0 | 1 | 1 |
FDS | 248 | 248.0 | 248.0 | 384.0 | 384 | 23 | 1080.3 | 0.0 | 1 | 1 |
SBLI | 1 | 1024.0 | 1024.0 | 1024.0 | 1024 | 480 | 817.1 | 0.0 | 3 | 3 |
TPLS | 4 | 2048.0 | 4096.0 | 4096.0 | 4096 | 70 | 754.4 | 0.0 | 3 | 1 |
SIESTA | 1 | 2304.0 | 2304.0 | 2304.0 | 2304 | 57 | 703.7 | 0.0 | 4 | 3 |
DL_POLY | 1 | 512.0 | 768.0 | 1024.0 | 1152 | 114 | 680.7 | 0.0 | 3 | 2 |
Smilei | 2 | 256.0 | 384.0 | 384.0 | 1024 | 60 | 384.5 | 0.0 | 3 | 1 |
Arm Forge | 1 | 4096.0 | 8192.0 | 8192.0 | 8192 | 92 | 111.5 | 0.0 | 6 | 6 |
HemeLB | 32 | 256.0 | 256.0 | 256.0 | 2048 | 40 | 79.3 | 0.0 | 3 | 2 |
AxiSEM3D | 256 | 256.0 | 256.0 | 256.0 | 256 | 35 | 19.7 | 0.0 | 1 | 1 |
FEniCS | 160 | 160.0 | 160.0 | 160.0 | 160 | 2 | 5.8 | 0.0 | 1 | 1 |
ECOGEN | 256 | 256.0 | 256.0 | 256.0 | 256 | 1 | 2.1 | 0.0 | 1 | 1 |
Zacros | 512 | 512.0 | 512.0 | 512.0 | 512 | 5 | 1.2 | 0.0 | 1 | 1 |