Full system power down
Friday 29th August 2025 09:00 - Monday 15th September 18:00
Due to a significant Health and Safety risk, associated with our power supply to the site, action is required at the Advanced Computing Facility (ACF). There will be a full power outage to the site during this period. Specialised external contractors will be working on a 24/7 basis for the outage period replacing switchgear.
Users will not be able to connect to ARCHER2 and will not be able to access data on any of the ARCHER2 file systems. The system will be drained of jobs ahead of the power outage and jobs will not run during this period. Any queued jobs will remain in the queue during the outage and jobs will start once the service is returned. SAFE and the ARCHER2 website will be available.
Update Tuesday 16th September 17:00
The HPE ARCHER2 systems team continue the work to return ARCHER2 to full service. The login nodes and file systems are now powered on and the issue with the scratch file system has been resolved. The HPE team are now in the process of powering on the compute nodes. Once all of the nodes are powered up, the EPCC CSE team will complete application tests this evening. We hope to open the service to all users and release jobs to the compute nodes tomorrow morning and will notify users as soon as the service is available.
Update Tuesday 16th September 14:30
The HPE ARCHER2 systems team continue the work to return ARCHER2 to full service. Two network switches have been replaced and work is now progressing on the scratch file system but an issue has been identified. The team are working to resolve this and once the file systems are in good order, work will progress to power on the compute nodes whilse monitoring the power and cooling load in the data centre. We will provide further updates as we have them.
- Current System Load - CPU, GPU
- Service Alerts
- Maintenance Sessions
- Previous Service Alerts
- System Status Mailings
- FAQ
- Usage statistics
Current System Load - CPU
The plot below shows the status of nodes on the current ARCHER2 Full System service. A description of each of the status types is provided below the plot.
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Note: the long running reservation visible in the plot corresponds to the short QoS which is used to support small, short jobs with fast turnaround time.
Current System Load - GPU
- alloc: Nodes running user jobs
- idle: Nodes available for user jobs
- resv: Nodes in reservation and not available for standard user jobs
- plnd: Nodes are planned to be used for a future jobs. If pending jobs can fit in the space before the future job is due to start they can run on these nodes (often referred to as backfilling).
- down, drain, maint, drng, comp, boot: Nodes unavailable for user jobs
- mix: Nodes in multiple states
Service Alerts
The ARCHER2 documentation also covers some Known Issues which users may encounter when using the system.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Ongoing | Issue | 2025-06-25 13:00 | Ongoing issues with /work1 file system | An issue with the /work1 file system is currently being investigated. This is only impacting some jobs and examples of errors seen include Cannot read/write checkpoint; corrupt file, or maybe you are out of disk space. | Under investigation |
Maintenance Sessions
This section lists recent and upcoming maintenance sessions. A full list of past maintenance sessions is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Planned | Full | 2025-08-29 09:00 | 2025-09-15 18:00 | Full ARCHER2 System |
Users will not be able to connect to ARCHER2 and will not be able to access data on any of the ARCHER2 file systems. The system will be drained of jobs ahead of the power outage and jobs will not run during this period. Any queued jobs will remain in the queue during the outage and jobs will start once the service is returned. SAFE and the ARCHER2 website will be available. Update Monday 15th September 16:21 The return to service is taking longer than expected as careful testing is needed of both the power and cooling and the machine at every stage. Work will continue tomorrow to bring ARCHER2 back, and we will send out a further update in the middle of the day. |
Due to a significant Health and Safety risk, associated with our power supply to the site, action is required at the Advanced Computing Facility (ACF). There will be a full power outage to the site during this period. Specialised external contractors will be working on a 24/7 basis for the outage period replacing switchgear. |
Previous Service Alerts
This section lists the five most recent resolved service alerts from the past 30 days. A full list of historical resolved service alerts is available.
Status | Type | Start | End | Scope | User Impact | Reason |
---|---|---|---|---|---|---|
Resolved | Service Alert | 2025-08-25 14:30 | 2025-08-25 20:00 | Reduced number of compute nodes | Increased queue times and reduced node availability | We stopped new jobs from starting and are monitoring the cooling closely during unusually forecasted high temperatures in the Edinburgh area. |
Resolved | Service Alert | 2025-08-18 11:00 | 2025-08-18 18:00 | Resolution of external hostnames from ARCHER2 | Users will not be able to resolve external host names from ARCHER2 nodes for a short time | Testing backup DNS for data centre downtime later in the year |
System Status mailings
If you would like to receive email notifications about system issues and outages, please subscribe to the System Status Notifications mailing list via SAFE
FAQ
Usage statistics
This section contains data on ARCHER2 usage for Jul 2025. Access to historical usage data is available at the end of the section.
Usage by job size and length
Queue length data
The colour indicates scheduling coefficient which is computed as [run time] divided by [run time + queue time]. A scheduling coefficient of 1 indicates that there was zero time queuing, a scheduling coefficient of 0.5 means that the job spent as long queuing as it did running.
Software usage data
Plot and table of % use and job step size statistics for different software on ARCHER2 for Jul 2025. This data is also available as a CSV file.
This table shows job step size statistics in cores weighted by usage, total number of job steps and percent usage broken down by different software for Jul 2025.
Software | Min | Q1 | Median | Q3 | Max | Jobs | Nodeh | PercentUse | Users | Projects |
---|---|---|---|---|---|---|---|---|---|---|
Overall | 1 | 512.0 | 1280.0 | 8192.0 | 262144 | 2723804 | 3933345.6 | 100.0 | 907 | 117 |
VASP | 1 | 256.0 | 640.0 | 1024.0 | 16384 | 857992 | 799281.2 | 20.3 | 155 | 17 |
Unknown | 1 | 512.0 | 1536.0 | 9216.0 | 99104 | 778285 | 717913.3 | 18.3 | 434 | 84 |
Nektar++ | 1 | 5120.0 | 8192.0 | 12800.0 | 23040 | 783 | 296212.3 | 7.5 | 17 | 5 |
CP2K | 1 | 512.0 | 720.0 | 1024.0 | 5120 | 63565 | 237985.9 | 6.1 | 61 | 13 |
No srun | 1 | 256.0 | 1024.0 | 9216.0 | 262144 | 108342 | 229167.5 | 5.8 | 675 | 103 |
SENGA | 1 | 8192.0 | 33500.0 | 33500.0 | 33500 | 198 | 192393.7 | 4.9 | 7 | 3 |
GROMACS | 1 | 256.0 | 640.0 | 2048.0 | 12288 | 37975 | 163321.6 | 4.2 | 49 | 7 |
LAMMPS | 1 | 128.0 | 256.0 | 1024.0 | 131072 | 106225 | 152546.9 | 3.9 | 56 | 18 |
OpenFOAM | 1 | 1024.0 | 2048.0 | 4096.0 | 25600 | 2610 | 144009.2 | 3.7 | 55 | 17 |
Met Office UM | 1 | 576.0 | 1024.0 | 1296.0 | 12544 | 13916 | 113859.9 | 2.9 | 45 | 6 |
Xcompact3d | 128 | 32768.0 | 32768.0 | 32768.0 | 32768 | 91 | 92932.9 | 2.4 | 5 | 4 |
Python | 1 | 9216.0 | 16384.0 | 16384.0 | 25600 | 42180 | 90739.9 | 2.3 | 63 | 26 |
FHI aims | 1 | 384.0 | 512.0 | 1024.0 | 6000 | 20524 | 76286.2 | 1.9 | 28 | 5 |
CASTEP | 1 | 128.0 | 256.0 | 512.0 | 8192 | 80845 | 66387.9 | 1.7 | 43 | 7 |
NEMO | 1 | 768.0 | 5504.0 | 5504.0 | 6528 | 4094 | 53013.0 | 1.3 | 23 | 3 |
EDAMAME | 1331 | 171475.0 | 171475.0 | 171475.0 | 171475 | 55 | 52585.5 | 1.3 | 2 | 1 |
ChemShell | 1 | 1024.0 | 10752.0 | 12800.0 | 12800 | 2026 | 49848.1 | 1.3 | 15 | 5 |
OpenSBLI | 64000 | 64000.0 | 64000.0 | 64000.0 | 131072 | 15 | 45114.4 | 1.1 | 2 | 2 |
iIMB | 2048 | 2304.0 | 2304.0 | 9216.0 | 9216 | 76 | 34793.5 | 0.9 | 1 | 1 |
GENE | 1 | 3584.0 | 8192.0 | 8192.0 | 10240 | 776 | 33501.2 | 0.9 | 15 | 3 |
PeleLMeX | 16 | 512.0 | 2048.0 | 4096.0 | 8192 | 382 | 30876.4 | 0.8 | 3 | 1 |
ONETEP | 4 | 128.0 | 128.0 | 128.0 | 128 | 698 | 29480.2 | 0.7 | 5 | 2 |
Code_Saturne | 128 | 4096.0 | 4096.0 | 4608.0 | 16384 | 146 | 26459.1 | 0.7 | 9 | 4 |
Nek5000 | 6 | 65536.0 | 65536.0 | 65536.0 | 65536 | 62 | 25983.0 | 0.7 | 3 | 2 |
EPOCH | 1 | 1152.0 | 1280.0 | 2304.0 | 16000 | 919 | 21691.8 | 0.6 | 7 | 2 |
MITgcm | 2 | 112.0 | 126.0 | 356.0 | 384 | 24578 | 21060.1 | 0.5 | 10 | 2 |
Quantum Espresso | 1 | 192.0 | 384.0 | 512.0 | 6400 | 10261 | 18334.3 | 0.5 | 34 | 10 |
GS2 | 128 | 2560.0 | 2560.0 | 2784.0 | 4096 | 9457 | 14572.0 | 0.4 | 5 | 2 |
Smilei | 8 | 128.0 | 512.0 | 512.0 | 512 | 1780 | 12516.6 | 0.3 | 3 | 1 |
Hydro3D | 30 | 2000.0 | 73440.0 | 73440.0 | 73440 | 86 | 11842.0 | 0.3 | 3 | 3 |
CESM | 1 | 1536.0 | 1536.0 | 6400.0 | 6400 | 305 | 11252.3 | 0.3 | 4 | 1 |
CRYSTAL | 128 | 128.0 | 512.0 | 512.0 | 2560 | 2487 | 10301.9 | 0.3 | 4 | 2 |
SIESTA | 1 | 1280.0 | 2304.0 | 2304.0 | 2304 | 415 | 9061.3 | 0.2 | 4 | 2 |
VAMPIRE | 128 | 1024.0 | 1024.0 | 1024.0 | 2048 | 369 | 6921.1 | 0.2 | 5 | 3 |
NWChem | 16 | 128.0 | 128.0 | 384.0 | 512 | 547924 | 6132.9 | 0.2 | 9 | 5 |
ptau3d | 32 | 200.0 | 240.0 | 400.0 | 400 | 33 | 5891.8 | 0.1 | 2 | 2 |
CASINO | 128 | 640.0 | 640.0 | 1280.0 | 2560 | 61 | 4954.2 | 0.1 | 2 | 2 |
RMT | 1 | 640.0 | 1024.0 | 2432.0 | 2560 | 258 | 4830.1 | 0.1 | 6 | 1 |
WRF | 6 | 384.0 | 384.0 | 384.0 | 384 | 82 | 3763.5 | 0.1 | 4 | 1 |
NAMD | 4 | 512.0 | 512.0 | 512.0 | 1536 | 233 | 2677.0 | 0.1 | 7 | 5 |
DL_POLY | 64 | 512.0 | 640.0 | 1024.0 | 1152 | 188 | 2006.1 | 0.1 | 5 | 2 |
TPLS | 480 | 2048.0 | 2048.0 | 2048.0 | 2048 | 17 | 1900.4 | 0.0 | 2 | 1 |
a.out | 1 | 2048.0 | 2048.0 | 4096.0 | 4096 | 639 | 1883.9 | 0.0 | 10 | 6 |
FDS | 248 | 248.0 | 248.0 | 248.0 | 248 | 34 | 1480.8 | 0.0 | 1 | 1 |
ECOGEN | 256 | 512.0 | 1024.0 | 1024.0 | 1024 | 24 | 1295.5 | 0.0 | 2 | 2 |
PDNS3D | 1024 | 1024.0 | 1024.0 | 1024.0 | 1024 | 69 | 1272.6 | 0.0 | 2 | 1 |
SU2 | 1280 | 6400.0 | 6400.0 | 6400.0 | 12800 | 3 | 1186.7 | 0.0 | 1 | 1 |
HYDRA | 1 | 2560.0 | 3840.0 | 4480.0 | 4480 | 356 | 773.9 | 0.0 | 6 | 4 |
SBLI | 512 | 1024.0 | 1024.0 | 1024.0 | 1024 | 51 | 388.5 | 0.0 | 1 | 1 |
BOUT++ | 768 | 768.0 | 1344.0 | 1344.0 | 1344 | 5 | 172.2 | 0.0 | 1 | 1 |
ABINIT | 40 | 320.0 | 320.0 | 320.0 | 640 | 19 | 156.8 | 0.0 | 1 | 1 |
ludwig | 8 | 128.0 | 256.0 | 512.0 | 1024 | 218 | 130.7 | 0.0 | 1 | 1 |
HemeLB | 8 | 512.0 | 1024.0 | 2304.0 | 4096 | 518 | 91.0 | 0.0 | 2 | 2 |
Arm Forge | 1 | 128.0 | 256.0 | 512.0 | 1024 | 421 | 36.0 | 0.0 | 10 | 7 |
DL_MESO | 128 | 128.0 | 128.0 | 128.0 | 128 | 7 | 31.6 | 0.0 | 1 | 1 |
AxiSEM3D | 128 | 128.0 | 128.0 | 256.0 | 256 | 78 | 26.8 | 0.0 | 1 | 1 |
Amber | 96 | 96.0 | 96.0 | 96.0 | 96 | 1 | 16.2 | 0.0 | 1 | 1 |
CloverLeaf | 512 | 512.0 | 512.0 | 512.0 | 512 | 1 | 0.3 | 0.0 | 1 | 1 |
COSA | 8 | 8.0 | 8.0 | 8.0 | 8 | 10 | 0.1 | 0.0 | 1 | 1 |
PRECISE | 8 | 8.0 | 8.0 | 8.0 | 8 | 36 | 0.0 | 0.0 | 1 | 1 |