ARCHER2 Weekly Newsletter
By ARCHER2 Service on September 20, 2023
Tags:
- Network Upgrade Maintenance: Monday 18th September – Friday 22nd September
- RDFaaS File System
- The 11th ARCHER2 and Exascale eCSE software development call and Early Career eCSE Panel Observers call
- Seminar - Structure and delivery format for Online HPC courses - 9th October 2023 11:30am - 1pm BST Hybrid
- Reproducible research using containers - introduction to Singularity/Apptainer, University of Nottingham, 21 September 2023 09:30 - 16:00 BST
- Shared memory programming with OpenMP, Imperial College London, 28 - 29 September 2023 10:00 - 17:00 BST
- Recently added known issues
- Upcoming ARCHER2 training
Network Upgrade Maintenance
Monday 18th September 09:00 – Friday 22nd September
There will be a major Data Centre Network (DCN) upgrade which will impact the external network connectivity at the Advanced Computing Facility (ACF). This essential upgrade work will significantly increase the reliability of our DCN and greatly simplify the deployment of new services and network changes to existing services.
During the upgrade, users will be unable to connect to ARCHER2 via the login nodes and will not be able to access any data on the system which includes the /home, /work, NVMe and RDFaaS (/epsrc and /general) file systems. The serial QoS will also be unavailable.
The compute nodes will still be available and jobs will run to ensure a good utilisation of the service during this outage.
The SAFE will be available during the outage but there will be reduced functionality due to the unavailability of the connection to ARCHER2 such as resetting of passwords or new account creation.
The Service Desk will be available as usual throughout the outage and we will notify users as soon as ARCHER2 is returned to full service. We apologise for any inconvenience caused by this essential maintenance.
RDFaaS file system
The RDFaaS (Research Data Facility as a Service) file systems provide additional capacity for projects to store data that is not currently required on the compute nodes but which is too large for the home file systems.
If you do not currently have access to the RDFaaS and would like access for your active ARCHER2 project then please contact the Service Desk.
Further information for the RDFaaS can be found in our documentation
Please note the RDFaaS is not currently available for NERC funded projects.
The 11th ARCHER2 and Exascale eCSE software development call and Early Career eCSE Panel Observers call
We are very pleased to announce that the 11th ARCHER2 eCSE call (ARCHER2 eCSE11) opened on the 12 Sep 2023. For this call, in addition to proposals for codes to run on ARCHER2, we are also accepting proposals looking to prepare codes for a future Exascale successor, including for example GPU development. The deadline for submitting documents for technical evaluations is 16:00 on 10 Oct 2023, with the final deadline for proposal submission being 16:00 on 31 Oct 2023. The early career eCSE Panel Observers opened alongside the eCSE call and the deadline for applications for this call is 16:00 on 10 Oct 2023.
The 11th ARCHER2 eCSE call
Through a series of regular calls, Embedded CSE (eCSE) support provides funding to the ARCHER2 user community to develop software in a sustainable manner, to run on the ARCHER2 system, or to prepare for a future Exascale successor. Projects targeting a future GPU platform are welcome, as are projects looking at any aspect of preparing the ARCHER2 software stack for Exascale. Note projects that only target improvements for existing platforms (other than ARCHER2), rather than preparing for future Exascale systems, are not in scope.
Funding will enable the employment of an RSE (a Research Software Engineer, PDRA or equivalent) to work specifically on the relevant software to enable new features or improve the performance of the code. Funding can be requested for up to 18 person months of effort per project. There is flexibility in the way the effort is requested. For example, a project may have more than one person working on it or the effort could be spread over a greater number of calendar months, i.e. a staff member could work on the project at less than 100% alongside other commitments or it could be worked on by a staff member who doesn’t work full time. Please just get in contact with the ARCHER2 Service Desk if you wish to discuss possible scenarios.
For this call, software which results in research which is within the remit of EPSRC is eligible. Funding can be requested for staff located at the institution of the PI, third parties, or can include staff from the ARCHER2 centralised CSE support team at EPCC or a mixture of the above.
More details of how to apply are available by following the link to the eCSE calls page together with details of the application and review process, assessment criteria eligibility, etc.
A recent webinar can be found on our website which includes a walk through of the submission of a proposal.
ARCHER2 Early Career eCSE Panel Observers call
The eCSE Panel Meeting is the meeting where eCSE proposals are reviewed and decisions are made on which proposals to fund. As part of our commitment to encouraging and developing Early Career Researchers, we are offering a small number of places to such researchers to attend the eCSE panel meeting as Early Career Observers. The aim is to give Early Career Researchers a better insight into this competitive selection process, to assist them in the preparation of future funding proposals.
An insight into the experience from one of our recent Early Career Panel Observers
eCSE and observers call dates
Deadline for submitting ARCHER2-eCSE11 documents for technical evaluation: 16:00 on 10 Oct 2023
Deadline for final ARCHER2-eCSE11 submission: 16:00 on 31 Oct 2023
Deadline for Early Career eCSE Panel Observers call: 16:00 on 10 Oct 2023
Seminar - Structure and delivery format for Online HPC courses
9th October 2023 11:30am - 1pm BST Hybrid
A UNIVERSE-HPC Project seminar, but likely to be of interest to ARCHER2 community
With the rapid move to wide scale use of online training kickstarted by the COVID pandemic, there has been a lot of discussion about the best technologies to use for the delivery of lectures and practical for remote courses, but surprisingly little discussion about the high-level structure such as timetabling. In this talk, our speaker will discuss the pros and cons of a variety of formats that EPCC has used for online training before, during and after the pandemic, including intensive live courses; full or half days with gaps between them; recorded lectures with live practical; fully asynchronous courses such as MOOCs or online degree programmes.
Further details and registration
Reproducible research using containers - introduction to Singularity/Apptainer
University of Nottingham, 21 September 2023 09:30 - 16:00 BST
This course aims to introduce the use of containers with the goal of using them to effect reproducible computational environments. Such environments are useful for ensuring reproducible research outputs and for simplifying the setup of complex software dependencies across different systems. The course will introduce the use of Singularity/Apptainer containers but the material will be of use for whatever container technology you plan to, or end up, using. On completion of this course attendees should:
- Have an understanding of what Singularity/Apptainer containers are, why they are useful and the common terminology used
- Have a working Singularity/Apptainer installation on your local system to allow you to use containers
- Understand how to use existing Docker/Singularity/Apptainer containers for common tasks
- Be able to build your own Singularity/Apptainer containers by understanding both the role of a recipe in building containers, and the syntax used in Singularity/Apptainer recipes
- Understand how to manage Singularity/Apptainer containers on your local system
- Appreciate decisions that need to be made around containerising research workflows
- Appreciate issues around reproducibility in software, understand how containers can address some of these issues and what the limits to reproducibility using containers are
Shared memory programming with OpenMP
Imperial College London, 28 - 29 September 2023 10:00 - 17:00 BST
Almost all modern computers now have a shared-memory architecture with multiple CPUs connected to the same physical memory, for example multicore laptops or large multi-processor compute servers. This course covers OpenMP, the industry standard for shared-memory programming, which enables serial programs to be parallelised easily using compiler directives. Users of desktop machines can use OpenMP on its own to improve program performance by running on multiple cores; users of parallel supercomputers can use OpenMP in conjunction with MPI to better exploit the shared-memory capabilities of the compute nodes.
This course will cover an introduction to the fundamental concepts of the shared variables model, followed by the syntax and semantics of OpenMP and how it can be used to parallelise real programs. Hands-on practical programming exercises will be included, with access to HPC provided for the duration of the course.
Recently added known issues
The “Known Issues” page of the ARCHER2 Documentation https://docs.archer2.ac.uk/known-issues/ lists all current open known issues including a description of the issue, its symptoms and any work-arounds.
- No recent issues
Upcoming ARCHER2 Training
- Message-passing Programming with MPI, Online, always-open self-service course
- Shared Memory Programming with OpenMP, Online, always-open self-service course
- QM/MM with GROMACS + CP2K, Online, Always open - self-service course
- Reproducible research using containers - introduction to Singularity/Apptainer, University of Nottingham, 21 September 2023 09:30 - 16:00 BST
- Shared memory programming with OpenMP, Imperial College London, 28 - 29 September 2023 10:00 - 17:00 BST
- Building and installing packages with CMake: a hands-on introduction, online vt, Wednesday 4th October 2023 15:00 - 16:00 BST
Further details of upcoming training
We always welcome researchers wishing to present their work in a webinar - please contact the Service Desk if you would be interested in presenting your work.