Service Status

Login Nodes: Available

Last Updated: 01:15, Wednesday 22 February 2017


Compute Nodes: Available

Last Updated: 01:15, Wednesday 22 February 2017


Low Priority: Disabled

Last Updated: 01:15, Wednesday 22 February 2017


This page contains details of the status and usage of the ARCHER system.

Known Issues

Instability of login and post-processing (PP) nodes

You may currently be experiencing some problems with the ARCHER external login nodes and pre/post processing nodes.

This recent instability has been identified as a Lustre filesystem deadlock condition, which occurs intermittently during certain bulk read operations. The nature of this problem has been identified and Engineering teams are working hard to have a Lustre client fix available by the end of the month. In parallel, we are investigating the possibility of a workaround.

We will continue to monitor this issue and it will be discussed at the next ARCHER Management Board Meeting.

We will keep users informed and update you with further information as soon as it becomes available.

We hope that there will be an improvement soon and we thank you for your patience.

Open Issues

No open issues

Recently Resolved Issues

No recently resolved issues

Upcoming Maintenance Sessions

Lustre Patch : Wednesday 22nd Feb 2017 and Thursday 23rd Feb 2017
Assuming the patch is available on time, we will be following a similar plan to last week to apply it to the login nodes on Wed 22 Feb 2017 and to the PP nodes on Wed 22 Feb 2017 (espp1) and Thu 23 Feb 2017 (espp2).
The serial queue will be available throughout the work.

Users should:
  • login to logina.archer.ac.uk, 08:30-13:00 GMT, Wed 22 Feb 2017
  • login to loginb.archer.ac.uk, 13:00-18:00 GMT, Wed 22 Feb 2017
  • be aware that espp1 will be rebooted at 10:00 GMT, Wed 22 Feb 2017
  • be aware that espp2 will be rebooted at 11:00 GMT, Thu 23 Feb 2017

Live Usage/Queue Data

Total Usage

Usage history for:

This shows the number of nodes currently in use on ARCHER.

ARCHER usage graph.

Queue Status

Queue history for:

The graph below shows the status of the standard queue on ARCHER.

ARCHER standard queue status graph.

Live Usage/Queue Data by Job Length/Size

This section displays ARCHER usage and queue information broken down by the job class (class is based on size in nodes and run time in hours).

Queue Times by Job Size/Length

The heatmap below provides information on the mean queue times on ARCHER over past periods for different job sizes/lengths. The colour in the heatmap indicates the mean queue time as measured by scheduling coefficient for each class and the number in the box is the number of jobs of that class.

The scheduling coefficient is the ratio of run time to run time plus queue time and, as such, represents the fraction of the time in the queue that was spent running. A value of 1 indicates that the job started instantaneously and a value of 0.5 indicates that the job queued for the same amount of time that it ran for.

Usage history for:

ARCHER scheduling coefficient.

Usage by Job Size/Length

The heatmap below provides information on the usage on ARCHER over past periods for different job sizes/lengths. The colour in the heatmap indicates the number of kAU expended for each class and the number in the box is the number of jobs of that class.

Usage history for:

ARCHER usage by job size/length.

ARCHER Disk Status

RAID Check, fs2: Not Running

Last Updated: 01:15, Wednesday 22 February 2017


RAID Check, fs3: Not Running

Last Updated: 01:15, Wednesday 22 February 2017


RAID Check, fs4: Not Running

Last Updated: 01:15, Wednesday 22 February 2017


ARCHER Network Traffic

Data transferred on and off ARCHER in the last 72 hours (green = data transferred off ARCHER; blue = data transferred on to ARCHER).

ARCHER network traffic graph.

Data transferred on and off RDF in the last 72 hours (green = data transferred off RDF; blue = data transferred on to RDF).

ARCHER network traffic graph.