Skip to main content

This is a new service – your feedback (opens in a new tab) will help us to improve it.

SEGAS-00014 Service reliability

Last updated: 26 April 2024
Relates to (tags): Reliability, Infrastructure, SRE

Home Office systems should be able to withstand failures during various circumstances, both expected and unexpected, over a specified period, and in specified environments.


Requirement(s)

Service reliability MUST be observable relative to defined service level expectations

Reliability expectations should be clear and baselined so that service level objectives (SLOs) can be defined for measures that can keep Home Office services reliable.

Service MUST be tolerant to expected and unexpected failure of a data centre

In the event of an unexpected catastrophic event (flooding, fire, etc.) or an expected event (data centre maintenance), an entire data centre may be lost causing outages of services that are hosted in the affected data centre.

Although rare, services should consider the possibility and guard against such scenarios by spreading applications / services across multiple data centres (e.g. availability zones) either in active/active or active/passive configuration.

Service MUST be tolerant to expected and unexpected failure of physical or virtualised hardware as well as network

In the event of an unexpected event (hardware fault, withdrawal of virtualised hardware e.g. spot instances) or an expected event (maintenance activities) your service may suffer an outage.
Such scenario is more common and your service should guard against such scenarios by operating:

  • from multiple data centres
  • from multiple physical / virtualised nodes

Service MUST be tolerant to expected and unexpected failure of upstream services such as databases

In the event of an unexpected event (failure of an upstream application, database, etc.) or an expected event (upgrade, maintenance activities, etc.) your service may suffer an outage.
Your service should guard against such scenarios by:

  • handling the failure gracefully
  • attempting to reconnect
  • attempting to retry within reasonable limits of your service

When retrying transactions, particularly when an upstream service was down, avoid flooding the upstream service(s) with requests thus potentially creating a DDoS style attack on the upstream service, potentially causing further upstream service outages. Furthermore, implement a backoff strategy if your transactions allow it.

Service MUST be sized appropriately for normal operations and MUST be able to automatically scale as appropriate based on metrics

The service must be sized appropriately in terms of system resources (CPU, memory, storage, etc.) for normal operating conditions preventing unnecessary resource wastage.

All components of a service can be scaled up or out (or down or in) based on some metric (such as CPU, memory, thread utilisation or some appropriate metric) so as to ensure continued successful operation of the service during expected and unexpected peak or off-peak operating periods.

Ideally, scaling should be based on some metric that is indicative of an application’s performance (such as number of busy threads, queue depth) rather than system resources such as CPU and memory. Not scaling based on intelligent metrics could result in situations such as:

  • your application has exhausted threads and cannot service any more requests, while CPU and memory utilisation is low
  • a messaging queue has filled up, and there are not enough consumers to process messages

Basing scaling policies on system resources should be the fallback option, rather than the main criteria by which you scale.
Allowing for automatic scaling of a service not only enables a service to respond accordingly during peak or off-peak times, but is also a valuable cost saver.

Services should be continuously reviewed.

Service MUST be able to handle requests to protect the overall service

The service must be able to apply rate limiting rules to protect itself from being overwhelmed by external and internal clients, preventing such scenarios as DDoS style attacks either intentional or accidental.

Check that services respond accordingly during load and stress testing; usually services should respond with HTTP 503

Service MUST be able to tolerate expected load

Where required based off business requirements, the service can be load, soak or stressed tested to expected normal conditions as defined by the service volumetrics to:

  • ensure the application can manage expected load
  • identify and remediate performance bottlenecks
  • understand trigger points for scaling policies

Service MUST be able to tolerate expected stress

The service must be stressed tested beyond normal load conditions (at least 200% of expected peak load), so as to:

  • understand how the application behaves beyond normal conditions, and document any findings
  • ensure scaling policies behave as expected, and that normal service can be expected from unexpected load conditions
  • identify and remediate performance bottlenecks in the service and upstream / downstream systems and services

This is particularly important as there will be hidden limits that affect scaling capacity, such as limits on upstream systems, cloud accounts, or network throughput on appliances connecting to 3rd parties/POISE.

Service MUST be soak tested

The service must be soak tested to ensure normal operating conditions over time, from low, normal to high load situations, so as to:

  • understand how the service behaves over a period of time, and that garbage collection effectively prevents memory and CPU leaks over time
  • understand how the service recovers from high load situations and ensure it relinquishes unused resource back to the underlying infrastructure