16:25 PDT Sept 1st - We have an open enquiry with Amazon Web Services to give us full information on the root cause of this incident. There was a fail in one system at 06:17 that caused failover within the cluster within 2 minutes, and consequent latency of all database queries. After successful failover we expected a return to normal performance levels, but saw a continuation of high latency. Our DevSecOps team restarted specific services in order to restore normal performance levels. Once we receive more information from Amazon, we will post a more comprehensive description here.
11:00 PDT Aug 31st - We are still investigating the root cause of this incident, and awaiting further information from external sources.
10:00 PDT Aug 30th - Initial evidence has demonstrated that a deeper investigation will be required by our development team before we can publish a full post mortem on this incident. Next update can be expected by 9am PDT on Aug 31st. We appreciate your patience.