Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
We’ve confirmed that all systems are back to normal as of November 6th 2018 15:45 UTC. Our logs show the incident started on November 6th 2018 14:45 UTC and during the 60 minutes it took to resolve the issue, a peak of 2150 users experienced slow and failed commands in West Europe region. Sorry for any inconvenience this may have caused.
- Root Cause: Our current investigations point at a failover on SQL Azure layer as the cause for this incident.
- Chance of Re-occurrence: Medium at this point, we understand to some extent what happened.
- Lessons Learned: We are logging a number of repair items and investigating in depth why an Azure SQL failover got our DBs in a bad state.
- Incident Timeline: 14:45 UTC through 15:45 UTC.
- We are actively investigating performance issues with Azure DevOps.
- Some customers may experience slower performance than usual while accessing their accounts.