Outages
If you're here on November 18 2025, it's THIS outage you're after!
GitHub customers were facing degraded services late on Tuesday with the Microsoft-owned platform confirming issues across multiple services.
Actions, API requests, pages and pull requests all suffered issues.
The GitHub incident was first acknowledged by GitHub at 19:42 UTC.
“We experienced problems with multiple services, causing disruptions for some users. We have identified the cause and are rolling out changes to restore normal service. Many services are recovering, but full recovery is ongoing” GitHub confirmed at 20:04 UTC, shortly after the incident began.
GitHub’s last notable outage in late 2024 was later blamed on a “configuration change that impacted traffic routing within our database infrastructure, resulting in critical services unexpectedly losing database connectivity. There was no data loss or corruption during this incident.”
"We mitigated the incident by reverting the change,” it said at the time.
A major 11-hour Microsoft 365 outage in November 2024 meanwhile was triggered during a decommissioning workflow for an internal Microsoft 365 backend service. Before it was decommissioned, traffic to the service was not disabled and other services continued to send traffic to it.
As an incident reported noted, logic in routing requests to a backup endpoint if the now-decommissioned service was unreachable was in place, but the Microsoft Client Access Front End (CAFE) components managed these requests and their attempts to resolve requests headed for the decommissioned backend service included a synchronous call in an asynchronous code path. This led to threads being held for an extended time and they soon became exhausted,.
Even minor software changes across such large and complex organisations not irregularly trigger unforeseen issues. Last week Google Cloud took down services for millions of customers after deploying a change that caused an outage; a cold restart then triggered a thundering herd issue.
Google Cloud said in the wake of the incident that it would “enforce all changes to critical binaries to be feature flag protected and disabled by default; improve our static analysis and testing practices to correctly handle errors and if need be fail open... audit and ensure our systems employ randomized exponential backoff.”