❌

Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

Active - Virtual Machines and dependent services - Service management issues in multiple regions

2 February 2026 at 14:46

Impact statement: As early as 19:46 UTC on 2 February 2026, we became aware of an issue causing customers to receive error notifications when performing service management operations - such as create, delete, update, scaling, start, stop - for Virtual Machines (VMs) affecting multiple regions. These issues are also impacting services with dependencies on these service management operations - including Azure Arc Enabled Servers, Azure Batch, Azure Cache for Redis, Azure Container Apps, Azure DevOps (ADO), Azure Kubernetes Service (AKS), Azure Backup, Azure Load Testing, Azure Firewall, Azure Search, Azure Virtual Machine Scale Sets (VMSS), GitHub (see https://www.githubstatus.com)..

Current status: We determined that these issues were caused by a recent configuration change that affected public access to certain Microsoft‑managed storage accounts, used to host extension packages. We have applied our mitigation across all impacted regions and have performed validation checks to ensure that all affected resources have had their configurations updated. At this stage, customers should see signs of recovery across regions. We are currently monitoring downstream services for any further impact. Our next update will be provided by 08:00 UTC, approximately 2 hours from now, or sooner if we have progress to share.

Active - Managed Service Identity and depended service operation failure in East US and West US

2 February 2026 at 19:15

Impact Statement: Beginning at 00:15 UTC on 03 February 2026, customers utilizing Managed Identity for Azure resources may have encountered issues when trying to create, update, delete, or acquire tokens in the East US and West US regions. These errors impacted dependent services such as Azure Synapse Analytics, Azure Databricks, Azure Stream Analytics, Azure Kubernetes Service, Microsoft Copilot Studio, Azure Chaos Studio, Azure Database for PostgreSQL Flexible Servers, Azure Container Apps, and Azure AI Video Indexer. Please be aware that this is not an exhaustive list of affected services, and further updates will be provided as additional information is obtained.

Current Status: The underlying Managed Identity service has been stabilized, and key mitigation actions-including targeted service restarts and controlled traffic throttling-have been successfully applied.

As a result, most customer workloads are now recovering, and traffic is being safely and gradually ramped up to normal levels while we continue close monitoring. While core functionality has largely been restored, a small number of dependent workloads may still experience intermittent delays or retries as the environment continues to fully stabilize.

We will continue to actively monitor all dependent services and address any remaining tail issues.

The next update will be provided within 60 minutes, or as events warrant.

Mitigated – Networking reduced availability in East US

18 March 2025 at 09:09

What happened?

Between 13:09 UTC and 18:51 UTC on 18 March 2025, a platform issue resulted in an impact to a subset of Azure customers in the East US region. Customers may have experienced intermittent connectivity loss and increased network latency sending traffic within as well as in and out of East US Region.Β 

At 23:21 UTC on 18 March 2025, another impact to network capacity occurred during the recovery of the underlying fiber that customers may have experienced the same intermittent connectivity loss and increased latency sending traffic within, to and from East US Region.


What do we know so far?

We identified multiple fiber cuts affecting a subset of datacenters in the East US region at 13:09 UTC on 18 March 2025. The fiber cut impacted capacity to those datacenters increasing the utilization for the remaining capacity serving the affected datacenters. At 13:55 UTC on 18 March 2025, we began mitigating the impact of the fiber cut by load balancing traffic and restoring some of the impacted capacity; customers should have started to see service recover starting at this time. The restoration of traffic was fully completed by 18:51 UTC on 18 March 2025 and the issue was mitigated.Β 

At 23:20 UTC on 18 March 2025, another impact was observed during the capacity repair process. This was due to a tooling failure during the recovery process that started adding traffic back into the network before the underlying capacity was ready. The impact was mitigated at 00:30 UTC on 19 March after isolating the capacity impacted by the tooling failure.Β 

At 01:52 UTC on 19 March, the underlying fiber cut has been fully restored. We continue working to test and restore all capacity to pre-incident levels.Β 

Our telemetry data shows that the customer impact has been fully mitigated. We are continuing to monitor the situation during our capacity recovery process before confirming complete resolution of the incident.

An update will be provided in 3 hours, or as events warrant

❌
❌