On 2 Feb 2026, between 20:13:34 to 22:56:04 UTC, Octopus Cloud customers in West US 2 and West Europe may have experienced failed deployments or failed runbook runs due to Ubuntu Dynamic Worker steps failing on Leasing timeout.
This disruption was caused by Azure failing to provision Virtual Machines across multiple regions - see Azure Issue FNJ8-VQZ on Azure Status History.
Octopus Cloud Dynamic Workers are isolated virtual machines that we provide as part of our Octopus Cloud Subscription offering as a way to execute deployment and runbook steps and scripts, without needing to run on the Octopus Server or deployment targets themselves. Customers can use both Windows and Ubuntu Dynamic Workers.
Octopus provides a dynamic worker pool of these virtual machine types from which, as required by your deployment/runbook steps, your Octopus Cloud will exclusively lease a freshly provisioned dynamic worker VM for a limited time.
Octopus Cloud implements multiple layers of protection to ensure Dynamic Workers’ availability and minimize customer impact in cases of service disruptions:
All dates and times below are in UTC
Feb 2 2026:
18:52: 1st failed Dynamic Worker provisioning - At this point, our Dynamic Workers Service continued supplying workers successfully from the pools. However, since we couldn’t provision new workers, the pools started depleting.
20:13: 1st Ubuntu Dynamic Worker lease failed in West US2 Azure region (once the pool was depleted) - start of customer impact
20:47: Octopus on-call was paged after 3 Dynamic Worker lease requests failed in West US2. The on-call then started the incident to investigate the issue
20:47-21:20:
West Europe and East Australia Azure regions. At this point we realized that this was a multi-region outage in Azure and decided to open a support ticket with them.21:24: an on-call engineer opened a Sev A support ticket with Azure
21:31: We published the initial partial outage alert for Octopus Cloud on https://status.octopus.com/
21:33: 1st Dynamic Worker lease failed in West Europe
21:42: Azure acknowledged the multi-region issue and reported that they are investigating it.
22:56: We saw the last Dynamic Worker lease failure. After this time, we were able to provision the required Virtual Machines and return the Dynamic Workers successfully to all lease requests.
23:46: After verifying that all Dynamic Worker pools have been restored and we didn’t see any additional provisioning failures, we updated the incident status to “Mitigated”, and updated the Status page.
Feb 3 2026:
6:05: Azure confirmed that the issue was fully resolved on their side.
21:38: Incident was resolved and Status page updated
Octopus Cloud uses Azure Virtual Machines in order to supply Dynamic Workers to customers. During this Azure outage, we couldn’t provision new Azure Virtual Machines for our Ubuntu Dynamic Workers. Our pre-provisioned Dynamic Workers’ pools continued supplying Dynamic Workers for additional
West US 2 and West Europe before they were depleted and customer requests for new Dynamic Workers started failing. It’s worth noting that East Australia customers were not impacted because the pools in this region didn’t deplete during the incident. Additionally, customers that already had a Dynamic Worker leased at the time of the outage were not impacted (unless their Dynamic Worker expired so a new one was requested during the incident).
We made preparations to switch to our Dynamic Workers Standby Service on different Azure regions. However, once we realized that this was a multi-region outage, we decided to revert the switch since it wouldn’t have resolved the issue.
Once we saw that we couldn’t supply Dynamic Workers from our standby regions, we turned off non-essential services (e.g. Instance Upgrades) to preserve the available Dynamic Workers for customer use.
Once Azure mitigated the issue on their side, our system recovered automatically and resumed providing new Dynamic Workers successfully.
Octopus takes service availability seriously. Despite the difficulty with upstream cloud provider outages, especially ones that are widespread across multiple regions, we fully review and remediate any outages that occur. We do this so that we’re continuously improving and maintaining the best possible service we can.
Following our post mortem, we identified the following improvements to our system to help identify and mitigate (where possible) similar issues earlier:
We apologize to our customers for any disruption and inconvenience as a result of this incident.
We have started work on the identified remediations to ensure that we can detect similar incidents more quickly and reduce the impact on our customers as much as possible.