Investigating reports of on-demand call failures on US server

Incident Report for Boostlingo

Postmortem

During the incident window, some customers experienced intermittent failures when initiating new on-demand calls (including audio, IVR, and custom routing flows). The impact was primarily limited to call setup and routing: ongoing calls were largely unaffected, but customers using more complex routing logic reported higher failure rates and incomplete call initiations.

The issue was caused by excessive growth and load in a call-routing datastore and related services, which reached an operational threshold and degraded routing performance. Under this elevated load, dependent components (including SignalR hubs and callback handling) began reconnecting and failing intermittently, preventing new routing entries from being processed reliably. Our engineering team validated Azure infrastructure health, stabilized the affected routing components, and service recovered as load normalized and routing entries decreased.

To prevent recurrence, we are enhancing monitoring/alerting on datastore growth and service load thresholds, implementing optimized routing data lifecycle management to avoid unbounded growth, and improving observability to detect routing degradation earlier—before it impacts customers.

Posted Feb 02, 2026 - 15:06 PST

Resolved

We have concluded our monitoring and observed no errors during our monitoring window. While we are marking this incident as resolved we have also implemented additional automated monitoring and will continue to check database stability. We will share a postmortem in the next 72 hours.
Posted Jan 28, 2026 - 14:14 PST

Monitoring

Our team made updates to the database to address the auto scaling issues that led to degradation with call routing and connection. We are monitoring the change to ensure that all calls are routing as expected. At this time we are not experiencing platform degradation.

We will share additional technical details about this issue and our remediation in a postmortem on this page in the next 72 hours.
Posted Jan 28, 2026 - 13:28 PST

Identified

Our team has identified a database auto-scaling issue and is working on mitigation. Over the last fifteen minutes we have seen increased connection rates as our team has taken action.

We will continue to share updates through our status page as our team takes action.
Posted Jan 28, 2026 - 12:38 PST

Update

Our team continues to investigate issues with on demand calls successfully routing and connecting to interpreters. At this time we are experiencing degraded performance with some calls failing to route as expected or connect fully to an interpreter.

If you experience issues when placing a call we recommend waiting 30-60 seconds and trying again as we are seeing many calls routing and connecting as expected.
Posted Jan 28, 2026 - 11:39 PST

Update

We are continuing to investigate this issue.
Posted Jan 28, 2026 - 11:22 PST

Investigating

We are investigating reports of calls failing to connect across the US server. We will follow up with additional information here as soon as we have it. Please subscribe to ensure you receive updates as soon as they are available.
Posted Jan 28, 2026 - 11:19 PST
This incident affected: Boostlingo Voice (Boostllingo Voice IVR, Boostlingo Network Traversal Service, Boostlingo Speech Recognition, SIP Proxy), Boostlingo Video (Boostlingo Group Rooms, Boostlingo Communication REST API), Boostlingo Portal (Boostlingo Requestor Portal), and Twilio Voice, North America, Twilio SMS Long Code, North America.