Blog

Your DR needs a DR


The core problem: Your on-prem disaster recovery (DR) assumes regional failures: a power outage, a hardware crash, a natural disaster. It doesn’t account for an entire country being targeted by drone strikes and missile attacks. To survive geopolitical disruption, you need public-cloud-based DR that can spin up in a completely different country in minutes—not a backup system that’s still inside the same borders as your primary.


Three weeks ago, Iranian drone strikes damaged three AWS data centers in the UAE and Bahrain. Banks went down. Payment platforms went dark. Ride-hailing apps stopped working. For the first time in history, a hyperscaler’s physical infrastructure was directly targeted in a military conflict. “Geopolitical risk” used to be a line item in a risk register that nobody took seriously. Now, data centers are strategic targets alongside oil refineries, power plants, and transportation hubs. The threat model just changed overnight.

And I know exactly what some of you are thinking: See? The public cloud isn’t safe either.

Here’s what you should actually be thinking: My on-prem DR is in the same country as my primary system. That’s the real problem the conflict in the Middle East just exposed—not that data centers can be hit, but that your DR strategy assumes disasters stay local. When an entire country’s infrastructure is under attack, a backup in a different city has the same level of exposure.

The real lesson from the AWS strikes

After the drone strikes knocked out two availability zones in the UAE and one in Bahrain, AWS told its customers to migrate workloads to other regions. And the sophisticated ones did—within hours. Their services came back online. Not because the damaged data centers were repaired, but because the AWS public cloud has 39 regions across the globe. Microsoft Azure has 70. Google Cloud has 43.

You lose one, you move to another. That’s not a vulnerability. That’s the most resilient architecture on the planet.

Now compare that to your on-prem disaster recovery setup.

You have two data centers: maybe active/active, maybe a warm standby. They’re in different cities, maybe even different parts of the country. But they’re in the same country. When a conflict targets your entire nation’s infrastructure—the way Iran targeted the UAE and Bahrain—having your DR in a different city doesn’t help. There is no “migrate to another region.” You’re calling your vendor, who will quote you six to nine months to ship hardware and stand up a new system.

How do I know? Because I’ve lived it.

Been there, done that

In early 2024, Zain Sudan lost both of its on-prem data centers during the Sudanese civil war. It was a complete network blackout: charging down, tens of millions of dollars a week in lost revenue. Zain’s incumbent vendor, Ericsson, projected a six-to-nine-month recovery timeline. Ship hardware to a new data center. Fly in engineers. Install the software. The standard playbook.

Looking for a better answer, Zain called Totogi, where I’m the CEO. We deployed Totogi Charging-as-a-Service on AWS in 18 days. Voice charging went live first, followed by data and SMS. We restored service to over 20 million subscribers. Today, Zain uses Totogi as its public cloud backup—because now that it knows what it’s like to have a charging system that can spin up anywhere in minutes, it’s not giving that up.

The operational model is simple: keep subscriber data synced, and when disaster strikes, switch over. No complex dual-write architectures. No real-time synchronization headaches. Just a pragmatic approach that works when everything else has failed.

That was the proof point. What’s happening right now in the Middle East is the wake-up call.

Zain was just the beginning

Totogi hasn’t been sitting still in the meantime. The initial Zain deployment, in 2024, used the first generation of Charging-as-a-Service. Since then, the platform has matured.What we offer now is a spectrum of operational readiness—each tier faster, more resilient, and more capable than the last. You choose the level that matches your risk tolerance and your budget.

Tier 1: Business continuity includes a nightly subscriber and balance sync on basic voice, data, and SMS plans with a four-hour failover window. This approach keeps the lights on and the revenue flowing while you recover your primary system. For minimal cost, you get maximum coverage for the scenario nobody plans for.

Tier 2: Accelerated failover gives you bulk balance synchronization that cuts the failover window dramatically. You also get full bundle and product support—not just basic plans, but the actual offers and promotions your subscribers expect, like integrated wallet top-ups and low-balance notifications. These are the operational details that make the difference between “technically charging” and “actually running a business.”

Tier 3: Near real-time readiness provides continuous synchronization, ensuring a minimal data loss window. At this tier, your public cloud layer isn’t waiting for a nightly dump, but tracking your operational state in near real-time. If disaster hits, the gap between your last known state and your recovery state approaches zero. This is DR that behaves like a live system, deployed to any AWS region on the planet, dormant until the moment you need it.

Every deployment we do makes the next one faster and more capable. That’s the structural advantage of solving this problem in production, under pressure, and refining it continuously.

No other charging vendor has done this. Not Ericsson. Not Amdocs. Not Nokia. None of them have deployed a public cloud DR layer for a competitor’s charger during an active conflict and kept 20 million subscribers online. Totogi has. The integration patterns, the sync architecture, the failover playbook: it all exists because we deployed it under real conditions. That’s not something you replicate from a slide deck.

Reinforce your disaster recovery

I’m not suggesting you rip out your existing DR system. I’m telling you to add a layer on top of it: a public cloud environment that sits dormant until you need it, then scales instantly when you flip the switch. You have your primary on-prem charger, your on-prem DR twin, and now a public cloud third layer that can spin up in any AWS region on the planet.

Your primary goes down? Fail over to your on-prem DR as usual. Both go down? Burst to the public cloud—in a completely different country. If your country is affected by conflict, your charging DR can be running in Frankfurt. Or Mumbai. Or São Paulo. Or all three. You can move your most critical revenue system out of the country entirely, in minutes.

On-prem DR can’t do that. The public cloud can.

Totogi can do this with any on-prem charging system. We sit alongside your existing infrastructure as the emergency public cloud layer, syncing subscriber data, balances, and electronic data records (EDRs). When disaster strikes, traffic redirects to Totogi on AWS—in whatever region is safe.

A DR for your DR

Since the conflict started three weeks ago, Totogi’s phones have been ringing. Operators across the Middle East and Africa are all asking us the same question Zain Sudan asked two years ago: Can Totogi be the public cloud DR for our DR?

The answer is yes. You choose the tier. We deploy it to whatever region is safe.

The conflict in the Middle East just showed us that on-prem DR has a single point of failure we never really planned for: geography. The good news is that public cloud eliminates that failure mode entirely. The even better news is that Totogi Charging-as-a-Service is ready to be your public cloud DR layer.

You don’t have to replace a thing. You just add a public cloud layer that lets you move your charging system out of the affected region—to any region on the planet—in minutes.

For operators in the Middle East, disaster recovery is no longer a hypothetical. It’s happening right now. The public cloud gives you virtually instant access to relocate your business operations out of the region. Use it!

Get in touch with the Totogi team to set up public cloud DR for your charger today.

Recent Posts

  1. Take back your telco
  2. The five myths of agentic AI
  3. A read-only ontology is just a data lake with better marketing
  4. Amdocs’ AI reckoning
  5. You built a data lake. You needed a river.


Get my FREE insider newsletter, delivered every two weeks, with curated content to help telco execs across the globe move to the public cloud.

Get started

Contact Totogi today to start your cloud and AI journey and achieve up to 80% lower TCO and 20% higher ARPU. 

Discover

Take note

Reimagine

By using cloud bursting for disaster recovery, telcos can slash costs and eliminate the headaches of maintaining an on-prem system. Learn more about how it works.


Connect

Engage

Connect with an expert today!

Set up a meeting to learn how the Totogi platform can unify and supercharge your legacy systems, kickstarting your AI-first transition.
Test Drive

Transform

Charging-as-a-Service

When Zain Sudan turned to Totogi’s Charging-as-a-Service to ensure business continuity, the solution protected revenue and improved customer satisfaction.


Frequently Asked Questions

1. Why isn’t on-premises disaster recovery enough anymore?

Traditional on-prem DR was built for local failures: a power outage, a hardware crash, a flood. It was never designed for a scenario where an entire country’s infrastructure becomes a military target. When that happens, your “backup” data center in a different city has the exact same problem as your primary: it’s still inside the same borders. Geography is the single point of failure no one planned for—until now.

2. Didn’t the AWS strikes in the UAE prove that the public cloud isn’t safe either?

No—they proved the opposite. When Iranian drone strikes damaged two AWS availability zones in the UAE and one in Bahrain, customers were able to migrate their workloads to other regions within hours. That’s resilience by design. AWS has 39 global regions. Azure has 70. Google Cloud has 43. Losing one region is an inconvenience. Losing your entire country’s infrastructure with nowhere to move? That’s an existential problem—and exactly what on-prem DR cannot solve.

3. How quickly can a telecom operator actually fail over to a public cloud charging system?

Faster than you think—if you’ve set it up in advance. When Zain Sudan lost both of its on-prem data centers during the Sudanese civil war, Totogi deployed Charging-as-a-Service on AWS in 18 days and restored service to over 20 million subscribers. That was the first deployment, under live wartime conditions, with no playbook. With the architecture Totogi has built since then, failover windows are measured in minutes. Compare that to the six-to-nine months Ericsson quoted Zain to ship hardware and stand up a new system. It’s a big difference.

4. Does adding a public cloud DR layer mean ripping out existing infrastructure?

Absolutely not. The whole point is that you’re adding a layer, not replacing anything. You keep your primary on-prem charger, and your on-prem DR twin. You add a third layer—a public cloud environment that sits dormant, costs a fraction of your existing DR, and can spin up in any AWS region on the planet the moment you need it. If your country becomes a war zone, your charging system can be running in Frankfurt, Mumbai, or São Paulo in minutes. Your on-prem DR cannot do that. The public cloud can.

5. What does public cloud disaster recovery actually cost compared to traditional DR?

Most operators are already paying close to 100% of their primary system cost to run a DR twin, because it has the same hardware, same software licenses, and same headcount to keep it current. Even so, when disaster actually strikes, operators generally just restore the primary because the twin hasn’t been properly tested or synced. A public cloud DR layer costs a fraction of that, because you’re paying for storage and minimal compute until you actually need it. There’s no duplicate hardware. No second set of licenses. No second team. It’s a category change in DR economics.