Rain Lag

The Analog Reliability Story Trainyard Compass: Hand‑Mapping How Tiny Routing Choices Turn Into Big Outages

How to use story-driven, analog mapping to reveal how small routing and infrastructure decisions become large-scale outages—and how to build a practical reliability practice without waiting for the next crisis.

The Analog Reliability Story Trainyard Compass

Hand‑Mapping How Tiny Routing Choices Turn Into Big Outages


Modern systems fail in non‑obvious ways. A tiny local routing tweak, a “temporary” firewall rule, a minor DNS shortcut—each looks harmless in isolation. But under the right pressure, those small decisions combine, amplify, and suddenly you’re explaining a major outage to executives and customers.

This post introduces a practical way to see those failure paths before they see you: the Analog Reliability Story Trainyard Compass—a mouthful, but a simple idea. You map your systems like a trainyard, tell stories about how trains (traffic, dependencies, failures) move through it, and use that analog map as a compass to navigate reliability risks.

It’s deliberately low‑tech: paper, whiteboards, sticky notes, and conversation. Because once you can see how the tracks connect, it becomes easier to understand how small routing and infrastructure choices can propagate into big, cascading outages.


Why You Need an Analog Map in a Digital World

Most organizations have:

  • Network diagrams (often stale).
  • Dependency graphs (partial at best).
  • Cloud dashboards (too detailed to see the big picture).

What’s usually missing is a story‑friendly, human‑scale map of business risk. Not “where are the servers?” but:

Where does money flow, where do critical promises to customers live, and along which paths can failure realistically travel?

You can’t answer that with an auto‑generated diagram alone. You need a shared mental map that helps you:

  • Explicitly map business risks to systems and dependencies.
  • See how local decisions could create a global failure path.
  • Prioritize reliability work around actual impact, not just technical neatness.

That’s where the Trainyard Compass comes in.


The Trainyard Metaphor: Tracks, Switches, and Cascades

Imagine your architecture as a trainyard:

  • Trains = traffic, jobs, user requests, data flows.
  • Tracks = network paths, message queues, replication links, API calls.
  • Switches = routing rules, DNS configs, feature flags, load balancers.
  • Stations = services, databases, external providers.

In a quiet period, the yard looks orderly. But under load or during a partial failure, a small routing decision—one extra track, a misconfigured switch, a slightly overloaded station—can send trains down paths you didn’t intend.

Now add interdependence: stations depend on each other for power, signaling, capacity. This is where models like the Motter–Lai model from network science are relevant. Their core idea:

  • Components don’t fail in isolation.
  • When one component fails or is overloaded, its load shifts to neighbors.
  • Those neighbors can overload in turn, causing a cascade of failures.

In your world, that might look like:

  • A routing change pushes more traffic to a secondary region.
  • That region’s database replicas suddenly take 2× load.
  • Replication lags, read queries slow, queues back up.
  • Other services retry more often, increasing load further.
  • Eventually, both regions are degraded—and your status page lights up.

The failure didn’t start with “the database went down.” It started with a tiny routing choice in a dense, interdependent network.


Step 1: Map Business Risk, Not Just Topology

Before drawing any trains or tracks, ask: What do we actually care about losing? Examples:

  • The ability for customers to log in.
  • The ability to take payments.
  • The integrity of order records.
  • The timeliness of alert notifications.

For each critical business function, capture:

  • What breaks for the customer if this fails.
  • Rough financial or reputational impact (even a 3‑level scale helps).
  • Time sensitivity (seconds, minutes, hours).

Then connect those business functions to supporting systems:

  • “Login depends on: identity API, user DB, network path to IdP, DNS.”
  • “Payments depend on: checkout API, payment gateway, message bus, card network, email provider.”

This gives you a business‑centric starting point. Your trainyard isn’t just technical; it’s aligned to revenue and promises.


Step 2: Hand‑Draw the Trainyard

Get a whiteboard or paper. Invite engineers from multiple teams. Then, hand‑draw your trainyard for a single business function. Keep it rough, but include:

  • Stations (services / components): Draw boxes.
  • Tracks (paths / dependencies): Draw arrows with labels like “HTTP,” “VPN,” “replication,” “Kafka topic.”
  • Switches (routing and decisions): Mark load balancers, DNS, feature flags, and failover rules.

Ask explicitly:

  • Where do we assume traffic goes under normal conditions?
  • Where can traffic go under failover conditions—even the weird ones?
  • Which stations are shared by multiple trains (shared databases, shared queues, shared caches, shared network segments)?

The goal is not precision; it’s shared understanding of structure and interdependence.


Step 3: Story the Cascades with a Traceability Matrix

Next, turn this drawing into a set of stories about how things go wrong. To make that systematic, use a threat and vulnerability traceability matrix.

At its simplest, your matrix connects:

  • Vulnerabilities / Weaknesses
  • Triggering Events
  • Propagation Path
  • Incident Archetype / Business Impact.

Example (simplified):

VulnerabilityTriggerPropagation PathIncident Archetype
Secondary region LB has no rate limitsFailover sends 80% of traffic to secondaryLB allows spike → app nodes saturate → DB replicas overload → replication lag → read timeouts"Failover cascade" –
Degradation in both regions

As a group, walk through:

  1. Pick a small local choice. A NAT rule, a feature flag, a new routing policy, a shared queue.
  2. Invent a plausible trigger. A failover, a traffic spike, a partial provider outage, a misconfigured deployment.
  3. Follow the train. Where does load go? What switches flip automatically? What shared stations get hit?
  4. Name the archetype. Is this a “retry storm,” “failover cascade,” “slow partial outage,” “silent corruption,” or something else?

Writing this down creates traceability: you can see how a minor issue can escalate into a named, recognizable incident pattern. That’s far more actionable than “we might have a problem somewhere.”


Step 4: Treat Reliability Like a Red Team Exercise

Security teams use red teaming to think adversarially about defenses. Reliability teams can do the same for architectures and routing decisions.

In a reliability red team session:

  • One group plays “blue team” and explains how the system is supposed to work.
  • Another group plays “red team” and asks, “How could this fail spectacularly?”

The red team looks for:

  • Single shared stations that become hidden bottlenecks.
  • Switches that have asymmetric behavior under failure (e.g., fail open vs fail closed).
  • Paths that only exist under rare conditions (manual failover, emergency playbooks).
  • Places where retries, timeouts, or backpressure are unclear or inconsistent.

The mindset shift is key:

You’re not trying to prove the system is robust. You’re trying to break it on paper before it breaks in production.

This adversarial framing makes teams more honest about constraints and unknowns, and it surfaces routing choices that deserve better safeguards.


Step 5: Practice in Safe Chaos Environments

Thinking is not enough. Teams need to practice handling failure paths. But full‑on GameDays can be heavy:

  • They’re logistically complex.
  • They can be risky to run in production.
  • They’re intimidating to teams new to chaos engineering.

Start smaller with safe chaos environments and guided exercises:

  • Use a staging or sandbox environment that mirrors critical topology (even if at smaller scale).
  • Run scripted, low‑blast‑radius experiments: drop a dependency, slow a network link, simulate a DNS misconfig.
  • Use your analog map as a scenario script: “Let’s walk the exact cascade we sketched last week and see if reality matches the story.”

These workshops help teams:

  • Validate or correct the analog trainyard map.
  • Discover hidden coupling or unexpected routes.
  • Practice detection, communication, and mitigation.

You’re building muscle memory for outage response, not just diagrams.


Step 6: Grow from Exercises to a Culture of Reliability

Over time, you can mature from small workshops to:

  • Semi‑regular mini‑GameDays scoped to a single incident archetype (e.g., “retry storm day”).
  • Cross‑team reliability councils that maintain shared trainyard maps and traceability matrices.
  • Pre‑deployment risk reviews where new routing changes or dependencies are walked through on the trainyard map.

Full‑scale GameDays still have their place—especially for validating cross‑org readiness—but they become just one tool in a continuous reliability mapping practice, not a rare event.

The goal is a culture where:

  • Every significant routing or infrastructure choice prompts the question: “Where does this put new trains and switches in our yard?”
  • Teams routinely hand‑map new services and dependency changes.
  • Business leaders understand which tracks their revenue actually runs on.

Conclusion: Draw the Yard Before the Trains Derail

Outages rarely start as big, obvious failures. They begin as small, local choices made without a full picture of how interdependent the system really is.

By using an analog, story‑driven approach—the Trainyard Compass—you can:

  • Explicitly map business risks to systems and routes.
  • See how tiny routing and infrastructure choices can propagate.
  • Use a traceability matrix to connect vulnerabilities to concrete incident archetypes.
  • Run red team–style reliability sessions to stress your architecture on paper.
  • Practice safely in chaos workshops before jumping to heavyweight GameDays.

You don’t need perfect tooling to start. You need a whiteboard, the right people in the room, and the willingness to ask:

If this one little switch flips the wrong way, what trains might it send into what stations—and how bad would the crash really be?

Answer that regularly, and you’ll spend more time improving reliability—and less time writing post‑mortems about how “a small change” led to a very big outage.

The Analog Reliability Story Trainyard Compass: Hand‑Mapping How Tiny Routing Choices Turn Into Big Outages | Rain Lag