Rain Lag

The Analog Incident Story Mobile Harbor: Docking Past Outages in a Rotating Paper Port on Your Desk

How a simple rotating paper “harbor” on your desk can transform incident reviews from dry technical reports into living stories that improve reliability, resilience, and team culture.

The Analog Incident Story Mobile Harbor: Docking Past Outages in a Rotating Paper Port on Your Desk

What if your incident history didn’t just live in dashboards, tickets, and forgotten postmortems—but in a small, rotating paper “harbor” that sat on your desk? Each outage docks there like a tiny ship, carrying not just metrics and timelines, but feelings, tradeoffs, and lessons.

This is the idea behind the Analog Incident Story Mobile Harbor: a low-tech, physical way to explore reliability, culture, and failure. It’s part visualization, part memory palace, and part conversation starter. And surprisingly, it can help your team design more resilient systems.

In this post, we’ll explore why reliability is more than uptime, how emotional retrospectives reveal hidden truths, and how a simple paper port can make complex reliability concepts accessible to everyone.


Why Incidents Need More Than Technical Postmortems

Most incident reviews start the same way:

  • What broke?
  • When did it start?
  • How long did it last?
  • How do we prevent it from happening again?

Those are important questions, but they’re incomplete. They miss a crucial layer: how people felt and behaved before, during, and after the incident.

An emotional retrospective adds questions like:

  • When did people first sense something was off, even before alerts fired?
  • Who felt safe speaking up, and who stayed silent?
  • Which parts of the process felt clear, and which felt chaotic?
  • Where did people feel shame, blame, or fear of consequences?

These emotional signals often point to deeper issues:

  • A culture that punishes mistakes, so people hide early warning signs.
  • Hero culture—only a few people are seen as “fixers,” causing burnout and bottlenecks.
  • Confusing runbooks, unclear ownership, and noisy alerts nobody trusts.

By capturing emotion alongside facts, your incident history becomes more than a list of outages. It becomes a map of your team’s psychological safety, knowledge gaps, and process friction.

This is where an analog, visual tool like the Mobile Harbor shines. It forces you to slow down, tell the story, and make room for how people actually experienced the event—not just what Grafana or CloudWatch recorded.


Reliability Is More Than Uptime

We often use “reliability” as shorthand for "the service is up." But real reliability is more nuanced. It includes:

  • Resilience: How your system behaves during and after failures.

    • Fault tolerance – Can it absorb certain failures without users noticing?
    • Graceful degradation – If something must break, does it break softly?
    • Recovery – How quickly and safely can it return to a good state?
  • Availability: Do users get the promised level of quality when they expect it?

    • Not just “HTTP 200” responses, but fast enough, correct enough, often enough.

Reliability isn’t binary; it’s always about:

What did we promise? To whom? Under what conditions? And did we keep that promise?

An incident might look small technically but be devastating for user trust. Or it might be huge internally (a scary internal system outage) but barely visible to customers. Your reliability work should be guided by clear, negotiated expectations—not a vague “be more reliable.”


Assuming Failure: Designing for the Inevitable

Modern systems are too complex to be failure-free. You must assume that at some point:

  • A critical component will malfunction.
  • A cloud region, third-party API, or database will experience an outage.
  • Performance will degrade under an unexpected workload.
  • You’ll hit resource limits you didn’t anticipate.

That’s not pessimism; it’s realism.

Reliability, then, can’t be “bolted on” later. It must be designed in from the start, across:

  • Code – Timeouts, retries with backoff, idempotent operations, input validation.
  • Infrastructure – Redundancy, failover plans, capacity planning, isolation.
  • Operations – Runbooks, incident drills, clear roles, escalation paths.

The analog harbor metaphor helps here: imagine each failure mode as a “ship” that will try to dock someday. Do you have:

  • A berth prepared (a known response pattern)?
  • Tugboats ready (tools, runbooks, people)?
  • A lighthouse (alerts and dashboards) to see it coming?

Assuming failure doesn’t mean you accept chaos. It means you accept reality—and then calmly design for it.


Defining “Good Enough”: Reliability Requirements That Guide Design

You can’t design for reliability in a vacuum. You need negotiated reliability requirements that describe:

  • User experience expectations – How slow is too slow? Which flows are mission-critical vs. nice-to-have?
  • Data needs – Can data be eventually consistent, or must it be strongly consistent? What’s the acceptable data loss window, if any?
  • Workflow traits – Are users batch-processing overnight or depending on real-time interactions?
  • Unique workload behaviors – Spiky traffic? Long-running jobs? Heavy writes? High read volume?

When these are explicit, they:

  • Guide architecture choices (e.g., multi-region vs. single-region, caching strategy, queuing).
  • Clarify tradeoffs (e.g., latency vs. consistency, speed vs. cost, simplicity vs. redundancy).
  • Define what “good enough” looks like before, during, and after failures.

In the Mobile Harbor, each incident card can include:

  • Which reliability requirement was violated.
  • How badly.
  • Whether the original requirement was realistic.

Sometimes you’ll realize the problem wasn’t just the outage; it was a mismatch between what users needed and what you promised.


Observability and Shared Visibility: Seeing the Storm Together

Fast detection and thoughtful learning depend on observability and shared visibility:

  • Metrics – Quantitative signals: latency, error rates, saturation, throughput.
  • Logs – Detailed event records for context and debugging.
  • Traces – End-to-end view across services to understand where time and errors occur.
  • Alerts – Curated, actionable signals that something meaningful is wrong.

But tools alone aren’t enough. You also need:

  • Shared views – Dashboards that different teams can read and reason about together.
  • Shared language – Agreed names for services, dependencies, and user journeys.
  • Shared ownership – Ops, dev, product, and support seeing the same reality, not arguing from different slices.

Your analog harbor can help here by:

  • Listing the dashboards or graphs that were (or weren’t) helpful during the incident.
  • Marking which signals fired too late, too often, or not at all.
  • Showing how many teams needed to coordinate and what visibility they had.

When people can literally point at a paper card and say, “This graph lied to us,” it becomes much easier to refine observability than when complaints are buried in a closed incident ticket.


Simplicity vs. Single Points of Failure: Finding the Balance

“Simplify the system” is great advice—until you realize you’ve created a giant single point of failure disguised as a clean abstraction.

Simplicity helps reliability because it:

  • Reduces the total surface area for bugs and misconfigurations.
  • Makes mental models easier to hold and share.
  • Speeds up debugging and onboarding.

But oversimplification can:

  • Centralize too much responsibility in one service, one team, or one database.
  • Hide complex failure modes behind "magic" components.
  • Make it impossible to degrade gracefully when that central piece fails.

On your Mobile Harbor, each incident ship can note:

  • Did complexity cause this (too many moving parts, hard-to-reason interactions)?
  • Or did oversimplification cause this (one service doing too much, a single database or queue everyone depends on)?

This keeps the team honest: you’re not just chanting “microservices” or “monolith” as ideology—you’re actually tracking how design choices behave in the real sea of production.


Prototyping Reliability with a Rotating Paper Port

So how does the Analog Incident Story Mobile Harbor actually work?

Imagine a small, rotating cardboard or wooden base on your desk—a lazy Susan of incident memory. Around the edge, like docks in a circular harbor, you place incident cards or mini “ships.” Each one represents a real event.

Each card might include:

  • Name: A memorable, human-readable incident name ("The Tuesday Timeout Storm").
  • Timeline: Start, detection, mitigation, resolution.
  • Impact: Users affected, workflows broken, promises violated.
  • Reliability aspects: Availability, resilience behaviors, recovery patterns.
  • Observability notes: What helped, what misled you.
  • Design questions: What did this reveal about assumptions, complexity, or architecture?
  • Emotional snapshot: A few words about how the team felt—confused, frustrated, proud, panicked, calm.

You can:

  • Rotate the harbor during team meetings and pick a ship at random to review.
  • Cluster similar incidents together (e.g., all caused by timeouts, all related to a specific dependency).
  • Use stickers or colored markers to highlight patterns (e.g., red for observability gaps, blue for unclear ownership).

This physical model is a prototype and teaching tool:

  • For new joiners: they can “walk the harbor” and get a visceral feel for your reliability history.
  • For stakeholders: it turns abstract reliability into tangible stories they can touch.
  • For engineers: it surfaces patterns that may not pop out in a linear list of tickets.

Suddenly, reliability isn’t a vague aspiration. It’s a living, rotating museum of past storms and future improvements.


Conclusion: Dock Your Past to Navigate Your Future

Reliable systems aren’t defined only by clean architecture diagrams or stellar uptime percentages. They’re shaped by:

  • How you experience incidents as people.
  • How you negotiate expectations with users.
  • How you design for failure instead of pretending it won’t happen.
  • How you observe reality across teams, not just in silos.
  • How you balance simplicity with redundancy and flexibility.

The Analog Incident Story Mobile Harbor is a playful but powerful way to bring all of this into view. By docking your past outages in a rotating paper port on your desk, you make space for:

  • Emotional retrospectives that uncover culture and process issues.
  • Concrete discussions about resilience, availability, and tradeoffs.
  • Collaborative learning that includes engineers, product, support, and leadership.

You don’t need fancy hardware to build more reliable software. Sometimes, all it takes is a pen, some paper, a rotating stand—and a willingness to look honestly at the storms you’ve already sailed through.

Your incidents are already telling stories. The question is: will you listen, learn, and redesign—before the next ship comes into harbor?

The Analog Incident Story Mobile Harbor: Docking Past Outages in a Rotating Paper Port on Your Desk | Rain Lag