Back to blog

16 Jul 2025

Performance Optimisation: Turning Milliseconds Into Margin

A practical framework for linking engineering performance work to conversion, retention, and cost efficiency.

performanceoptimisationarchitecturereliabilityproduct

Performance optimisation cover

Performance engineering has a communication problem.

Engineers say "p95 latency" and "tail amplification." Business leaders hear "we need more time for technical tuning." The meeting ends with agreement that performance matters, followed by funding decisions that suggest the opposite.

That gap is avoidable.

Performance work gets investment when teams describe it as a commercial lever, not a technical hygiene task. Faster systems usually mean:

  • better conversion,
  • stronger retention,
  • lower support cost,
  • reduced infrastructure waste,
  • higher delivery confidence in peak periods.

If those links are explicit, optimization stops being optional.

The business case is stronger than most teams realize

A widely cited Deloitte analysis (published with Google) showed that very small speed improvements can move engagement and conversion metrics materially in retail scenarios. Exact impact varies by context, but the direction is consistent: user patience is short and performance is part of product quality.

In B2B and enterprise settings, the impact often appears in different metrics:

  • workflow completion rates,
  • support contact volumes,
  • task abandonment,
  • NPS drivers,
  • renewal confidence.

In short: performance is user trust, measured in milliseconds.

Why performance programs fail

Three recurring causes:

1) Isolated infrastructure focus

Teams tune servers while ignoring application logic, data access design, or client-side bottlenecks.

2) No journey-level prioritization

All endpoints are treated equally even though business value is concentrated in a few user journeys.

3) Success measured technically, not commercially

Latency improves, but no one tracks conversion, drop-off, or cost-to-serve effects.

A performance program without product context is expensive calibration.

A layered optimization strategy that actually scales

I use four layers.

Layer 1: Journey-Critical Experience

Prioritize user journeys tied directly to trust or revenue:

  • sign-in,
  • search/discovery,
  • checkout/transaction,
  • account and support pathways.

Set thresholds by customer impact, not engineering convenience.

Layer 2: Architecture and Data Flow

Address structural latency:

  • chatty service calls,
  • inefficient query patterns,
  • duplicated business logic,
  • weak caching strategy,
  • noisy cross-service dependencies.

Most severe latency issues are architecture issues wearing infrastructure clothes.

Layer 3: Runtime and Capacity Engineering

Now optimize:

  • autoscaling behavior,
  • queue management,
  • concurrency tuning,
  • payload size and compression,
  • CDN and edge strategy.

Only after layers 1 and 2 are understood.

Layer 4: Operational Readiness

Ensure performance remains stable during change:

  • release guardrails,
  • performance regression tests,
  • canary thresholds,
  • rollback criteria,
  • incident runbooks for degradation events.

Performance without operational resilience is temporary.

A practical metric stack

Track both technical and commercial indicators.

Technical indicators

  • p95/p99 latency by critical journey,
  • error rate under load,
  • saturation and queue depth,
  • cache hit ratio,
  • dependency timeout frequency.

Commercial indicators

  • conversion and completion rate by journey,
  • abandonment at latency breakpoints,
  • support contact volume by performance incident type,
  • infrastructure cost per successful transaction.

Tie these in one dashboard. If they live in separate systems, optimization decisions drift.

Lessons from peak-demand environments

In high-demand windows, performance debt surfaces brutally. Teams can ship features all year and still lose confidence if key journeys degrade under load.

In product environments with seasonal peaks, optimization work around checkout, identity, and mobile rendering often created more commercial value than additional feature volume. That is not anti-innovation. It is sequencing.

Sometimes the most strategic feature in Q4 is "the app still works when everyone shows up."

Engineering practices that reduce performance regressions

A few practices consistently work:

Performance budgets in CI

Define acceptable thresholds and fail builds for significant regressions in key flows.

Synthetic + real-user monitoring

Synthetic tests catch deterministic failures. RUM catches real-world variability.

Dependency SLIs and SLOs

Your service quality is bounded by dependency behavior. Make it visible.

Feature-flagged rollout for heavy changes

Allow quick rollback when performance impact exceeds thresholds.

Incident taxonomy for performance

Categorize incidents by root type (query, cache, third-party latency, capacity, client rendering). Trend analysis becomes actionable.

The cost narrative executives understand

Leaders often support performance investment when you show two curves:

  1. Cost of inaction:

    • revenue leakage,
    • support burden,
    • incident cost,
    • reputation impact.
  2. Cost of intervention:

    • engineering effort,
    • platform spend,
    • temporary delivery trade-offs.

When framed this way, performance investment becomes portfolio management, not technical preference.

AI-assisted optimization: useful and risky

AI tools can help performance engineering by:

  • identifying suspicious query patterns,
  • proposing refactoring paths,
  • summarizing profiling outputs,
  • generating baseline test scenarios.

But keep controls tight. AI-generated changes can optimize one metric while harming another. Human review remains mandatory for critical paths.

A funny but true lesson: never trust a "perfect" optimization PR that also claims to have improved everything by 40%. Nature is usually not that generous.

Product and engineering partnership model

Performance programs succeed when product and engineering co-own targets.

  • Product defines experience thresholds and user impact priorities.
  • Engineering defines technical pathways and risk controls.
  • Finance/ops sees cost implications.

Without this partnership, teams optimize for the loudest stakeholder of the week.

The first 90 days of a performance reset

If performance confidence is low, run this sequence:

Days 1-15

  • baseline journey metrics,
  • identify top bottlenecks,
  • map dependency risk.

Days 16-45

  • implement highest-leverage fixes,
  • add regression controls,
  • instrument cost-performance correlation.

Days 46-90

  • expand to second-tier journeys,
  • harden operational playbooks,
  • publish business impact summary.

This creates momentum and credibility quickly.

Common mistakes to remove immediately

  • treating p50 as success metric when users feel p95,
  • running load tests on unrealistic traffic patterns,
  • optimizing backend while frontend render remains slow,
  • shipping large changes without performance canaries,
  • delaying observability because "we know where the issue is."

You rarely know where the issue is.

Closing thought

Performance optimization is one of the few engineering disciplines where technical excellence and business value align naturally. Faster, more stable systems make customers happier and operations cheaper.

If your performance work feels underappreciated, improve the narrative and the measurement model, not just the query plan.

When performance is treated as a product capability, it compounds. When treated as occasional maintenance, it decays.

And yes, your future self will thank you for writing the rollback criteria before launch, not during the incident call.

Mobile Reality and the Backend Comfort Blanket

A recurring pattern in performance programs is backend overfocus. Backend profiling matters, but users experience total journey latency, not server pride.

In product environments with heavy mobile usage, major improvements often come from:

  • reducing client payload size,
  • minimizing render-blocking resources,
  • simplifying hydration and state initialization,
  • tuning image delivery and caching policies,
  • controlling third-party script impact.

The uncomfortable truth is that some of the worst customer pain sits in frontend orchestration and third-party overhead, while teams spend months optimizing internal APIs that were never the primary bottleneck.

One practical method is to run a monthly "journey teardown" from the customer device perspective. Record real-world sessions, identify where time is actually spent, and prioritize interventions accordingly. This discipline prevents teams from optimizing what is easy to measure instead of what users feel.

Performance Reviews That Influence Budget Decisions

If you want performance work to survive quarterly prioritization, package results in terms executives can act on:

  • what changed technically,
  • what changed for customer behavior,
  • what changed for commercial outcomes,
  • what risk remains and why.

A simple before/after narrative per critical journey is often more persuasive than large dashboards:

  1. Baseline pain and business impact.
  2. Interventions applied.
  3. Measured technical improvement.
  4. Measured customer/commercial effect.
  5. Next constraint to tackle.

This format helped in high-pressure delivery environments where capacity competition was intense. Performance work kept investment not because it was loudly defended, but because evidence showed it was one of the highest-ROI engineering activities available.

Performance optimization is ultimately a credibility game. When teams consistently connect milliseconds to money, customer trust, and operational resilience, prioritization becomes much easier.

Choosing Performance Bets Like a Portfolio Manager

When capacity is limited, pick optimization bets with explicit return logic:

  • expected user impact,
  • expected commercial impact,
  • implementation effort,
  • operational risk reduction.

This prevents the team from spending months on technically elegant changes that do not move customer outcomes. The best performance portfolios usually combine one high-impact journey bet, one architectural debt reduction bet, and one reliability hardening bet each quarter.

If teams keep this portfolio lens, performance work stops feeling like defensive maintenance and starts behaving like strategic growth enablement. That shift in perception is usually what unlocks sustained executive sponsorship.

It also makes performance planning conversations dramatically less adversarial.

Better conversations usually produce better technical decisions.

And better technical decisions protect margin.

That is the point.

References and Data