Why 70% of Platform Engineering Teams Fail (And the 5 Metrics That Predict Success)

· 24 min read

Quick Answer (TL;DR)

Problem: 60-70% of platform engineering teams fail to deliver impact, with 45% disbanded within 18 months.

Root Cause: Most teams hire engineers when they need product managers. Only 33% have PMs despite 52% saying they're crucial.

Key Statistics:

  • 45% don't measure anything (can't prove ROI)
  • 10% average Backstage adoption vs Spotify's 99%
  • DORA: 8% throughput DROP, 14% stability DROP initially
  • Platform teams cost $450K-$3.75M annually

Success Metrics: Product Manager (Y/N), Developer NPS over 20, Voluntary adoption over 50%, Time to value under 30 days, Pre-platform baseline exists

When NOT to Build: Under 100 engineers, no real pain yet, can't dedicate 3-5 FTE

Full analysis below ↓

🎙️ Listen to the podcast episode: Why 70% of Platform Engineering Teams Fail - Jordan and Alex discuss the critical PM gap, metrics blindness, and the Monday morning action plan to get your platform team on track.


The Knight Capital Story

August 1, 2012. 9:30 AM Eastern. Knight Capital Group deploys new trading software.

By 10:15 AM, the system had sent $7 billion in unintended orders to the market. $440 million evaporated in 45 minutes—because their deployment platform lacked proper validation and rollback capabilities.

Fast forward to 2025. Platform engineering is supposed to prevent exactly these disasters. 83% of organizations now have platform initiatives underway. 80% of large enterprises will have platform teams by 2026.

Yet here's the uncomfortable truth: 60-70% of platform engineering teams fail to deliver impact. 45% are disbanded or restructured within 18 months.

Not because they chose the wrong tools. Not because Kubernetes is too hard. But because they made one critical mistake: they hired engineers when they needed product managers.

Only 33% of platform teams have PMs, despite 52% saying PMs are crucial to success. That 19-point gap? That's your failure rate right there.


Key Statistics (2024-2025 Data)

StatisticValueSourceContext
Platform failure rate60-70% fail to deliver impactThe New Stack, Fast Flow Conf 2025Not all fail completely, but majority can't demonstrate ROI
Disbandment timeline45% within 18 monthsThe New Stack, Team TopologiesDisbanded or restructured due to lack of ROI proof
Product Manager gapOnly 33% have PMsPuppet State of DevOps 2023Despite 52% saying PMs are "crucial" (2024)
Zero measurement45% measure nothingState of Platform Engineering 2024Can't prove value when leadership asks
Backstage adoption10% average (vs 99% at Spotify)Helen Greul, SpotifyProduct thinking vs engineering thinking
Initial performance impact8% throughput drop, 14% stability drop2024 DORA ReportPlatforms make things worse before better
Team size threshold100+ engineers for meaningful ROIMultiple sourcesBelow this, overhead exceeds benefit
Annual platform cost$450K (3 people) to $3.75M (25 people)Calculated from researchVs $1.5M-$18M productivity gains

The Failure Statistics

The 70% Number: What It Really Means

The statistics are sobering. 60-70% of platform projects fail to deliver impact according to research presented at Fast Flow Conf 2025 and reported by The New Stack. Even more concerning: 45% are disbanded or restructured within 18 months.

This isn't universal failure. Many platforms technically work—they run, they deploy applications, they have users. The failure is in demonstrating ROI and achieving meaningful adoption. When budget reviews arrive or leadership changes, these platforms can't justify their existence.

Consider that 55% of platform teams are less than 2 years old. Most organizations are still in the early, immature phase where failure patterns emerge.

The DORA Paradox

Platform teams were supposed to increase developer productivity. The reality tells a different story.

The 2024 DORA Report found that platform teams initially decreased throughput by 8% and change stability by 14%. Organizations saw only a 6% overall performance gain initially—far below expectations.

Why? Initial complexity, learning curves, forced migrations, and the overhead of new abstractions. Platforms make things worse before they make them better—a reality most platform teams fail to communicate to stakeholders.

The Adoption Crisis

The most damning statistic comes from Backstage adoption data. External Backstage adopters achieve 10% average adoption while Spotify—the creator—achieved 99% voluntary adoption.

Same tool. 89-point adoption gap.

This isn't a technology problem. It's a product management problem. Organizations invest heavily in beautiful platforms only to achieve 5-10% usage rates because nobody chose them. One-third of teams struggle with resistance to adoption, with many platforms seeing only 20% workload migration after significant investment.

💡 Key Takeaway

Only 33% of platform teams have product managers, yet 52% say PMs are crucial. This 19-point gap explains why most platforms fail. You can't engineer your way out of a product management problem.


The 7 Failure Modes

1. Lack of Product Mindset

The Problem:

68% of platform teams lack product thinking—only 32.3% follow platform-as-a-product approaches. They treat platforms as IT projects instead of products, with no user research, feedback loops, or adoption metrics. They build features engineers want, not what developers need.

Real Example:

A large insurance company featured in a Microsoft case study built a technically beautiful platform. Voluntary adoption? Zero. They had to mandate usage and tie it to performance reviews—a clear product failure signal.

Multiple organizations report building "amazing platforms" but achieving only 20% workload migration after significant investment. The platform exists, but developers actively avoid it.

Impact:

Low voluntary adoption (the 10% Backstage average), platform teams become bottlenecks instead of accelerators, and wasted engineering investment with no business value.

The Fix:

  • Hire a Platform Product Manager (only 33% currently have one)
  • Treat developers as customers with choices: User research, NPS surveys, feedback loops
  • Voluntary adoption over mandates: Build golden paths so compelling developers choose them
  • Spotify's approach: Quarterly developer interviews, NPS tracking, achieved 99% voluntary adoption

Success Metric: Product Manager role exists (Y/N), NPS over 20

Platform Team Performance: With vs Without Product Manager

MetricWith PM (33% of teams)Without PM (67% of teams)Impact
Voluntary Adoption50%+ typical (Spotify: 99%)10% average (Backstage external)3x adoption rate
Disbandment Rate~23% within 18 months~60% within 18 months2.6x survival rate
NPS Score20-76 (Spotify)Under 0 typicalProduct thinking drives satisfaction
Time to ValueUnder 30 days achievable60+ days commonPM prioritization speeds delivery
ROI ProofCan demonstrate with metrics45% measure nothingPM establishes baselines

Source: Aggregated from Puppet State of DevOps 2023, DORA 2024, Backstage adoption data


2. Poor Metrics / Can't Prove ROI

The Problem:

45% of platform teams don't measure anything at all. Measurement becomes an afterthought. They focus on outputs (features built) instead of outcomes (developer productivity). When leadership asks for ROI, they have no answer.

Real Example:

Knight Capital's $440 million loss in 45 minutes resulted partly from lack of deployment validation metrics. Many platform teams report "high success" because they're not collecting contradictory data—a zero-metric success illusion that shatters during budget reviews.

Impact:

Platform teams get disbanded when they can't justify budget during economic downturns. It becomes impossible to demonstrate whether the platform helps or harms (recall: DORA found 8% throughput decrease). Leadership questions ROI and cuts funding.

The Fix:

  • Establish baseline BEFORE platform launch: DORA metrics, deployment frequency, lead time, developer NPS
  • Track business outcomes, not just adoption: Time saved, cost reduction, security incidents prevented
  • The benchmarkable trio: Market share (% teams using platform), onboarding time, NPS score
  • Example ROI: One organization calculated 156.5% ROI ($3.7M net effect) for onboarding 10 teams/year over 3 years
  • Quarterly measurement cadence: NPS surveys every quarter, DORA metrics continuously

Success Metric: Pre-platform baseline exists (Y/N), Can prove ROI with numbers

📊 Need help calculating ROI? Check out the Platform Engineering ROI Calculator with proven frameworks, templates, and real data from 5 company sizes to prove value to executives.


3. Forced Adoption ("Build It and They Will Come")

The Problem:

"Build it and they will come" is not viable for internal developer platforms. Mandated usage leads to resentment, shadow infrastructure, and reduced trust. Developers work around the platform, ignore golden paths, or actively advocate against adoption.

Real Example:

The Backstage adoption crisis tells the story: 10% average vs Spotify's 99%—same tool, different approach. The insurance company mentioned earlier mandated platform usage and tied it to performance reviews—a morale disaster.

90% of external Backstage adopters fail to achieve meaningful adoption despite the tool's technical excellence.

Impact:

Developers build shadow infrastructure to avoid the platform. Low usage makes the platform economically unviable. Platform team morale crashes as they watch their work ignored.

The Fix:

  • Make compliance the easiest path: Golden paths more convenient than alternatives
  • Start with developer pain points: Talk to developers across teams, identify common problems
  • Pilot with friendly teams: Build confidence before scaling
  • No mandates—ever: Voluntary adoption is the only sustainable approach
  • Spotify's approach: Started with highest-value use cases, invested in developer relations, marketed internally, achieved 99% voluntary adoption after ~1 year

Success Metric: Voluntary adoption over 50% within 12 months

💡 Key Takeaway

The Backstage Paradox: Spotify achieves 99% voluntary adoption with product thinking. External orgs average 10% with engineering thinking. Same tool, 89-point gap—the difference is product management.


4. Misaligned Stakeholders / No Executive Sponsor

The Problem:

Platform teams treated as "after-hours side projects" without leadership backing usually stall within months. No executive sponsor means no one to secure funding, resources, or strategic alignment. 45% are disbanded within 18 months—often during budget cuts without executive protection.

Real Example:

A Reddit discussion captured this perfectly: "Shiny new CTO mandated shift from 'DevOps' to 'Developer Experience' with no explanation"—classic stakeholder misalignment. Teams without executive sponsors lose budget during economic downturns. Platform leads who can't "speak the language" of each stakeholder group (engineering, product, finance, leadership) get caught in political crossfire.

Impact:

Platform team sandwiched between conflicting priorities, budget cuts during economic uncertainty, team disbanded before achieving meaningful impact.

The Fix:

  • Secure executive sponsor from day one: Senior leader who provides direction and secures budget
  • Align platform metrics with business goals: Show how platform supports revenue, customer satisfaction, or cost reduction
  • Head of Platform Engineering role: Responsible for stakeholder alignment, budget planning, North Star Metrics
  • Quarterly business reviews: Present platform impact in business terms (time saved → features shipped faster → revenue)

Success Metric: Executive sponsor identified (Y/N), Quarterly business reviews scheduled


5. Over-Engineering (Too Complex, Too Early)

The Problem:

"Boiling the ocean"—trying to solve every problem, address every edge case. Teams spend months engineering bulletproof systems, then discover unforeseen failures anyway. The guidance to "build for 90% of use cases" is wise—the remaining 10% introduces exponential complexity.

The 2024 DORA Report confirms this: platform teams decreased throughput 8% and stability 14% because complexity slows teams down.

Real Example:

Uber's "Death Star" architecture scaled to 1,000+ microservices with tangled dependencies and overlapping APIs—so complex engineers gave it that nickname. Spotify's early Hadoop single point of failure brought down their entire event delivery system.

Teams build Backstage for 50 engineers or deploy service mesh before having microservices. The "perfect platform covering everything" results in cluttered systems that don't satisfy a single use case well.

Impact:

Development velocity DECREASES instead of increasing. The platform becomes a maintenance burden instead of productivity multiplier. Developers find the platform more complex than the AWS console. Long time-to-value makes the platform vulnerable to budget cuts.

The Fix:

  • Start small and stay lean: "A good platform is just big enough, but not bigger"
  • Ruthless simplification: Every configurable component creates mental overhead
  • Opinionated simplicity: Make decisions FOR developers rather than exposing all options
  • Build backend before UI: Focus on solid APIs and orchestration before adding UI
  • Golden paths for 80% of teams: Don't try to support every workload
  • Start with: Streamline basic developer touchpoints and CI/CD, then introduce complexity incrementally

Success Metric: Time to first value under 30 days, Complexity score (number of required steps under 20)

💡 Key Takeaway

DORA found platform teams decreased throughput by 8% and stability by 14% initially. Platforms make things worse before they make them better—plan accordingly and set stakeholder expectations.


6. Wrong Team Structure / Skills Gap

The Problem:

Underestimating maintenance effort is the #1 Backstage adoption pitfall. Organizations need an average 2.5 engineers minimum (frontend, backend, part-time DevOps) for Backstage alone. Finding good platform engineers is challenging—there are limited candidates passionate about infrastructure.

Teams trying to build developer portals without frontend engineers consistently fail. Only 33% have product managers—they have technical chops without product sense.

Real Example:

Backstage adoption failures often trace to teams attempting deployment with 1-2 backend engineers when they need frontend + backend + DevOps minimum. The "Platform as a Bucket" antipattern emerges: the platform becomes a dumping ground, a huge mess with lack of ownership and focus.

The insurance company with forced adoption had technical skills but no PM—they couldn't sell the value proposition, so they mandated it instead.

Impact:

Platform becomes a bottleneck instead of accelerator. High turnover and burnout from overload. Platform lacks required capabilities (decent UI, good documentation). Team can't balance competing demands.

The Fix:

  • Balanced team composition: DevEx engineers (developer delegates) + Infrastructure engineers (I&O delegates) + Product Manager
  • Typical platform team: 3-5 people minimum, 6-10 for supporting multiple dev teams
  • Required roles: Software engineers (45%), platform engineers (44%), developers (40%), product/project managers (37%), I&O pros (35%), SREs (16%)
  • Hire for product mindset: Engineers who understand developer pain points
  • Include frontend skills if building portal: Don't assume backend engineers can build good UIs
  • Platform Product Manager (crucial): Leads strategy, roadmap, user research, feedback loops

Success Metric: Team has PM + frontend + backend + ops (Y/N), Team size 3-5 minimum


7. Ignoring Developer Experience (DX)

The Problem:

Platform teams don't realize UX is as important as existence of self-service. Developers hate losing access to underlying technologies for the sake of abstraction. Overly complex platforms with unfamiliar config formats, no documentation, and inconsistent APIs drive developers away. Ticket-based workflows lead to bottlenecks and reduce autonomy.

"Engineers went back to AWS console" is a common refrain—the platform was harder to use than going direct.

Real Example:

Backstage's 10% adoption problem stems partly from platform complexity and lack of clear prioritization—developers get overwhelmed by sheer volume of data with no alerting or task management.

Platform teams build technically perfect platforms developers hate using. Over-abstraction creates a "dangerous divide between platform engineers and developers" where developers need to tweak configs but find themselves locked out. One organization built beautiful infrastructure but developers still used shadow IT because the platform UX was terrible.

Impact:

Developers avoid the platform entirely or use it minimally. Shadow infrastructure proliferates. Platform team confused why "nobody appreciates our work." Low NPS scores (can be under -20 for legacy platforms).

The Fix:

  • Developer-centric design from day one: Involve app dev teams from the beginning
  • Measure developer satisfaction: NPS surveys quarterly, qualitative feedback loops
  • Good NPS benchmarks: Over 0 is good, over 20 is great, over 50 is amazing, over 80 is top percentile (Spotify Backstage achieved 76)
  • Focus on lowest-scoring pillar: Organizations should spend 80% of time on pillar with worst NPS
  • Build golden paths that are easiest path: Compliance should be more convenient than non-compliance
  • Documentation and support: Tutorials, help channels, peer learning
  • Start with developer pain points: Talk to developers across teams, find common pain points platform could remove
  • Avoid over-abstraction: "Build golden paths, not cages"

Success Metric: Developer NPS over 20, Documentation coverage over 80%, Support response time under 4 hours


The Success Framework: 5 Predictive Metrics

Metric 1: Product Manager Exists (Y/N)

Why It Matters: Only 33% have PMs, yet 52% say crucial. This gap predicts failure.

Binary Predictor:

  • ✅ Has PM → 3x more likely to achieve over 50% adoption
  • ❌ No PM → 67% more likely to be disbanded within 18 months

What Good Looks Like:

  • Dedicated Platform Product Manager (not part-time)
  • Owns roadmap, prioritization, adoption metrics
  • Conducts quarterly user research
  • Builds tight feedback loops with developer teams

Metric 2: Pre-Platform Baseline Established (Y/N)

Why It Matters: 45% don't measure anything. Can't prove ROI without baseline.

What to Measure BEFORE Platform Launch:

  • DORA metrics: Deployment frequency, lead time, change failure rate, MTTR
  • Developer satisfaction: NPS score, qualitative feedback
  • Onboarding time: Days to first commit for new engineers
  • Incident frequency and resolution time
  • Cloud/infrastructure costs

What Good Looks Like:

  • Baseline measured 3-6 months before platform launch
  • Metrics documented and shared with stakeholders
  • Clear targets defined (e.g., "reduce deployment time from 45min to 10min")
  • Quarterly measurement cadence established

Metric 3: Developer NPS over 20

Why It Matters: NPS predicts adoption. Under 20 = developers avoiding platform. Over 50 = amazing.

NPS Benchmarks:

  • Under -20: Legacy platforms, developers actively avoid
  • 0-20: Platform exists but developers neutral/slightly positive
  • 20-50: Good platform, developers recommend to peers
  • 50-80: Amazing platform, developers evangelize internally
  • Over 80: Top percentile (rare, Spotify Backstage achieved 76)

Developer NPS Score Interpretation

NPS RangeStatusWhat It MeansAction Required
Under -20🔴 CriticalDevelopers actively avoid platform, shadow IT proliferatingEmergency intervention: Remove mandates, conduct user research, simplify immediately
-20 to 0🟠 PoorPlatform exists but developers neutral/slightly negativeMajor improvements needed: Focus on DX, gather feedback, fix top 3 pain points
0 to 20🟡 AcceptablePlatform functional, developers slightly positiveContinue improving: Address feedback, optimize golden paths
20 to 50🟢 GoodDevelopers recommend to peers, voluntary adoption growingMaintain quality: Keep feedback loops, prioritize based on usage
50 to 80🟢 AmazingDevelopers evangelize internally, high voluntary adoptionBest practice: Document success, share learnings, sustain excellence
Over 80🏆 Top PercentileExceptional (rare) - Spotify Backstage achieved 76Industry leading: Publish case studies, share at conferences

Measurement: Quarterly NPS surveys to all developers who've interacted with platform. Source: DX research and Spotify Backstage data.

What Good Looks Like:

  • Quarterly NPS surveys sent to all developers
  • Follow-up interviews with detractors to understand pain points
  • 80% of platform team time spent on lowest-scoring pillar
  • Visible improvements quarter-over-quarter

Metric 4: Voluntary Adoption over 50% Within 12 Months

Why It Matters: Forced adoption = failure. Spotify achieved 99% voluntarily.

Adoption Phases:

  • Months 0-3: Pilot with friendly teams (target: 10-20% adoption)
  • Months 3-6: Early majority (target: 30-40% adoption)
  • Months 6-12: Mainstream (target: over 50% adoption)
  • Months 12-18: Laggards (target: over 70% adoption)

What Good Looks Like:

  • No mandates or top-down enforcement
  • Golden paths so compelling developers choose them
  • Platform team focuses on highest-value use cases first
  • Internal marketing and developer relations investment
  • Success stories documented and shared

Metric 5: Time to First Value Under 30 Days

Why It Matters: Long time-to-value makes platform vulnerable to budget cuts. Complexity kills.

Time to Value Milestones:

  • Day 1: Developer can access platform
  • Day 7: Developer deploys first "hello world" app
  • Day 14: Developer deploys real workload
  • Day 30: Developer sees productivity improvement (faster deployments, better observability, etc.)

What Good Looks Like:

  • Onboarding docs clear enough for junior engineer to follow
  • Self-service capabilities (no ticket required)
  • Quick wins visible early (e.g., automated deployments, easy rollbacks)
  • Feedback loop: Measure onboarding time, optimize based on data

💡 Key Takeaway

If you have fewer than 100 engineers, a dedicated platform team likely costs more than it saves. Wait for real pain (snowflake environments, DevOps helpdesk, onboarding over 1 week) before investing $450K-$3.75M annually.


Decision Framework: Should You Build a Platform Team?

Team Size Threshold: 100+ Engineers

Why This Matters:

  • Under 10 engineers: Not worth it (platform overhead exceeds benefits)
  • Around 30 engineers: Platform work becomes someone's full-time job
  • 100+ engineers: Meaningful positive effect possible
  • 500+ engineers: Platform team becomes critical

The Math:

  • 3-person platform team costs around $450K annually
  • Needs to save over $450K in developer productivity to break even
  • At 50 engineers: $9K/engineer/year savings required—difficult
  • At 100 engineers: $4.5K/engineer/year savings—achievable
  • At 1000 engineers: $450/engineer/year savings—easy

At scale, the ROI becomes compelling. A 25-person platform team costs $3.75M annually but can deliver $18M in productivity gains—380% ROI. But you need the engineering headcount to justify that investment.

When to Build: ROI by Organization Size

Organization SizeRecommendationAnnual CostBreak-Even SavingsROI Potential
Under 50 engineers❌ Don't build$450K (3 people)$9K/engineer/yearNegative - overhead exceeds benefit
50-100 engineers⚠️ Consider 1 platform engineer$150K (1 person)$1.5K-3K/engineer/yearMarginal - wait for pain
100-500 engineers✅ Build 3-5 person team$450K-750K$900-4.5K/engineer/yearPositive - achievable ROI
500-1000 engineers✅ Build 6-10 person team$900K-1.5M$900-3K/engineer/yearStrong - clear value
1000+ engineers✅ Build 15-25 person team$2.25M-3.75M$2.25K-3.75K/engineer/yearExceptional - 380% ROI

Note: Costs assume $150K fully-loaded per platform engineer. ROI calculated from DORA productivity improvements and time savings.


Pain Point Checklist

Build Platform Team If:

  • ✅ Snowflake environments (every team deploys differently)
  • ✅ DevOps helpdesk (platform engineers fielding constant questions)
  • ✅ Onboarding over 1 week (new engineers can't be productive quickly)
  • ✅ Deployment frequency under 1/day (slow, manual processes)
  • ✅ Multi-cloud complexity (managing AWS + Azure + GCP)
  • ✅ Compliance requirements (SOC 2, HIPAA, etc.) applied inconsistently

DON'T Build Platform Team If:

  • ❌ Under 100 engineers total
  • ❌ No real pain yet (just following trends)
  • ❌ Can't dedicate 3-5 FTE to platform team
  • ❌ No executive sponsor secured
  • ❌ Can't measure ROI or establish baselines

When NOT to Build

Be Brutally Honest:

Most platform initiatives are premature optimization. You're probably too early if you're still figuring out product-market fit. Wait for real pain before investing—pain is your signal that the investment will pay off.

Alternatives:

  • Use managed platforms (Render, Fly.io, Railway) until you hit their limits
  • Hire DevOps engineer embedded in product teams instead of separate platform team
  • Start with documentation and best practices before building tooling

The Cost of Failure:

  • $450K-$3.75M annually wasted if platform disbanded
  • 18 months of engineering time lost
  • Team morale damage
  • Trust erosion ("leadership doesn't know what they're doing")

💡 Key Takeaway

45% of platform teams don't measure anything. By the time leadership asks for ROI during budget reviews, it's too late to establish baselines. Measure BEFORE you build.


The Remediation Playbook

If You're in the 70%: Early Warning Signs

Warning Sign 1: Low Adoption (under 30% after 6 months)

  • Action: Conduct user research with non-adopters
  • Ask: "Why aren't you using the platform?"
  • Fix: Address top 3 pain points identified
  • Timeline: 30 days to implement fixes

Warning Sign 2: No Metrics Being Tracked

  • Action: Establish baseline TODAY (not when platform "ready")
  • Measure: DORA metrics, developer NPS, onboarding time
  • Fix: Quarterly measurement cadence
  • Timeline: 14 days to first measurement

Warning Sign 3: Developers Complaining About Complexity

  • Action: Simplification sprint
  • Cut: Remove features with under 10% usage
  • Focus: Golden path for 80% of teams
  • Timeline: 60 days to simplified v2

Warning Sign 4: No Product Manager

  • Action: Hire Platform PM immediately (or reassign existing PM)
  • Alternative: Train platform engineer in product thinking
  • Focus: User research, feedback loops, roadmap prioritization
  • Timeline: 90 days to PM onboarded

Warning Sign 5: Forced Adoption / Mandates

  • Action: Remove mandates immediately
  • Apologize: Acknowledge mistake to developer community
  • Rebuild: Focus on making platform so good developers choose it
  • Timeline: 6 months to restore trust

If You're Starting: Avoid These Mistakes

Month 0: Before You Begin

  • Secure executive sponsor (VP+ level)
  • Establish baseline metrics (DORA, NPS, onboarding time)
  • Hire Platform Product Manager (or train engineer in product thinking)
  • Define success criteria (adoption %, NPS target, ROI target)
  • Validate pain points (talk to 20+ developers across teams)

Month 1-3: MVP

  • Build for 1-2 highest-value use cases only
  • Pilot with 2-3 friendly teams
  • Weekly feedback sessions with pilot teams
  • Iterate based on feedback (not your assumptions)
  • Document golden path (onboarding guide)

Month 4-6: Scale

  • Achieve over 50% adoption with pilot teams
  • Measure NPS (target: over 20)
  • Expand to early majority teams (voluntary only)
  • Build internal case studies (time saved, productivity gains)
  • Quarterly business review with leadership (show ROI)

Month 7-12: Mainstream

  • Achieve over 50% overall adoption
  • Establish quarterly measurement cadence
  • Hire additional platform engineers as needed
  • Focus 80% of time on lowest NPS pillar
  • Plan for sustainability (documentation, support, maintenance)

💡 Key Takeaway

"Build it and they will come" is not viable for IDPs. Forced adoption leads to resentment and shadow infrastructure. Spotify achieved 99% adoption voluntarily—no mandates, just compelling golden paths.


Conclusion: The Product Manager Problem

The Central Thesis:

68% of platform teams don't follow platform-as-a-product approach. Only 33% have product managers, yet 52% say PMs are crucial. This 19-point gap explains the 45% disbandment rate.

You can't engineer your way out of a product management problem.

The Uncomfortable Truth:

Platform engineering is 20% technology, 80% product management. Choosing Kubernetes vs ECS doesn't matter if nobody uses your platform. Backstage is technically excellent—the 10% adoption rate proves it's a product problem, not a technical problem.

You hired engineers when you needed a product manager.

The Path Forward:

  1. Hire Platform Product Manager (or train engineer in product thinking)
  2. Establish baseline metrics BEFORE launch
  3. Build for 90% of use cases, ruthlessly simplify
  4. Voluntary adoption only—no mandates ever
  5. Measure NPS quarterly, fix lowest-scoring areas

The 5 Predictive Metrics (Checklist):

  • Product Manager exists (Y/N)
  • Pre-platform baseline established (Y/N)
  • Developer NPS over 20
  • Voluntary adoption over 50% within 12 months
  • Time to first value under 30 days

If you can check all 5, you're in the 30% that succeed. If you can't, you're in the 70% that fail.

The question isn't whether platform engineering works. It does—when treated as a product, not a project.

The question is: Will you hire a product manager before leadership asks for ROI?

💡 Key Takeaway

Platform engineering is 20% technology, 80% product management. Choosing Kubernetes vs ECS doesn't matter if nobody uses your platform. Hire the product manager first, then the engineers.


📚 Learning Resources

Platform-as-a-Product

Product Management for Platforms

Case Studies

Metrics and Measurement



Last updated: October 28, 2025