Back to Blog

When AI Masters D, P Becomes Your Career Leverage

Mar 03, 2026 · 9 min read · Harsha Cheruku

If AI writes the code, drafts the PRD, summarizes the research, and generates ten strategy options in 30 seconds…

what exactly are you getting paid for?

That’s not a rhetorical scare tactic. It’s the core career question of this decade.

In The D/P Framework, I argued that every workflow is a mix of deterministic (D) and probabilistic (P) steps.

In Life in D/P, I argued this pattern isn’t just product architecture — it’s basically life.

This post is the uncomfortable version:

  • AI is compressing D work fast
  • P work is becoming the bottleneck
  • P work is also where the risk lives — and the cost of getting it wrong compounds

So this is not just a productivity shift. It’s a responsibility shift.


The Great D/P Inversion

For years, most careers rewarded high-volume deterministic output:

  • write code
  • run analyses
  • produce docs
  • execute process

Now AI does more of that in less time.

Which means the bottleneck moves upstream:

  • Which problem is worth solving?
  • Which trade-off are we willing to live with?
  • Which risk is acceptable?
  • Which option do we commit to?

That upstream work is P.

So the real change is not “AI is faster.” The real change is where value is created.


Why P Work Is More Valuable (and More Dangerous)

People call P work “ambiguous.” True. But that word is too polite.

P work is expensive because the cost of error is asymmetric.

D errors are bounded

A failing test. A broken endpoint. A wrong SQL query. A malformed payload.

Bad? Sure. But: - Detectable — tests, linters, monitoring catch them - Bounded — the blast radius is usually one component, one feature - Reversible — fix the bug, deploy the patch, move on

D errors cost hours or days. Rarely more.

P errors compound

Building the wrong feature for 2 quarters. Choosing the wrong market. Hiring into the wrong org model. Optimizing the wrong north star metric.

These aren’t bugs. These are commitments that cascade.

Here’s the compounding problem: a wrong P decision doesn’t just fail in isolation. It poisons the D work downstream.

Pick the wrong market? Now: - Every feature built for that market is wasted effort - Every hire optimized for that market is misaligned - Every quarter spent there is opportunity cost against the right market - The data you collected validates the wrong assumptions

A bad D decision costs you a sprint. A bad P decision costs you a year — and you often don’t know it was wrong until the year is over.

D mistakes are defects. P mistakes are bets that went wrong.

And that’s the crux: P-heavy work carries higher cognitive load not because it’s hand-wavy, but because the downside compounds silently. By the time you see the signal, you’ve already invested heavily in the wrong direction.

This is also why P-heavy roles pay more. You’re not being paid for output volume. You’re being paid to absorb uncertainty and be accountable for decisions where the feedback loop is months, not minutes.


Ambiguity Is the Surface. Risk Is the Core.

People say P work is hard because it’s “ambiguous.” That’s half the story.

P decisions are hard because they combine:

  • Incomplete information — you never have enough data to be certain
  • Conflicting constraints — stakeholders want contradictory things
  • Delayed feedback — you won’t know if you were right for months
  • Partial reversibility — you can pivot, but not without sunk cost

You’re not picking “right vs wrong.” You’re picking a risk profile you’re willing to own — with your name on it.

And this is exactly why P roles are harder to automate. Not because AI can’t handle ambiguity (it can brainstorm options all day). But because someone has to absorb the consequences and still make a call.

AI can say “here are 5 options with trade-offs.” AI cannot say “I’ll stake my reputation on option 3, and here’s why I’m comfortable with the downside.”

That’s a human function. Maybe permanently.


AI Expands Options. Humans Own Consequences.

Can AI make good P decisions? Yes — in bounded contexts where the cost of error is low.

AI is already excellent at:

  • generating option sets (10 go-to-market strategies in 30 seconds)
  • surfacing trade-offs (pros/cons, risk matrices)
  • scenario expansion (what if X happens? what about Y?)
  • first-pass prioritization (rank by effort, impact, confidence)
  • stress-testing assumptions (poke holes in your logic)

But AI doesn’t have skin in the game. It’s weaker when decisions require:

  • Value hierarchy resolution — which objective wins when they conflict?
  • Organizational context — the politics, relationships, and unwritten rules that aren’t in any document
  • Accountability for irreversible downside — who gets fired if this fails?
  • Commitment timing — when to decide vs. when to wait for more signal

This gives us the operating model for the AI era:

AI widens the option space. Humans collapse it into commitments.

The best teams aren’t replacing human judgment with AI. They’re using AI to make human judgment better informed — more options considered, more trade-offs surfaced, more blind spots caught — before a human makes the call.


Why Coding Is AI’s Best Playground

Coding success is not a mystery. It’s an environment design win.

AI thrives in coding because coding has strong D rails:

  1. Deterministic syntax Code parses or it doesn’t.

  2. Deterministic evaluation Tests, linters, and CI provide immediate pass/fail signals.

  3. Explicit specs Requirements, contracts, and acceptance criteria are often clear.

  4. Fast feedback loops You can iterate in minutes.

  5. Error localization Stack traces and logs point to likely failure zones quickly.

Now compare that to strategy, product direction, or management:

  • ground truth is delayed
  • success criteria are contested
  • social context matters more than syntax
  • “correctness” is probabilistic until reality responds

So yes — AI is great at coding partly because coding has built an ecosystem where quality is legible. Errors are cheap, feedback is fast, and correctness has a definition.

Now notice what AI doesn’t do well in coding: choosing the architecture. Deciding between microservices and monolith. Picking the right abstraction. Those are P decisions within a D-dominant field — and they’re still where senior engineers earn their premium.

Lesson for every domain: if you want better AI outcomes, build stronger D rails around your P decisions. Make success measurable. Make feedback loops shorter. Make error detection automatic. The more D scaffolding you build around P work, the more AI can help with the P.


The Training Data Problem: We Overtrain D, Undertrain P

Most learning artifacts are D-heavy:

  • tutorials (“how to set up Kubernetes”)
  • docs (“API reference for Stripe”)
  • examples (“here’s a React component”)
  • SOPs (“follow these 12 steps to deploy”)
  • “best practices” (“always use connection pooling”)

Great for execution. Weak for judgment.

What we rarely document:

  • why option 3 beat option 1
  • what assumptions were accepted and which were challenged
  • what risks were consciously taken (and what the fallback was)
  • what would have changed the decision

Consider: there are thousands of tutorials on how to build a recommendation engine. There are almost none on how Spotify decided to build a recommendation engine instead of improving search, what trade-offs they weighed, and what they would have done differently.

The tutorials teach D. The decision narrative teaches P. We have an ocean of the first and a puddle of the second.

That missing layer is exactly what creates leverage.

If your org documents only execution, you train better execution assistants. If your org documents reasoning — the why, not just the what — you build a decision advantage that compounds.


The Real Force Multiplier: Decision Infrastructure

If AI gets very good at D, your moat is not output volume. It’s decision quality at speed.

Build this infrastructure:

1) Framing templates

Before solutioning, force clarity on: - objective - constraints - non-negotiables - failure modes - decision horizon

2) Risk-adjusted scoring

Don’t rank by upside alone. Add: - downside magnitude - reversibility - confidence level - time-to-signal

3) Decision journals

Capture: - what you chose - what you rejected - what assumptions you made - what would invalidate your choice

4) Pre-mortems + kill criteria

Before launch, define: - what failure looks like - what evidence triggers stop/pivot

5) AI/Human protocol

Decide explicitly: - where AI can auto-act - where human sign-off is mandatory - where escalation happens on low confidence

This is how you move fast and avoid high-cost blind commitments.


Career Strategy: Move Up the D/P Stack (Without Losing D Fluency)

Overreaction to avoid: “D is dead.”

No. D fluency still matters — for feasibility judgment, quality intuition, and credibility with your team. The PM who can’t read code, the strategist who can’t build a model, the manager who can’t review a design — they lose the ability to evaluate their own P decisions.

But pure D production is being commoditized. The gap is widening between people who produce output and people who decide what output matters.

Career upside comes from becoming excellent at:

  • Framing — defining the problem before anyone solves it
  • Prioritization — saying no to good ideas because you’ve committed to a better one
  • Risk modeling — knowing which bets to take and which to walk away from
  • Trade-off communication — explaining what you’re sacrificing and why it’s worth it
  • Commitment under uncertainty — making the call when 60% of the data says go and 40% says wait

Don’t abandon execution literacy. Just stop treating execution volume as your moat.

Your moat is the quality of your decisions — and your willingness to be accountable for them.


Closing: The New Premium Is Accountable Judgment

As AI gets better at deterministic execution, one thing becomes clearer:

The highest-leverage person is the one who can make high-cost decisions under uncertainty — and explain why.

That’s P.

And the winners won’t be:

  • people who ignore AI,
  • or people who outsource judgment to AI.

The winners will be people who can do both:

  1. use AI to widen and sharpen options
  2. own the final commitment when cost-of-error is real

You don’t beat AI by racing it on D. You build leverage by owning P.


Part 3 of the AI-Native PM series. Part 1: The D/P Framework. Part 2: Life in D/P. Building in public at fullstackpm.tech. Follow along on X @fullstackpmtech.

LinkedIn X

Enjoyed this?

Get posts like this in your inbox. No spam, unsubscribe anytime.

Comments

No comments yet. Be the first to share your thoughts!