Hard Questions
Is AI making my team look productive without doing more useful work?
The social default
The default is the social one — Shane Parrish's vocabulary for "what other smart people are doing." When a Hacker News piece on appearing productive in the workplace hits 1,256 points and 492 comments, it tells you that the suspicion is widespread: the AI tooling that's shipping in 2026 produces more outputs without producing more outcomes. Dashboards rise. Customer revenue does not.
The default is mostly correct for AI deployed away from the constraint. Productivity theater is the predictable failure mode of AI rolled out everywhere. The honest answer is that the worry has two valid versions, and the diagnostic for whether your deployment is real comes from a forty-year-old operations novel.
The slower thinking
The two valid versions of the worry
Version 1: Dashboards lie
Productivity dashboards count outputs. Drafts shipped, tickets closed, briefs sent, summaries produced. AI agents are very good at producing outputs. The dashboard goes up. The business outcome — revenue protected, deals closed, decisions actually made — does not necessarily move.
Outputs and outcomes diverge in any system that measures the wrong thing. AI accelerates output production by an order of magnitude. Without changing the measurement, the divergence widens.
Version 2: Theater spreads
A team that ships more outputs without more outcomes signals to the rest of the org that "AI is working." Other teams ship their own AI tooling. Output volume across the org goes up. Outcome volume across the org doesn't. The Hacker News thread on appearing productive has the workplace version of this dynamic; AI just multiplies the volume.
What separates real lift from theater
The diagnostic comes from Eli Goldratt's The Goal (1984) and the field of operations research that grew from it. The argument is listed in detail in Goldratt was right about AI and The constraint is the coordination layer. The short version is this:
In any connected system, throughput is determined by the constraint — the slowest, most context-rich step in the chain. Improvements anywhere else accumulate as inventory, not as throughput. A 10% productivity boost on a marketer who is waiting on the CMO's brief produces zero additional briefs. The dashboard registers the marketer's improved velocity. The pipeline does not register more campaigns.
AI deployed away from the constraint produces dashboard inflation by definition. AI deployed at the constraint produces real lift — the constraint role unblocks five other people, and the lift compounds.
The Monday-morning diagnostic
A test you can run in the next twenty-four hours. It works as well in retrospect on past AI deployments as it does on current ones.
- At your team's next standup or Monday meeting, listen for the question being asked. Is it "what's blocking us?" or is it "look at all the things we shipped"?
- If it's the first, the team is operating around a constraint. The AI deployment that helps will be deployed at that constraint. The lift will be real.
- If it's the second, the team is in output mode. AI deployed in output mode produces theater. The dashboard will look great. The business will not move.
The Monday question is not vibes. It's a structural diagnostic. The "what counts as an outcome" Field Guide piece walks through how to translate the diagnostic into actual measurement.
The Block frame at scale
In February 2026, Block — Jack Dorsey's payments company — laid off about four thousand people and published "From Hierarchy to Intelligence," arguing that AI replaces the coordination work of middle management. Two readings of the bet are possible.
The unkind reading: theater at the org scale. The company shipped fewer outputs (post-layoff) and called it more outcomes.
The kind reading: structural. Block bet that the integration work middle managers used to do — context-sharing across teams in real time — is what AI does well, when the substrate is built (Goose, Block's open-source agent framework donated to the Linux Foundation; Anthropic's Model Context Protocol). If the bet pays out, the work that survives is the work that doesn't scale, and the work that goes is the coordination layer that does.
We are eighteen months from knowing which reading is right at Block's scale. The role-level version of the same bet — one constraint, one agent, real harness — is operating at smaller companies and the diagnostic for theater vs. lift is the Monday question.
The hardest version of this question
The hardest version is not "is my team in theater right now?" The hardest version is: do I want them to be? An organization that measures outputs has institutional reasons to keep measuring them. The CMO whose dashboard is up is the CMO whose budget is renewed. The PM whose tickets-closed metric is up is the PM who got promoted last cycle.
AI deployed away from the constraint serves those institutional incentives perfectly. It also drives the company into output-mode at exactly the moment when the actual leverage is at the integration point. The Brynjolfsson-Li-Raymond NBER paper on call-center AI is the closest thing to empirical evidence we have on this — the lift is real for low-tier workers, marginal for the top, and the question of whether the institution measures it correctly is open.
What the honest answer looks like
If you're asking this question seriously, the answer is some version of: yes, the risk is real; the diagnostic is the Monday question; the fix is to deploy at the constraint and measure outcomes, not outputs. The synthesis essay is the long version of why this is the right frame. KORA-01 in the renewal-risk Monday memo case is a worked example of what real lift looks like.
The vibe deployments — agents at non-constraint roles, with no audit log, posting work in private apps — produce theater reliably. The engineered ones, deployed at the constraint, with the four signals satisfied, produce lift. The pattern is consistent across the cases we've seen.
Sources
Appearing productive in the workplace, Hacker News thread, 2026
E. Goldratt, The Goal, North River Press, 1984
J. Dorsey, R. Botha, From Hierarchy to Intelligence, Block, 2026
E. Brynjolfsson, D. Li, L. Raymond, Generative AI at Work, NBER Working Paper 31161, 2023
Most AI agents don't survive production, Hacker News thread, 2025
What would have to be true for the opposite to be correct
- The agent is deployed somewhere other than the team's actual constraint role.
- The team's primary measurement is output volume rather than outcome impact.
- The agent's work isn't visible in Slack to the rest of the team — it ships in a private app drawer.
- Other roles aren't unblocked by the agent's output, and the constraint hasn't moved since deployment.
- The Monday-morning question on your team is "look at what we shipped" rather than "what's blocking us?"
Where to next
- → Read Goldratt was right about AI — the structural argument
- → Read The constraint is the coordination layer
- → Read What counts as an outcome — the measurement question
- → See KORA-01 at the constraint — the renewal-risk Monday memo
- → Continue to the Roster
- → Email leadership about the specific theater pattern you're worried about
- → I'm about to deploy AI everywhere — read the Hard Questions index for fit