Blog

  • Every knowledge worker is a manager now

    Every knowledge worker is a manager now

    Every knowledge worker is a manager now. Agentic AI has turned individual contributors into managers of AI agents, and first-line managers into leaders of managers of agents. The job descriptions have not caught up yet. The operating models have not caught up yet. The reskilling plans have not caught up yet. All of that is lagging the capability frontier by twelve to eighteen months — and the organizations that close that gap first will operate at a structurally different throughput than the ones still writing job descriptions for the jobs that existed in 2023.

    The shift: agentic AI crosses the line from tool to colleague

    For the first year and a half after ChatGPT, the thing called “AI” in most organizations was a better search box. A more patient editor. A faster rough-draft generator. Useful, but still a single-interaction tool. You asked, it answered, you moved on. The job of the knowledge worker did not fundamentally change — they just had a slightly sharper pencil.

    What changed in the eighteen months leading into 2026 is the arrival of agentic models. The word “agent” in that context is not marketing. An agent is a system that can do a sequence of things, hold state across those steps, make decisions about what to do next, use tools, and come back with a completed multi-step task. That is a categorically different interaction than “ask question, get answer.” It is closer to “give a junior colleague an outcome to produce and trust them to produce it.” The commercial consequence of that shift is the subject of this post.

    The role change: ICs become managers of agents

    The individual contributor job has silently changed. Writing short summaries of long content — once a junior-to-mid task — is now an agent task. The human role is to specify the outcome, check the output, and decide what to do with it. Meeting preparation — the pre-meeting brief of background, context, attendees, prior touchpoints — is now an agent task. The human role is to feed the context, review the brief, and adjust the framing. Drafting a first pass of almost any structured document — a proposal, a plan, an analysis — is now an agent task. The human role is the editor, not the author of the first draft.

    The common thread is that the IC’s job has shifted from doing to specifying outcomes and judging output. Those are management skills. Not in the metaphorical sense — in the literal sense. Framing a task clearly enough that someone (or something) else can execute it. Evaluating whether the execution meets the specification. Deciding when to iterate and when to ship. These are exactly the skills that used to distinguish a first-line manager from a senior IC, and they have become baseline requirements for an IC working with agents.

    The org change: first-line managers become leaders of managers of agents

    If every IC is now a manager of agents, then every first-line manager is now a leader of managers of agents. Their job is no longer to supervise execution — the agent is doing the execution. Their job is to coach the humans on their team in how to specify outcomes, how to judge output, how to know when an agent is producing garbage, and how to scale their orchestration over time. That is a completely different job than the first-line management job of three years ago, and it requires a different skill set.

    Two structural consequences follow. First, the middle management layer compresses because a first-line manager leading managers-of-agents can reach further than one managing direct executors — the coordination overhead per report drops when the reports are themselves operating on a multiplier. Second, the definition of “span of control” stretches, but not infinitely: the Dunbar layers still govern the number of humans a manager can hold relationships with, even if each of those humans is now operating agents underneath them. The org chart can get flatter. It cannot get unbounded.

    The strategic consequence: orchestration is now a baseline skill, not an advanced one

    The skill that used to distinguish senior managers from junior ones — the ability to frame work so someone else can execute it and judge whether their execution is good — is now a baseline IC capability. Orchestration is the new baseline. Writing is the new baseline. Judgment about output quality is the new baseline. The organizations that will operate at structurally higher throughput over the next five years are the ones that reskill their IC population around these baseline orchestration skills, rather than hiring more specialists who each do one thing well.

    Talent leverage, not headcount, becomes the scoreboard. A commercial organization that operates at 300 humans with strong orchestration capability can outproduce a commercial organization that operates at 600 humans with legacy IC job descriptions. The difference is not about working harder. It is about operating model. The 300-human organization has fewer Dunbar breakpoints, shorter decision loops, less cross-functional friction, and a higher per-seat agent-multiplier. All of that is the consequence of a single structural decision made at the job-description layer.

    So what boards should do

    Three actions sit on the CEO agenda over the next two quarters. First, rewrite the IC job descriptions for every knowledge-worker role in the organization so that orchestration and output judgment are explicit baseline capabilities, not bonus ones. Second, rewrite the first-line management job description so that coaching for orchestration is the core of the role, not supervision of execution. Third, audit the reskilling plan against the assumption that every knowledge worker in the organization is now a manager and needs to be trained as one — because the capability frontier has already shipped and the only question is whether the organization catches up in quarters or in years.

    Boards that do not require a reskilling plan at this scope are budgeting against an operating model that does not exist anymore. The plan does not need to be perfect. It needs to exist. The gap between organizations that have this plan and organizations that do not is the structural competitive advantage of the next five years, and it is already being measured — in throughput, in decision velocity, in the quiet retention of the top performers who can see the gap coming.

  • Model deprecation is the new continuity risk

    Model deprecation is the new continuity risk

    OpenAI announced the discontinuation of the Sora web and app experiences on April 26, with the Sora API following on September 24. The first deprecation triggers in two weeks. Enterprises that built workflows on Sora since launch are not facing a model upgrade — they are facing a workflow rebuild on a four-month timeline. This is the first prominent enterprise-facing AI deprecation event of the cycle, and the precedent it sets matters more than the specific product involved.

    Model deprecation is no longer a developer-tier concern. It is an enterprise governance question that deserves a place on the risk committee agenda. The real shift is happening here: AI dependency without continuity is becoming a board-level risk in 2026.

    The shift: dependency without continuity guarantees

    The pattern of the past two years has been to build agent workflows on whichever foundation model was demonstrably best at the time, with little contractual commitment from the model provider about how long that model would remain available. Provider terms have improved — Azure OpenAI’s twelve-plus-six-month commitment for generally available models is the strongest standard in market — but most enterprises have not negotiated equivalent terms with their chosen providers. They built on capability, not on continuity.

    When the provider sunsets the model, the enterprise’s options are bad. Migrate to a successor model that may behave differently in subtle ways — requiring re-validation of every governed use case. Renegotiate at the eleventh hour for extended access at unfavorable terms. Or absorb the operational disruption of the workflow simply not working until rebuilt.

    The Sora event is small in dollar terms but large in precedent. The next deprecation will involve a more enterprise-critical model, and the enterprises that did not see this one coming are not going to see that one coming either.

    The role change is the addition of an AI continuity discipline

    Inside enterprises that take this seriously, a discipline is emerging that did not exist in 2024 — AI continuity management. The work overlaps with vendor management, with disaster recovery, with model risk management, and with regulatory compliance, but it is structurally distinct from all of them. The discipline involves maintaining an inventory of model dependencies by workflow, negotiating continuity commitments at procurement, running successor-model regression tests on a regular cadence, and ensuring that the documentation chain meets the rebuild-readiness standard.

    Most enterprises have not staffed this discipline. The accountabilities are scattered across teams that do not coordinate. The procurement team negotiated the model contract a year ago without a continuity clause. The deployment team is building production dependencies on the model without thinking about migration cost. The risk team has not flagged model deprecation as a category. When the deprecation announcement lands, the company finds out it has no plan.

    The fix is straightforward in concept and slow in practice. Add continuity commitments to the procurement template. Build a model-dependency inventory. Designate an owner for AI continuity at the executive level. Run quarterly successor-model tests. None of this is hard. It is just unglamorous work that does not get done unless someone owns it.

    The strategic consequence is renewed buy-versus-build math

    Continuity risk changes the calculus of where to deploy AI capability. For workflows where the cost of unplanned migration is high — regulated workflows, mission-critical operations, customer-facing experiences with high switching costs — the case for either fine-tuning a frontier model into a controlled deployment, partnering with a vendor offering enterprise-grade continuity commitments, or building on open-weight models the enterprise can host indefinitely is stronger than it was in 2024. The case for relying on whichever model is best on a benchmark this quarter is weaker.

    The math is not simple. Open-weight models lag the frontier, sometimes meaningfully. Self-hosting carries operational cost that the proprietary providers absorb. The vendor lock-in to a single proprietary provider, even with the best continuity terms, is a different kind of risk than open-weight self-hosting carries. Each enterprise has to make this trade-off based on the workflow’s tolerance for capability lag versus its tolerance for continuity disruption.

    What is no longer defensible in 2026 is treating model continuity as someone else’s problem. The Sora sunset is small. The next one will not be.

    So what boards should do this quarter

    Add model deprecation to the risk committee agenda. The first deprecation event lands in two weeks. The board should at minimum understand which workflows are exposed and what the migration plans are.

    Demand a model-dependency inventory. Which workflows depend on which models from which providers, with which contractual continuity commitments. If this inventory does not exist, building it is the priority.

    Reconsider the buy-versus-build posture for mission-critical AI workflows. The 2024 default — use whichever proprietary model is best — was rational at the time. In 2026, with the deprecation precedent now visible, that default deserves an explicit reconsideration. Continuity is becoming a form of resilience. The boards that price it in this quarter will not be the ones rebuilding workflows under deadline.

    References and links

  • Team sizes are not design choices. They’re cognitive limits.

    Team sizes are not design choices. They’re cognitive limits.

    Team sizes are not design choices. They are cognitive limits. The recurring numbers that show up in military units, religious communities, hunter-gatherer bands, and commercial organizations are not management philosophy. They are a property of the animal doing the work, and any organizational structure that pretends otherwise pays a measurable tax in friction, communication overhead, quiet attrition, and decisions that arrive three weeks late.

    Two. Four to six. Eight to twelve. Twenty to twenty-five. Fifty. One hundred and fifty. The specific numbers recur across centuries and industries. In the Roman legion and the US Marines. In religious communities and hunter-gatherer bands. In tech companies, sales organizations, and the advice experienced managers give each other about when to split a growing team. It is not a coincidence. It is cognitive architecture. The constraint is no longer technology. The constraint has always been the brain doing the coordinating.

    Dunbar’s layers

    The research most commercial leaders eventually bump into is Robin Dunbar’s. Dunbar is a British anthropologist who, in the early 1990s, proposed that the size of a primate’s social group is constrained by the size of its neocortex. Extrapolating from primate data, he estimated the human number at around 150 — the number of people with whom any one of us can maintain a stable, recognisable, mutually-active relationship. He published it in the Journal of Human Evolution in 1992, and the number has been running through management literature ever since.

    The part that gets talked about less, but matters more, is that Dunbar’s 150 is not a single flat layer. It is the outer ring of a nested set, each layer roughly three times larger than the one inside it:

    • ~5 — your closest support group. The people you would call in a real emergency.
    • ~15 — your sympathy group. People whose loss would significantly affect you.
    • ~50 — your band or clan. People you know well enough to share deep context with.
    • ~150 — your active community. Stable, recognisable, mutually reciprocal relationships.
    • ~500 — acquaintances.
    • ~1500 — faces you can still recognise.

    These layers show up in the research almost regardless of whether the subject is a tribal society, an office workforce, or a social-network friend graph. And they map astonishingly well onto the team sizes that commercial organizations stumble toward by trial and error — not because anyone read Dunbar, but because the alternatives don’t work.

    The military got there first

    Armies have been experimenting with how to organize humans under extreme stress for two thousand years, and they arrived at exactly these numbers through pure selection pressure. Smaller was too fragile. Larger fell apart under fire. The numbers that survived are the numbers that work.

    A Roman legion’s smallest unit was the contubernium — eight soldiers who shared a tent, a mule, a mess, and most of their waking life. Eight. Right at the boundary between the 5-person inner layer and the 15-person sympathy group. The Romans knew nothing about neocortex ratios. They noticed that a group of eight held together in a way that a group of four or a group of sixteen did not.

    The modern US Marine Corps fireteam is four. The squad is roughly 13. The platoon is 30 to 40. The company is 100 to 150. The same ratios, twenty-one centuries later. The cognitive limits haven’t moved, because the brain they are about hasn’t.

    The tech industry rediscovered the same numbers

    The technology industry discovered the same structure and gave it different names.

    Jeff Bezos’s two-pizza rule — a team should be small enough to be fed by two pizzas — is a practical restatement of the 5-to-8 cognitive sub-layer. Amazon did not get there via anthropology. They got there by watching their own product teams stall every time they grew past the point where the whole group could fit around one table.

    Scrum teams are officially 7 ± 2 — the current Scrum Guide recommends 3 to 9 members — which echoes George Miller’s 1956 paper on the working-memory limit of around seven chunks. Miller was not writing about teams. The cognitive limit he found on how many things we can juggle at once maps cleanly onto how many people we can coordinate without losing track of where everyone is.

    Fred Brooks, in his 1975 book The Mythical Man-Month, observed that adding people to a late software project makes it later, because every new person increases the number of pairwise communication channels by roughly n(n–1)/2. Seven people means 21 channels. Ten means 45. Fifteen means 105. The coordination tax is quadratic, and it surfaces as “mysterious” slowdowns at exactly the team sizes where the math stops being manageable.

    W. L. Gore & Associates, the Gore-Tex company, built Dunbar’s number directly into its real-estate strategy. Founder Bill Gore had a rule: every time a building exceeded 150 employees, they built another building. He was running Dunbar’s ceiling inside his facility planning decades before Dunbar had published the paper.

    The Ringelmann effect, documented in 1913 and one of the oldest findings in social psychology, is the same story in a different register: as group size grows, the effort each individual contributes goes down. People pull harder on a rope when there are fewer of them holding it. Max Ringelmann measured it with actual rope-pulling experiments, and the finding has been replicated many times since in workplace and sports settings.

    The role change: the first-line manager span is a cognitive limit, not a cost line

    A first-line manager’s direct-report span is not a matter of preference for most cognitive work. It sits around 5 to 7. Push it to 10 and managers stop coaching and start triaging. Push it to 15 and the role has reverted to being an individual contributor with a different title. Organizations that scale cleanly keep that first layer tight even when the spreadsheet says it is expensive — because the spreadsheet is not pricing the coordination tax that a wider span produces downstream.

    The org change: 50 and 150 are hard boundaries

    The sub-team that actually owns a piece of work should be closer to 5 than to 10. Not because small teams are faster in principle, but because the communication-overhead curve gets steep fast after 7. Bezos was right about this, and almost every high-performing team of any reasonable size runs its real work through an informal group of four or five — regardless of what the reporting structure says on the org chart.

    When a function crosses 50 people, it needs an operational substructure. Tribes, chapters, pods, whatever the label — or the Dunbar sympathy layer breaks. When the people in a team stop knowing each other well enough that a death in someone’s family would visibly register with everyone, culture starts dying quietly. By the time anyone notices, six months have usually been lost.

    When an organization crosses 150, it runs two cultures whether the leadership admits it or not. The question is only whether the split is designed deliberately or happens by default. Organizations that handle the ceiling well accept it and build deliberate boundaries. Organizations that handle it poorly spend years pretending 400 people are “all one team.”

    The strategic consequence: org design is surrender, not construction

    Good organizational design is mostly a process of surrender. The cognitive architecture of the humans running the teams picks team sizes for you, and the only real choice is whether to build the org chart around what actually works or to fight it and pay the tax. Every commercial organization that has tried to force a bigger number — a 12-person manager span, a 30-person “small team,” a 300-person “family culture” — has either quietly subdivided itself into groups that look suspiciously like the Dunbar numbers, or lost the thing that made it work.

    AI augmentation does not move the cognitive ceiling. It moves the throughput below the ceiling. An IC managing four AI agents is still operating inside a span of four. A manager coordinating seven sub-teams of augmented ICs is still operating inside a Dunbar-5 layer. The numbers that governed organizational design before agents are the numbers that will govern it after.

    So what boards should do

    Boards should design operating models around the Dunbar layers and treat AI-augmented throughput as a multiplier on what each cognitive unit can do — not as a license to stretch the unit past its ceiling. The specific actions sit at four layers: first-line spans at 5 to 7 even under headcount pressure; sub-team ownership at 5; operational substructure at 50; deliberate cultural boundaries at 150. These are not target numbers. They are discovered numbers. Every other structure is an argument with biology, and biology does not negotiate.

    The Roman legions did not know about neocortex ratios. The Marines do not design their fireteams around anthropology papers. Jeff Bezos did not cite Dunbar when he ordered the pizzas. All three converged on the same numbers because the numbers are a property of the animal doing the work, not the work itself. The job of an organizational designer is to notice this — and then get out of the way.

    References

    • Dunbar, R. I. M. (1992). “Neocortex size as a constraint on group size in primates.” Journal of Human Evolution.
    • Dunbar, R. I. M. (2010). How Many Friends Does One Person Need? Harvard University Press.
    • Miller, G. A. (1956). “The Magical Number Seven, Plus or Minus Two.” Psychological Review.
    • Brooks, F. P. (1975). The Mythical Man-Month: Essays on Software Engineering.
    • Hackman, J. R. (2002). Leading Teams: Setting the Stage for Great Performances. Harvard Business School Press.
    • Ringelmann, M. (1913). Early social-loafing experiments, Annales de l’Institut National Agronomique.
    • Gladwell, M. (2000). The Tipping Point. Popularised Gore’s rule of 150 for the management audience.
    • The Scrum Guide — current recommended team size: 3 to 9 members.
  • Vertical AI is winning the deployment race

    Vertical AI is winning the deployment race

    Gartner’s April read says eighty percent of enterprises will have adopted at least one vertical AI agent by year-end, and thirty percent of all enterprise AI deployments will be vertical-specific. Bessemer’s vertical AI report from this month is even more direct: vertical AI companies founded after 2019 are reaching eighty percent of traditional SaaS contract values while growing four hundred percent year-over-year. This is not a minor adjustment to the deployment landscape. It is a structural redirection of where the value of agentic AI accrues.

    For boards in 2026, the implication is that the right framework for thinking about AI vendor strategy is no longer horizontal-versus-vertical. It is which verticals you bet on, and how early. Deployment speed defines advantage in this cycle, and the deployment race is now a vertical-by-vertical race.

    The shift: vertical specialization beats horizontal generality at the workflow layer

    Horizontal AI tools — the chat assistants, the general-purpose copilots, the broad productivity overlays — are still the largest category by usage. They are not the largest category by enterprise value. The reason is structural. A horizontal copilot is good at fifty things. A vertical agent is excellent at five things that are deeply embedded in a specific workflow.

    When the enterprise needs to extract value, depth wins over breadth. Abridge in clinical documentation. Harvey and EvenUp in legal. Hebbia in financial research. Specialized clinical-coding agents at major payers. The vertical players ship integrations into existing systems, understand the regulatory and accuracy constraints of the domain, and deliver outcomes that horizontal tools cannot match without significant configuration effort that customers refuse to undertake.

    The defensibility of vertical players is also higher than the market priced in 2024. The data flywheel inside a regulated vertical is genuinely hard to replicate. The customer relationships are stickier because switching costs include re-credentialing within the regulator’s expectations, not just re-implementing software.

    The role change is the chief AI buyer becomes a portfolio manager

    Inside enterprises, the executive responsible for AI vendor strategy is increasingly running a portfolio of vertical specialists alongside the foundation-model contracts. The horizontal tools form a substrate. The vertical agents form the high-value layer. The portfolio manager has to balance ROI realization against integration overhead, and has to decide which verticals to deepen versus which to defer.

    The skill set for this role is closer to portfolio investment management than to traditional procurement or IT leadership. The portfolio manager has to read product roadmaps, anticipate vendor consolidation, manage concentration risk, and time entry into emerging verticals where category leaders have not yet emerged. None of this is in the standard procurement or CIO playbook.

    Most large enterprises have not formally structured this role yet. The work is happening inside the CIO function or inside individual line-of-business AI initiatives, with no portfolio-level coordination. The result is double-procurement of overlapping vertical capability and missed early-mover advantage in verticals where the category leader will not stay reasonably priced for long.

    The strategic consequence reshapes acquisition strategy

    For enterprises in regulated industries — banks, insurers, hospital systems, large law firms, accounting firms — the vertical-AI thesis has a direct M&A implication. The category leaders in each vertical are trading at premium multiples now and will trade at higher multiples by 2027 once their data flywheels and customer concentrations are visible in audited financials. The window for acquisition at reasonable multiples is open in 2026 for most verticals. It will close.

    For incumbents who do not acquire, the implication is partnership at scale. The vertical specialists need distribution that incumbents already have. The incumbents need capability that the specialists already have. The deal terms will tilt toward the specialists as their growth rates remain visible. Incumbents that delay partnership decisions to 2027 will pay more for less favorable terms.

    For boards governing AI strategy, the directive question is whether the company is buying or building or partnering for vertical AI capability — and whether that decision is being made deliberately for each vertical, or by default by the absence of a decision. Default-by-absence is the mode most large enterprises are operating in. It is the most expensive mode.

    So what boards should do this quarter

    Map the AI vendor portfolio with horizontal versus vertical breakdown. If the breakdown is more than two-thirds horizontal, the company is missing the value-creating layer. If it is unmapped, that is a more urgent finding.

    Designate an executive owner for vertical AI portfolio strategy with explicit authority across line-of-business silos. The decisions are too consequential to be made silo by silo. The horizontal-tool decisions can stay with the CIO. The vertical-agent decisions need a portfolio view.

    For each major vertical relevant to the business, assign a clear posture: acquire, partner, build, or wait. Defaulting to wait by not deciding is the same as deciding to wait — and in most verticals it is the wrong decision in 2026. Execution speed will separate leaders from followers in this cycle.

  • What 47 unicorns in one quarter actually means

    What 47 unicorns in one quarter actually means

    What was announced

    In Q1 2026, 47 startups crossed the billion-dollar valuation threshold for the first time — the largest single-quarter cohort in over three years. The pace is concentrated at the seed and early-stage end. Global venture funding hit roughly $300 billion in the quarter, of which 80% — about $242 billion — flowed to AI companies. Four companies (OpenAI, Anthropic, xAI, Waymo) absorbed 65% of all capital deployed.

    What it means

    Two things become visible at the same time. First, the market is willing to underwrite billion-dollar valuations earlier in the company lifecycle than at any point since the late-2020 boom. The valuation framework is no longer derived from realized revenue. It is derived from deployed compute and team density. Second, capital concentration at the top has reached a level where four companies define the cost of capital for everyone else. A new AI startup raising in 2026 is competing for the same dollars that just priced OpenAI at $122 billion.

    The early-stage explosion and the late-stage concentration are two symptoms of the same conviction: capital has decided that AI is a winner-take-most market, and it is funding accordingly.

    Andreas’s Take

    My read on this: the unicorn count is a lagging indicator of a much earlier decision. That decision was made — quietly, by capital allocators — when the consensus shifted to a single conviction: AI capability gaps will widen, not narrow, over the next decade. From that conviction two strategies follow logically: fund the few names that might dominate the frontier (concentration), and over-fund the early stage so that whatever the next breakthrough looks like, you own a piece of it (proliferation). The 47 new unicorns are the proliferation half.

    I don’t think this is a bubble in the conventional sense. A bubble is a price disconnect from fundamentals. What we’re seeing is a price connection to a forecast about fundamentals. If the forecast is right — capability gaps widen, AI returns accrue disproportionately to a few players — today’s valuations are conservative. If it’s wrong, half of these unicorns will not survive their next priced round.

    What I’d say to boards and CFOs reading these numbers: don’t take comfort from “the market is hot.” Take instruction. Capital is signaling where it expects the next moat to form. The companies absorbing the capital are absorbing optionality, not just dollars.

    Recommendation

    Three things for leaders watching this market:

    1. Treat unicorn-count reports as competitive intelligence, not social proof. Look at which unicorns and what they are building — that is the signal of where the market expects gaps to open.
    2. Reassess your own compute and talent allocation against the new benchmark. If AI startups can attract billion-dollar valuations on team and compute alone, your incumbent organization is competing for the same talent at a different cost basis.
    3. Stress-test your strategic plan against a scenario where capability concentration plays out. What does your business look like if three or four frontier labs control the compute infrastructure and all serious AI deployment runs through them?

    References and related signals

  • The agentic year begins underprepared

    The agentic year begins underprepared

    The year opens with a measurable gap. McKinsey’s 2026 trust maturity survey, fielded in December and January, puts twenty-three percent of organizations into the scaling phase for agentic systems and thirty-nine percent into experimentation. The remaining majority — nearly two thirds — has not yet begun scaling AI across the enterprise. The capability frontier moved twelve to eighteen months faster than the operating models around it. That gap is no longer an experimentation question. It is the year’s defining strategic risk.

    The boards that close this gap first will not be using better models than their competitors. They will be running organizations that can metabolize what the models already do. The constraint is no longer technology. It is adoption — and adoption is a leadership problem.

    The shift is structural, not cyclical

    Agentic systems are not a new feature inside a familiar product. They are a new class of worker. They take a goal, decompose it into steps, hold state across those steps, call other tools, recover from errors, and return a completed unit of work. That changes what a job is, not how a job is done.

    The 2025 narrative — copilots, productivity boosts, ten percent uplift — is over. The 2026 question is harder. What units of work no longer require a human originator? What units of work now require a human reviewer instead of a human executor? Which decisions can be delegated to a system that explains its reasoning? The companies asking these questions on a Monday morning are reorganizing. The companies still benchmarking model accuracy are stalling.

    The shift is one-way. No board will vote in 2027 to remove agentic systems from a workflow they reduced from forty hours to four. The architectural choices made this year will compound.

    The role change has already happened on the ground

    Inside organizations that have actually shipped agentic systems, the role redefinition is happening informally, by individual contributors, ahead of any HR process. A senior analyst who used to write three reports a week now reviews twelve agent-drafted reports a week and signs off on the analysis. A staff engineer who used to write three pull requests a day now reviews fifteen agent-generated pull requests a day. An account manager who used to draft proposals now edits proposals the agent has built from CRM context.

    The work that survives is judgment, taste, accountability, and relationship. The work that does not survive is execution under specification. Job titles still describe the second category. Job content has already shifted to the first.

    First-line managers feel this most acutely. They were trained to manage humans doing execution work. They are now managing humans doing review work, who in turn are managing systems doing execution work. That is a different management discipline — closer to portfolio management of automated processes than to people management of execution teams.

    The organizational consequence is delayering

    Span of control widens when the work below each manager becomes more automated and more reviewable. McKinsey’s parallel work on the state of organizations points in the same direction: companies that scale agentic systems also flatten by removing one to two layers of middle management. The economic logic is direct. Middle layers existed to translate strategy into execution and to coordinate the humans doing that execution. When the execution is increasingly handled by systems and the translation is increasingly handled by models, the layer is doing less.

    This is not the 2024 layoff cycle that hit individual contributors. This is a 2026 reorganization that compresses the manager-of-managers layer. It is structurally different and politically harder. The people most threatened by it are the people running the budget meetings about it.

    Organizations that resist the delayering will have a temporary cost advantage and a permanent decision-velocity disadvantage. Decision cycles compress when fewer humans need to be in the loop. The competitor who removed two layers will commit to a market move three weeks faster. Over a year, that compounds into a different market position.

    So what boards should do this quarter

    Two actions belong on the Q1 agenda. First, demand a workforce plan that names the units of work moving from human execution to human review, with a twelve-month horizon. Vague AI strategies are no longer acceptable as deliverables; the question is which jobs, which tasks, which review cadences, which accountability lines.

    Second, name an executive owner for the operating-model redesign — not for AI strategy as a separate track, but for the way the company will be organized around the systems it has already deployed. The CHRO and the COO are the natural owners. The CTO is not. The technology decision is downstream of the operating-model decision, and treating it as upstream is how organizations end up with sophisticated tools and a 2023 org chart.

    The year that just started will be measured by the gap between capability and operating model. The companies that close it first set the pace for the rest of the decade. The risk is not moving too fast. The risk is moving too late. Execution speed will separate leaders from followers.