<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Shift*Academy]]></title><description><![CDATA[Your digital leadership learning companion to help make sense of AI and emerging technologies, with practical guidance and techniques for implementation.]]></description><link>https://academy.shiftbase.info</link><image><url>https://substackcdn.com/image/fetch/$s_!dGVA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3669b15-b056-4bcd-828c-fa5d230cc563_256x256.png</url><title>Shift*Academy</title><link>https://academy.shiftbase.info</link></image><generator>Substack</generator><lastBuildDate>Sun, 26 Apr 2026 08:56:30 GMT</lastBuildDate><atom:link href="https://academy.shiftbase.info/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Shiftbase Ltd]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[academy@shiftbase.info]]></webMaster><itunes:owner><itunes:email><![CDATA[academy@shiftbase.info]]></itunes:email><itunes:name><![CDATA[Lee Bryant]]></itunes:name></itunes:owner><itunes:author><![CDATA[Lee Bryant]]></itunes:author><googleplay:owner><![CDATA[academy@shiftbase.info]]></googleplay:owner><googleplay:email><![CDATA[academy@shiftbase.info]]></googleplay:email><googleplay:author><![CDATA[Lee Bryant]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[How can Enterprises Create their own Agentic AI Evaluation Capabilities?]]></title><description><![CDATA[Why agentic systems are outpacing our ability to judge them and how we can evaluate their work effectively rather than just sampling outputs]]></description><link>https://academy.shiftbase.info/p/how-can-enterprises-create-their</link><guid isPermaLink="false">https://academy.shiftbase.info/p/how-can-enterprises-create-their</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 21 Apr 2026 14:37:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7rxv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Agents can now plan, act, escalate, and collaborate across tools and workflows. Early compositions are starting to deliver real outcomes, and in controlled settings, they can look surprisingly capable. Tasks are completed, metrics move and dashboards turn green.</p><p>But in many cases, we do not actually understand how these systems are behaving.</p><ul><li><p>We see the outputs they produce, but not the paths they take to get there.</p></li><li><p>We measure completion, but not reliability.</p></li><li><p>We observe success, but struggle to explain why it occurred, or whether it will happen again under slightly different conditions.</p></li></ul><p>This creates a dangerous illusion: systems that appear to be working long before we have any real confidence that they are.</p><p>Part of the issue is that most evaluation approaches have not caught up with the shift from tools to systems. We are still judging agents in the same way we judged early language models: by sampling outputs, spot-checking responses, or measuring task completion against predefined criteria. These methods can tell us whether something worked once. They tell us very little about how it behaves over time, under pressure, or at scale.</p><p>Worse still, agents can learn to satisfy these surface-level checks without actually developing the underlying capability we think we are testing. They pass the test, but for the wrong reasons. They follow patterns rather than understanding intent, they optimise for success conditions that do not hold outside the test environment and in practice, this means systems can degrade silently, masking fragility behind plausible performance.</p><p>As we begin to move toward Outcome-as-Agentic Solutions, this gap becomes harder to ignore. When a system is responsible for delivering a result, not just completing a task, the stakes change. It is no longer enough for something to work occasionally, or under supervision. It needs to behave consistently, adapt appropriately, and recover when things go wrong.</p><p>And yet, most organisations do not have a clear way to evaluate whether that is happening - we are deploying systems we do not yet know how to judge.</p><h2>Evaluating Agent Design</h2><p>If evaluation is no longer about checking outputs, then what replaces it?</p><p>Ideally, each organisation should design and run their own capability to observe, test, interpret, and improve agentic systems as they operate, as an ongoing function embedded into how systems are built and run. This is what an <strong>Agent Evaluation Capability</strong> provides.</p><p>At its core, it is the ability to make system behaviour visible and legible. To move from surface-level indicators of success to a deeper understanding of how outcomes are achieved, where fragility exists, and how performance evolves over time.</p><p>This is not a single tool or framework. It is a composed capability, built across systems, data, software, processes, and skills.</p><p><strong>Core Systems:</strong> Evaluation requires infrastructure. This includes the pipelines that capture system activity, the environments used to simulate scenarios, and the mechanisms for replaying or stress-testing behaviour. Without this layer, evaluation remains anecdotal, based on isolated observations rather than structured insight.</p><p>As systems become more complex, the ability to trace execution paths, reconstruct decisions, and observe interactions across components becomes essential.</p><p><strong>Data Sets:</strong> Agentic systems generate a different kind of data. Not just inputs and outputs, but traces: sequences of decisions, tool calls, escalations, and intermediate reasoning steps. Over time, these traces form a rich dataset that can be analysed to understand patterns of behaviour, identify failure modes, and detect drift.</p><p>Capturing and structuring this data is a prerequisite for meaningful evaluation.</p><p><strong>Software:</strong> On top of this foundation sit the tools that make evaluation possible. Evaluators, critics, and verifiers that assess system behaviour against defined expectations. Test harnesses that allow scenarios to be replayed and varied. Monitoring tools that surface anomalies and deviations in real time.</p><p>These components act as a counterbalance to generative systems, introducing layers of checking, validation, and interpretation.</p><p><strong>Services &amp; Processes:</strong> Evaluation is not just technical. It is operational. Teams need routines for reviewing performance, analysing failures, and refining system behaviour. This might include red teaming exercises, audit processes, incident reviews, and continuous testing loops.</p><p>Over time, these processes create institutional knowledge about how systems behave and how they can be improved.</p><p><strong>Skills:</strong> Finally, evaluation depends on people. Not just engineers, but individuals who can interpret system behaviour, understand risk, and make informed judgements about performance. This includes system thinkers, evaluators, and domain experts who can bridge the gap between technical signals and business outcomes.</p><p>As agentic systems become more embedded in organisations, these skills become critical.</p><p>Taken together, these elements form a capability that allows organisations to move beyond surface-level validation and toward a more robust understanding of their systems. Without it, agentic systems remain opaque.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hXS7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hXS7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hXS7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic" width="1000" height="1000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1000,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:139188,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/194914277?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hXS7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!hXS7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb9657dd-def3-4423-bbe2-d971c0b84a3a_1000x1000.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Key Design Principles for an Evaluation Capability</h2><p>Once evaluation is treated as a capability, a different set of design principles can emerge. These are not rules in the traditional sense, but patterns that help teams avoid the most common failure modes.</p><ul><li><p><strong>Evaluate behaviour, not just results.</strong> A correct outcome reached through a fragile or opaque path is not success. It is deferred failure. Systems need to be assessed on how they arrive at outcomes, not just whether they do.</p></li><li><p><strong>Design for failure visibility.</strong> Most systems do not fail cleanly. They degrade, drift, or produce plausible but incorrect results. Evaluation should make these conditions visible early, not hide them behind aggregate metrics or surface-level success rates.</p></li><li><p><strong>Separate generation from verification.</strong> The same system that produces outputs should not be solely responsible for judging them. Independent evaluators, critics, or verification layers create necessary tension in the system and reduce the risk of self-confirming behaviour.</p></li><li><p><strong>Make judgement legible.</strong> Decisions about system performance should be traceable and explainable, not buried in logs or reliant on individual interpretation. This is particularly important as systems scale and more stakeholders become involved.</p></li><li><p><strong>Treat evaluation as continuous.</strong> It is not a phase that happens before deployment. It is an ongoing function that evolves with the system, adapting to new conditions, new data, and new expectations.</p></li></ul><p>Evaluation cannot be just a checkpoint, it needs to be a continuous part of the system.</p><h2>Passing Tests Without Capability</h2><p>Agentic systems have the ability to pass tests without developing real capability. This is not new. In many domains, systems learn to optimise for the conditions under which they are evaluated, rather than the underlying objective. In agentic AI, this can manifest as agents that appear competent within test scenarios, but fail when those scenarios change in small but meaningful ways.</p><p>They follow patterns rather than understanding intent or exploit shortcuts that satisfy evaluation criteria without solving the actual problem. Often, they succeed in environments that resemble their training or testing conditions, but struggle to generalise beyond them.</p><p>The result is a form of surface competence.</p><p>From the outside, the system appears to work. Tasks are completed. Metrics improve. But beneath that surface, the system lacks the robustness required for real-world deployment. It cannot handle variation, ambiguity, or unexpected conditions without significant degradation.</p><p>This is particularly dangerous in enterprise settings, where early signs of success can drive rapid scaling. Systems that have not been properly evaluated for capability are extended into new contexts, where their limitations become more pronounced and more costly.</p><p>Without a structured approach to evaluation, these issues are often only discovered after failure.</p><h2>Deterministic Guardrails in Probabilistic Systems</h2><p>Agentic systems are inherently probabilistic. They operate under uncertainty, generating responses and actions based on patterns rather than fixed rules. This flexibility is part of their power, but it also introduces risk.</p><p>To manage this, organisations need to introduce deterministic elements into the system. Not to remove flexibility, but to bound it. To define where variation is acceptable and where it is not. These guardrails can take many forms.</p><ul><li><p>They may be constraints on what actions an agent is allowed to take.</p></li><li><p>They may be verification steps that check outputs against known rules or thresholds.</p></li><li><p>They may be escalation paths that transfer control to a human when confidence is low or risk is high.</p></li></ul><p>The key is that these elements are designed into the system, not added as an afterthought.</p><p>Too often, governance is treated as a layer of policy that sits outside the system. Documents are written, guidelines are issued, but the system itself remains unchanged. This creates a gap between intent and execution.</p><p>In effective agentic systems, governance is embedded. It is expressed through constraints, checks, and decision logic that shape behaviour in real time. Reliability is not something that emerges from probabilistic systems on its own. It is something that must be designed.</p><h2>What Evaluation Makes Visible</h2><p>The impact of an evaluation capability changes what teams can see, what they trust, and how they intervene. Systems are judged on whether they complete tasks. Once evaluation is introduced, the focus shifts and teams begin to understand how work is actually being done, where it breaks down, and what needs to change. Let&#8217;s explore some examples:</p><h3>Customer Onboarding</h3><p>A team defines success as &#8220;customer activated within 72 hours.&#8221; Without evaluation, performance is tracked at the end of the funnel. Activation rates go up or down, but the system remains largely opaque.</p><p>With evaluation in place, the system becomes legible. Teams can see where drop-offs occur, how agents respond, and which paths lead to successful activation. They can identify whether delays are caused by missing data or poor sequencing. They can see which users require escalation and why.</p><p>Over time, this shifts the work. Instead of reacting to missed targets, teams begin to refine the system itself, improving how onboarding is delivered rather than compensating for its failures.</p><h3>Incident Response</h3><p>In a typical setup, success is measured by resolution time. An incident is detected, addressed, and closed. If the system resolves the issue quickly, it is considered effective. Evaluation changes the frame.</p><p>Teams can see how the system detected the incident, which signals it prioritised, and how it coordinated response actions. They can identify missed signals, unnecessary escalations, or delays in decision-making. They can analyse how the system behaves under pressure, not just whether it resolves the issue.</p><p>This creates a different kind of confidence. Not just that incidents are resolved, but that the system can be relied upon when conditions change.</p><h3>Sales Qualification</h3><p>A sales team defines success as &#8220;90% of new pipeline entries fully qualified within five working days.&#8221; Without evaluation, progress is tracked through completion metrics. Entries are marked as qualified, and the process appears to work. With evaluation, the process becomes transparent.</p><p>Teams can see how qualification is being performed, whether agents are applying the right criteria, and where information is being missed or misinterpreted. They can detect patterns in stalled entries and identify where additional support is needed.</p><p>More importantly, they can distinguish between genuine qualification and superficial completion. This reduces the risk of pipeline inflation and improves the overall quality of sales activity.</p><p>Across these examples, evaluation allows teams to move from observing outcomes to interpreting behaviour. From reacting to results to shaping the systems that produce them.</p><p>Without this visibility, systems may appear effective while hiding fragility. With it, organisations gain the ability to refine, adapt, and trust the systems they are building.</p><h2>Getting Started</h2><p>Most organisations do not need to start from scratch.</p><p>They already have elements of this capability in place: logging systems, monitoring tools, performance reviews. The challenge is to connect these elements and extend them to cover agentic systems.</p><p>A practical starting point is to focus on a single outcome.</p><p>Choose a system that is already in use, or being developed, and begin by instrumenting it. Capture the traces of how it operates. Identify key signals that indicate success, failure, or drift.</p><p>Introduce simple evaluation loops. Review behaviour regularly. Look for patterns. Ask not just whether the system worked, but how. Expect gaps. There will be missing data, unclear signals, and behaviours that are difficult to interpret. This is part of the process. Each gap is an indication of where the capability needs to be strengthened.</p><p>Over time, these practices can be formalised. Evaluation becomes less ad hoc and more systematic. Insights accumulate. Confidence grows.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7rxv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7rxv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 424w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 848w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 1272w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7rxv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic" width="1000" height="2000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2000,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:207269,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/194914277?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7rxv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 424w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 848w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 1272w, https://substackcdn.com/image/fetch/$s_!7rxv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd71b1bf8-de95-4765-89d6-7160160ed2ab_1000x2000.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If evaluation is treated as something to be added later, it will always lag behind the system it is meant to support. Building it alongside the system is the only way to ensure it keeps pace.</p><p>Read on to explore the practicalities of evolving this emerging capability.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/how-can-enterprises-create-their">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Agents at the Ready? Yes and No...]]></title><description><![CDATA[Mythos, memex and management - how the various layers of agentic AI development are moving at different speeds and why we need to bring them together]]></description><link>https://academy.shiftbase.info/p/agents-at-the-ready-yes-and-no</link><guid isPermaLink="false">https://academy.shiftbase.info/p/agents-at-the-ready-yes-and-no</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 14 Apr 2026 14:30:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Yaws!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Agentic AI capabilities are developing within several pace layers at once - economic, infrastructure, capability readiness, and knowledge engineering - and it is getting harder to stay on top of these developments whilst tracking their interdependence. But organisational readiness is moving slower than each of them for most companies. Are agents ready for the big time yet, and if not, where should we focus our efforts?</p><h2>Anthropic in the news</h2><p>Anthropic has been making the news a lot recently, with <strong><a href="https://github.com/yasasbanukaofficial/claude-code">the leak of its Claude Code codebase</a></strong> (the harness code, not the model weights, etc) and more recently with the announcement that its new model preview Mythos is so powerful that it was held back from release until Anthropic built some counter-measures for its <strong><a href="https://www.wired.com/story/anthropics-mythos-will-force-a-cybersecurity-reckoning-just-not-the-one-you-think/">astonishing ability to detect security exploits in all kinds of commonly-used software</a></strong>:</p><blockquote><p><em>According to Anthropic, Mythos Preview crosses a threshold of capabilities to discover vulnerabilities in virtually any and every operating system, browser, or other software product and autonomously develop working exploits for hacking. With this in mind, the company is only releasing the new model to a few dozen organizations for now&#8212;including Microsoft, Apple, Google, and the Linux Foundation&#8212;as part of a <a href="https://www.wired.com/story/anthropic-mythos-preview-project-glasswing/">consortium dubbed Project Glasswing</a>.</em></p></blockquote><p>Both developments have created a lot of fear and noise, and Mythos points to the huge gap opening up between the (risky, arguably uncontrollable) art of the possible on one hand, and the architectures and control systems we have in place to govern and guide AI within complex organisations.</p><p>But it is a third, less dramatic development that I think is most immediately relevant to enterprise AI and the pursuit of new organisational operating systems: <strong><a href="https://www.anthropic.com/engineering/managed-agents">Claude Managed Agents</a></strong>, which they describe as <em>&#8220;a hosted service in the Claude Platform that runs long-horizon agents on your behalf through a small set of interfaces meant to outlast any particular implementation&#8212;including the ones we run today.&#8221;</em></p><p>It seems to be aimed at SaaS teams that want to embed Claude agents into their products, providing a managed infrastructure layer with composable APIs that sits between your system and Claude&#8217;s models. <strong><a href="https://kenhuangus.substack.com/p/how-anthropic-scaling-managed-agents?publication_id=1796302&amp;post_id=193754059&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Ken Huang provided a good overview of its three main components</a></strong> (harness, session log and memory) and some thoughts on its potential usage.</p><p><strong><a href="https://www.anthropic.com/engineering/managed-agents">Anthropic have called it a meta-harness</a></strong> that could help scale agentic AI and provide an abstraction layer that can cope with future changes and innovation:</p><blockquote><p><em>The challenge we faced is an old one: how to design a system for &#8220;programs as yet unthought of.&#8221; Operating systems have lasted decades by virtualizing the hardware into abstractions general enough for programs that didn&#8217;t exist yet. With Managed Agents, we aimed to design a system that accommodates future harnesses, sandboxes, or other components around Claude.</em></p></blockquote><p>This is one of the first serious attempts to define an agentic infrastructure layer, rather than just an agent.</p><h2>Is Agentic AI Showing up in Enterprise ROI?</h2><p>Adoption and investment analysis for enterprise AI usage is beginning to show the impact of agentic AI above and beyond individual productivity use cases for chatbots. <strong><a href="https://www.a16z.news/p/ai-adoption-by-the-numbers">A16z&#8217;s latest report</a></strong> (released last week) cites coding, search and support use cases as the most active in the firms they surveyed, with tech, legal and healthcare as the sectors most keen on AI adoption.</p><p>But we will not really begin to develop a full picture of enterprise AI&#8217;s impact until we make more progress building out the kind of architecture necessary for agentic automation to start operating as a managed layer of the tech stack.</p><p>McKinsey discussed this paradox - <strong><a href="https://www.mckinsey.com/capabilities/people-and-organizational-performance/our-insights/ai-is-everywhere-the-agentic-organization-isnt-yet">AI is everywhere, and yet the agentic organisation is not yet visible</a></strong> - in a recent podcast with Senior Partner Alexis Krivkovich:</p><blockquote><p><em>The real promise with agentic, relative to generative AI or previous evolutions of AI, is that you can have the equivalent of superhuman capabilities added to your teams. But the day-to-day workflows and the rituals around ways of working will need to fundamentally change.</em></p></blockquote><p>That&#8217;s what we mean when we say the operating model needs to shift. You need to think about how the hours of the day happen differently, the process of overseeing an agent population, how you engage in problem-solving as a team&#8212;and put the right governance and risk controls on top of that.*</p><blockquote></blockquote><p>AI is everywhere in the enterprise, in other words, but the work of <em>organising around it</em> has barely begun.</p><h2>Agents &amp; The Coasean Singularity</h2><p>But the real prize is not just greater efficiency in the old work model, but the so-called <em>Coasean Singularity</em> that occurs if/when agents drive workflow transaction costs towards zero, changing the entire premise of what organisations are and why they exist, which was the question Ronald Coase posed in his famous 1937 essay The Nature of the Firm.</p><p><strong><a href="https://www.nber.org/system/files/working_papers/w34468/w34468.pdf">NBER released an interesting working paper on this question in late 2025</a></strong>, which concluded that this shift could challenge existing market structures in good and bad ways, whilst also opening up entirely new forms of exchange beyond simple labour, jobs and contracting:</p><blockquote><p><em>The capacity of AI agents to dramatically reduce transaction costs as automated intermediaries could unlock new forms of market participation, enable previously infeasible mechanisms, and push allocative efficiency closer to competitive ideals. Yet the same forces that make agents attractive&#8212; their tireless persistence, computational superiority, and negligible marginal costs&#8212;also threaten to overwhelm existing market structures. The ultimate impact will depend critically on collective choices adopted regarding agent design, market structures, and regulatory frameworks.</em></p></blockquote><p><strong><a href="https://howardyu.substack.com/p/coase-vs-claude-and-the-future-of?publication_id=315949&amp;post_id=192932951&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Professor Howard Yu wrote an interesting piece on this recently</a></strong>, covering many angles that I think are useful starting points, such as Haier&#8217;s Rendanheyi model and Sangeet Paul Choudary&#8217;s work on unbundling and rebundling tasks and jobs. One of his key takeaways is that when transaction costs are lowered and supply can meet demand more efficiently, there is a tendency to create barbell-shaped markets where the premium and commodity ends grow, but the middle tier contracts.</p><p>Jack Dorsey&#8217;s rather chaotic financial services firm Block claims to be developing its own agentic operating system that will sever the link between headcount and output, but it is hard to know if their layoffs are the result of this innovation or just an unwinding of previous over-hiring.</p><p><strong><a href="https://thedigitalleader.substack.com/p/the-equation-just-broke-what-blocks">John Rossman wrote glowingly of Block&#8217;s intentions last week</a></strong>, and although I profess a degree of scepticism given Block&#8217;s history, there was one observation by Block&#8217;s Executive Officer in the piece that is worth highlighting:</p><blockquote><p><em>Before Block could restructure how work gets done, they had absolute clarity on what work needs to accomplish. This requires &#8220;Thinking in Outcomes&#8221;. Jennings described three non-negotiable outcomes that governed every restructuring decision &#8212; reliability (no outages), regulatory integrity (compliance teams untouched, full stop), and durable growth (roadmap commitments honored).</em></p></blockquote><p>Outcomes clarity and context are vitally important if we want reliable agentic architectures.</p><p>But we should not presume that agentic architectures will simply automate and wipe out jobs in complex organisations. As a layer of intelligence, automation and orchestration, agentic AI will make many new value creation methods possible, which could have an expansionist effect resulting in more activity and potentially also more or better human roles.</p><h2>Agentic Capabilities Still Maturing</h2><p>We have written a lot about emerging agentic capabilities and how companies hope they will develop to become a key new infrastructural layer in the organisational operating system stack, and the progress we are seeing is encouraging. But it is easy to conflate proofs of concept that demonstrate theoretical performance with mature, tested capabilities.</p><p><strong><a href="https://www.linkedin.com/pulse/agents-dont-know-what-good-looks-like-thats-exactly-luca-mezzalira-sgwte/?trackingId=XGn%2B%2FupMOGKSQTe618fEZw%3D%3D">Luca Mezzalira wrote a thoughtful post last month reflecting on an O&#8217;Reilly Network event that discussed agentic capabilities in some depth</a></strong>, which captured some nuanced critiques of current agentic AI state of the art. For example, as long as our agent testing is based only on behaviour and outputs, rather than its real capabilities, then we are at risk of missing deeper problems with how agents sometimes satisfy test conditions, which is not always honest in the sense people would understand the term. There is hidden risk here if we move too fast and assume too much.</p><p>Mezzalira&#8217;s take-aways from this are threefold.</p><p>First, he argues that introducing deterministic guardrails to probabilistic agents is non-negotiable.</p><p>Second, we don&#8217;t yet know what good looks like, so we should avoid the mis-steps that accompanied the micro-services movement, which had some parallels with agentic AI, but we should also be cognisant of the risk gap this opens up. If agents can already autonomously develop working exploits using a model like Mythos, not knowing what good looks like is a scary prospect.</p><p>Finally, we should recognise and accept that we are all beginners and there is a lot still to learn and discover about how agentic AI will play out in the real world, and especially how it will intersect with human behaviours and failure modes.</p><p>Recently, we have seen <strong><a href="https://arxiv.org/abs/2602.19260">several studies</a></strong> and <strong><a href="https://pathfounders.com/p/is-unlikelyai-the-most-undervalued?_bhlid=26da303bc3e9686449e01785c869e167e680e61b">startup announcements</a></strong> relating to neuro-symbolic AI systems and agents that might be part of the solution for deterministic agents in areas where we cannot tolerate probabilistic risk. Perhaps at a minimum, it is possible to use neuro-symbolic methods in testing or orchestrator agents to verify the output of other agents in the system.</p><h2>Memex Redux: Agents that Document your Knowledge</h2><p>There is one final development worth flagging, as I think it points to a bigger trend we will see in agentic enterprise AI in the future. <strong><a href="https://gist.github.com/karpathy/442a6bf555914893e9891c11519de94f">Andrej Karpathy shared a gist called LLM Wiki</a></strong> that helps us use AI agents to build, grow and curate our own structured personal knowledgebases, as Vannevar Bush first imagined in 1945, and Ted Nelson later attempted with project Xanadu:</p><blockquote><p><em>Most people&#8217;s experience with LLMs and documents looks like RAG: you upload a collection of files, the LLM retrieves relevant chunks at query time, and generates an answer. This works, but the LLM is rediscovering knowledge from scratch on every question&#8230;.</em></p><p>The idea here is different. Instead of just retrieving from raw documents at query time, the LLM <strong>incrementally builds and maintains a persistent wiki</strong> &#8212; a structured, interlinked collection of markdown files that sits between you and the raw sources.</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Yaws!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Yaws!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Yaws!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:224778,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/194175820?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Yaws!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!Yaws!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19c7e169-7ca5-4ac7-91f5-aae75c0b00e9_1408x768.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The idea of feeding our AI agents relevant information and knowledge, but asking them to maintain a structured knowledge graph or wiki is potentially helpful in overcoming the limits of memory and the context window when performing long-run or multi-stage tasks. I would like to imagine each agent we train and deploy in the enterprise could maintain its own knowledgebase to improve its learning and to stay updated on relevant developments that might impact its work.</p><p>Among the AI solopreneur and personal productivity communities, there is already a popular technique that involves combining Claude with the personal knowledge repository Obsidian to give personal agents better context to work with. <strong><a href="https://limitededitionjonathan.substack.com/p/stop-calling-it-memory-the-problem">But as one commentator pointed out, simply loading relevant wiki pages as .md files into an LLM&#8217;s context window is not really the same as a functioning memory</a></strong> that can be queried like a database, nor is it very scalable.</p><p>So whilst this is a great starting point for a simple memory and context system for people and agents, it would probably need to be paired with a real knowledge graph and/or database to allow agents to query, search and refactor the information they need without overwhelming the context window.</p><p>What would it mean for every enterprise agent to maintain its own structured knowledge base? What are the implications for how we build and govern knowledge infrastructure? Given how hard it has been over the years to persuade leaders that this kind of knowledge engineering should be a key part of their work, perhaps this development might point to us being able to accelerate world-building and context engineering using agents to do the hard work.</p><p>Within most organisations, there is a lot of talent and work going on across the operational, technical and (hopefully!) architectural layers that will support agentic AI. But it is worth repeating that only leaders will have the holistic view of how the economics, infrastructure, capability readiness, and knowledge engineering all work together to make reliable, productive agentic AI a reality.</p><p>Managed agents with well-curated knowledge repositories could lead us to the Coasean Singularity with all the economic benefits that entails. But if we don&#8217;t bring it all together in a considered and safe way, then Mythos points to what might go wrong.</p>]]></content:encoded></item><item><title><![CDATA[The Autonomous Project Manager]]></title><description><![CDATA[What happens when coordination becomes a system, not a role]]></description><link>https://academy.shiftbase.info/p/the-autonomous-project-manager</link><guid isPermaLink="false">https://academy.shiftbase.info/p/the-autonomous-project-manager</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 07 Apr 2026 15:03:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lgr5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="callout-block" data-callout="true"><p>Note: This piece uses project management as a lens to make a broader shift tangible. Much of our recent work has focused on how organisations move from implicit, human-carried coordination toward systems that are codified, legible, composable and coherent. These ideas are often discussed at an architectural level. Here, they are shown in the context where coordination pressure is most visible, allowing the underlying patterns to be observed in practice.</p></div><h2><strong>The Hidden World of Projects</strong></h2><p>Most projects do not fail because the work is unclear, but because alignment cannot be sustained as the work unfolds. Teams generally know what needs to be done, and plans are often internally consistent - the difficulty emerges in how that work is coordinated across time, teams and systems.</p><p>We often label coordination as a single activity, but in reality it is the ongoing effort required to keep multiple streams of work moving in a way that remains coherent. This includes managing dependencies between tasks, interpreting whether progress is meaningful or superficial, and maintaining a shared understanding of priorities as conditions change.</p><p>These signals exist in every project - task state transitions, message latency, document revisions and shifts in ownership all provide indicators of how work is actually progressing. In practice, these signals are fragmented across tools and interpreted manually. Teams reconstruct the state of the project through meetings, status updates and informal exchanges, often with a lag between what has happened and what is understood.</p><p>Most organisations have responded by increasing visibility. Dashboards, trackers and reporting layers provide more information about activity. This improves awareness, but it does not reduce the effort required to interpret that information. Progress is still reconciled manually, dependencies are still tracked unevenly, and alignment remains something that has to be actively maintained, producing a kind of project reporting theatre where the production of slide decks becomes more of a priority than the actual work of the project.</p><p>This is where project managers spend most of their time. They act as a coordination layer across fragmented systems and teams, stitching together signals, resolving inconsistencies and intervening when alignment begins to drift. Much of this work is invisible. It does not appear in plans or reports, but it determines whether the system holds together.</p><p>As complexity increases, this model begins to break down. More teams, more dependencies and faster cycles of change increase the volume of signals that need to be interpreted. Coordination effort scales with this complexity, and the organisation becomes increasingly dependent on individuals who can absorb ambiguity and maintain coherence under pressure.</p><p>Most coordination effort is invisible, and because it is invisible, it is rarely designed or improved.</p><h2><strong>The Autonomous Project Manager</strong></h2><p>Once coordination is recognised as the primary constraint, a different question becomes practical. Instead of asking how to improve planning or execution, attention shifts to how coordination itself is maintained and where it breaks down.</p><p>In most projects, this breakdown is visible in small, repeated patterns:</p><ul><li><p>Status updates arrive late or require interpretation.</p></li><li><p>Dependencies are identified, but their impact is not always tracked as conditions change.</p></li><li><p>Teams report progress locally, while the overall picture remains unclear.</p></li><li><p>Escalation often depends on someone noticing that something feels off, rather than on a consistent view of risk.</p></li></ul><p>These are not failures of discipline within the project team, they reflect the absence of a system that can continuously interpret what is happening across the project.</p><p>Recent advances in agentic AI make it possible to construct such a system. In practical terms, this creates a clear and immediate use case for a project management agent. The coordination work that currently sits with human project managers is already structured enough to be observed, interpreted and partially handled by a system operating across the same tools and signals. Models can now operate with persistent context, observe activity across multiple tools, and act over time rather than in isolated interactions. This allows coordination to be supported directly within the flow of work.</p><p>The idea of an autonomous project manager can be understood more concretely as a project management agent that operates across the coordination layer of the work. Rather than replacing the role, this agent takes responsibility for a set of recurring coordination tasks that are currently performed manually. It surfaces what coordination looks like when it is embedded into the system rather than carried by individuals. In practice, this defines a set of responsibilities that can be handled by a project management agent.</p><ul><li><p><strong>The agent continuously assembles a view of progress.</strong> Task state changes, document revisions and communication patterns are observed directly, allowing the current state of work to be assembled continuously. The need to reconstruct progress through meetings and reports is reduced.</p></li><li><p><strong>The agent tracks and updates dependencies as conditions change.</strong> When upstream work slips, the effect on downstream tasks can be traced and surfaced early. This reduces the lag between a change occurring and its implications being understood.</p></li><li><p><strong>The agent generates and distributes updates based on the current state of work.</strong> Updates reflect what is happening across the system, rather than relying on periodic summaries that quickly become outdated or inconsistent across audiences.</p></li><li><p><strong>The agent detects patterns that indicate risk and routes them appropriately.</strong> Patterns that indicate risk, such as repeated delays, conflicting signals or gaps in ownership, can be surfaced with context and directed to the appropriate individuals. This reduces reliance on informal escalation through messages and ad hoc discussions.</p></li></ul><p>These changes do not remove the need for coordination. They shift a significant portion of coordination work from manual effort to system support. Activities such as gathering updates, tracking dependencies and maintaining visibility can be handled continuously by the agent, reducing the need for project managers to reconstruct the state of the work.</p><p>For project managers, this changes the nature of the role. Time previously spent chasing updates, reconciling information and maintaining visibility can be redirected toward working with stakeholders, resolving trade-offs and shaping solutions as conditions evolve. Less time is spent reconstructing the state of the project, and more time is available to address ambiguity, resolve trade-offs and guide direction.</p><p>Coordination moves from something that is repeatedly reconstructed to something that is continuously maintained.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lgr5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lgr5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lgr5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:209483,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/193471985?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lgr5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!lgr5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a534357-5b8c-4bae-9408-64e10e3b0ec2_1408x768.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1><strong>Inside the Coordination System</strong></h1><p>The patterns described above are already present in most projects. Signals are generated continuously as work progresses, but they are fragmented, delayed and interpreted manually. A coordination capability brings these elements together into a continuous layer that can observe, interpret and respond as conditions change.</p><ul><li><p>The first component is <strong>sensing</strong>. Every project generates a stream of signals: task state transitions, changes in ownership, message latency between teams, document revisions and dependency updates. These signals exist across tools and communication channels, but they are rarely combined. As a result, no single view reflects the current state of the system. Bringing these signals together allows the project to be observed as it is, rather than as it was last reported.</p></li><li><p><strong>Interpretation</strong> builds on this. Signals on their own do not indicate whether a project is on track - a task moving to &#8220;in progress&#8221; may represent genuine progress, or it may reflect a delay in starting work. A completed task may unblock downstream activity, or it may mask unresolved issues. Interpretation requires relating signals to outcomes, timelines and dependencies. This allows the system to distinguish between movement and meaningful progress.</p></li><li><p><strong>Coordination</strong> follows from this understanding. Dependencies between workstreams can be treated as active relationships rather than static links in a plan. When upstream work slips, the system can trace the impact across downstream tasks, identifying where timelines need to adjust or where attention is required. This reduces the delay between a change occurring and its consequences being understood.</p></li><li><p><strong>Communication</strong> becomes part of this same layer. Instead of aggregating updates manually, the system can generate views of progress based on current conditions. Different stakeholders require different levels of detail, but the underlying information remains consistent. This reduces the need for parallel reporting structures that often diverge over time.</p></li><li><p><strong>Escalation</strong> provides structure to decision-making. Patterns such as repeated delays, conflicting signals across teams or gaps in ownership can be detected as they emerge. These situations can be surfaced with relevant context and directed to the appropriate individuals. This allows human attention to focus on points of uncertainty, rather than on monitoring routine activity.</p></li></ul><p>These components do not operate independently. They reinforce each other. Better sensing improves interpretation. Clearer interpretation supports more effective coordination. Structured coordination reduces the noise in communication. Consistent communication improves the quality of escalation.</p><p>Together, they form an emerging capability that allows alignment to be maintained continuously. The system does not remove variability from the project, but it reduces the effort required to detect, understand and respond to it.</p><p>Coordination becomes observable, interpretable and responsive, rather than fragmented and manually reconstructed.</p><h2><strong>Where Humans Remain Central</strong></h2><p>A coordination capability changes how effort is distributed across a project, but it does not reduce the need for human judgement. It changes where that judgement is applied and how it is informed.</p><p>When coordination is supported as a system, the backward looking reconstruction effort is reduced. Signals are assembled continuously, dependencies are made visible as they evolve, and the state of the project can be observed without manual aggregation. This creates space for attention to shift toward interpretation and decision-making.</p><p>Judgement becomes concentrated in areas where signals alone are insufficient. Prioritisation across competing objectives requires an understanding of intent that extends beyond task-level progress. Decisions about whether to absorb a delay, re-sequence work or escalate an issue depend on how outcomes are valued in context.</p><p>Ambiguity is another area where human input remains essential. Conflicting signals are common in complex projects. One team may report progress while another experiences blockage. A dependency may appear resolved in one system but remain open in another. Interpreting these situations requires an understanding of how work is actually being carried out, not just how it is represented.</p><p>Stakeholder dynamics introduce a further layer. Commitments, expectations and reputational considerations shape how decisions are made. These factors are rarely captured fully in systems, yet they influence how trade-offs are resolved and how issues are communicated.</p><p>Escalation, when structured effectively, directs these moments of uncertainty to the appropriate individuals. The system can identify patterns and assemble context, but decisions about how to respond remain with people who carry responsibility for outcomes.</p><p>This shifts the role of those involved in project leadership. More time and attention is freed up for shaping intent, resolving uncertainty and making decisions that affect the trajectory of the work.</p><h2><strong>The Failure Mode: More Activity, Same Misalignment</strong></h2><p>Introducing a coordination layer does not automatically improve performance and alignment. It can also increase the volume and speed of activity without changing how that activity is interpreted.</p><p>This pattern is already familiar. Most organisations have invested in tools that improve visibility, yet still rely on manual effort to reconcile what those signals mean. The result is more information, but not necessarily better alignment.</p><p>This leads to a form of coordination that is technically consistent but operationally ineffective. Activity is tracked, updates are generated, and alerts are surfaced, yet the underlying coherence of the work does not improve.</p><ul><li><p>Communication can become a source of friction rather than clarity.</p></li><li><p>Escalation can place additional strain on decision-makers if it is not well structured.</p></li></ul><p>There is also a tendency to formalise coordination prematurely. Systems that impose rigid workflows can struggle to accommodate how work actually happens. Teams adapt by working around the system, reintroducing informal coordination alongside the formal structure.</p><p>These patterns share a common characteristic. The system responds to signals, but does not improve how those signals are interpreted or acted upon. Coordination effort is redistributed, but not reduced.</p><p>This is the same failure mode that appeared in earlier waves of automation. Processes became faster and more visible, but remained fragmented. The system optimised activity within existing constraints rather than improving how the work held together.</p><h2>What are the conditions for success?</h2><p>When coordination is treated as a capability rather than a role, a different set of requirements becomes visible. These are not new technologies as such, but conditions that determine whether alignment can be sustained systematically.</p><ul><li><p><strong>Clarity of outcomes</strong> provides the foundation. The system needs a stable reference point against which progress can be interpreted.</p></li><li><p><strong>Shared context</strong> allows that interpretation to be consistent. Goals, roles, dependencies and constraints need to be visible in a form that can be related to ongoing activity.</p></li><li><p><strong>Access to systems and data</strong> determines the quality of signals available. Work that is dispersed across disconnected tools or constrained by interfaces limits the system&#8217;s ability to observe and respond. Coordination relies on the ability to relate activity across environments, rather than treating each system in isolation.</p></li><li><p><strong>Defined escalation pathways</strong> provide boundaries for decision-making. The system needs to know when to surface situations and how to direct them appropriately.</p></li><li><p><strong>Codified ways of working</strong> support continuity. Many coordination patterns already exist within organisations, but they are often informal and unevenly applied. Making these patterns explicit allows them to be observed, adapted and, where appropriate, supported by the system.</p></li></ul><p>These conditions make it possible for coordination to move from an informal practice to a structured capability. They allow alignment to be sustained through the interaction of systems and people, rather than relying on continuous manual intervention.</p><p>The effectiveness of coordination depends on how clearly the organisation can represent its own work, context and decision boundaries.</p><p>Read on for what this means for the future of the Project Manager role.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/the-autonomous-project-manager">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Agents of Progress or Agents of Chaos?]]></title><description><![CDATA[The delta between personal and enterprise agentic AI development is worrying, but perhaps combining the two offers a way to help overcome their limitations...]]></description><link>https://academy.shiftbase.info/p/agents-of-progress-or-agents-of-chaos</link><guid isPermaLink="false">https://academy.shiftbase.info/p/agents-of-progress-or-agents-of-chaos</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 31 Mar 2026 15:12:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!XMjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The <strong><a href="https://academy.shiftbase.info/p/how-we-survived-the-agent-apocalypse">OpenClaw moment we covered a few weeks ago</a></strong> was a wild ride. But in the age of YOLO, Hodl and r/wallstreetbets, it should come as no surprise that there is an apparently limitless supply of people <strong><a href="https://bsky.app/profile/mims.bsky.social/post/3mhsux67xpk2d">willing to hand over control of their personal computer</a></strong> to AI agents in pursuit of rapid progress.</p><p>One fascinating research study - <strong><a href="https://agentsofchaos.baulab.info/report.html">Agents of Chaos</a></strong> - let OpenClaw run riot within a controlled lab environment, and concluded:</p><blockquote><p><em>During a two-week experimental investigation, we identified and documented ten substantial vulnerabilities and numerous failure modes concerning safety, privacy, goal interpretation, and related dimensions. These results expose underlying weaknesses in such systems, as well as their unpredictability and limited controllability as complex, integrated architectures. The implications of these shortcomings may extend directly to system owners, their immediate surroundings, and society more broadly. Unlike earlier internet threats where users gradually developed protective heuristics, the implications of delegating authority to persistent agents are not yet widely internalized, and may fail to keep up with the pace of autonomous AI systems development.</em></p></blockquote><p>For those of us with a slightly lower risk appetite, Claude Cowork has also evolved very quickly as a more mainstream alternative to OpenClaw, and is incredibly impressive.</p><p>To take but one of many examples of personal agent setups, <strong><a href="https://promptedbyeric.substack.com/p/claude-cowork-might-be-the-most-consequential?r=9dv58&amp;triedRedirect=true">Eric Porres recently shared his own Claude Cowork harness</a> </strong>and lauded its ability to help him manage his portfolio of work activities in a more powerful and efficient way:</p><blockquote><p><em>the gap between &#8220;AI as chatbot&#8221; and &#8220;AI as operating system for your work&#8221; is closing fast. And Cowork is where that gap collapses for non-developers.</em></p></blockquote><p>At their impressive GTC event in San Jos&#233; recently, Nvidia placed great emphasis on this agentic inflection point as a pointer to where AI is headed next, and also crucially where it can start to show strong returns. <strong><a href="https://www.nvidia.com/gtc/keynote/">In his keynote, CEO Jensen Huang challenged companies to understand the nature of this shift</a></strong> with a typically provocative statement:</p><blockquote><p><em>Every company needs an OpenClaw strategy</em></p></blockquote><p>But this was not a crazy call for companies to let OpenClaw run riot in their organisations. <strong><a href="https://www.exponentialview.co/p/jensens-openclaw-thesis?publication_id=2252&amp;post_id=191613948&amp;triggerShare=true&amp;isFreemail=false&amp;r=9dv58&amp;triedRedirect=true">The message, as Azeem Azhar interpreted it, is a lot more far-reaching</a></strong>: it is about the shift from the model training era to one of inference and execution.</p><p>The dominant driver of AI progress (and NVIDIA&#8217;s revenue) was <em>training</em> compute: the vast one-time cost of training large foundation models. This emerging thesis says that the next scaling frontier is <em>inference-time</em> compute &#8212; spending more compute at the moment of generating a response, letting models &#8220;think longer&#8221; on hard problems (chain-of-thought, test-time search, etc.) rather than just being bigger. This changes the hardware economics significantly: inference demand is continuous, distributed, and latency-sensitive rather than concentrated in large training runs. It also opens up physical AI (robotics, autonomous systems) as a major new inference market.</p><p>The focus shifts from models to what we do with them - or as Azeem put it: <em><strong>&#8220;The harness is the revolution&#8221;</strong>&#8230;</em></p><blockquote><p><em>For AI, the harness moment happened at the tail end of 2025. Claude Code began to work reliably enough that you could leave it running overnight and trust what it had done in the morning. Not perfectly, but reliably enough. And that threshold, that &#8220;I can leave it to its own devices&#8221; threshold, changed everything. It changed what users asked AI to do. It changed how long tasks ran. It changed the token usage profile of every organization that crossed it.</em></p><p><em>Now, OpenClaw is the harness for the next layer.</em></p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XMjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XMjE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 424w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 848w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 1272w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XMjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic" width="800" height="533" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:533,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:78810,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/192740823?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XMjE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 424w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 848w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 1272w, https://substackcdn.com/image/fetch/$s_!XMjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80556756-6586-4ef9-8ac1-52bdfab40918_800x533.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">image credit: https://www.linkedin.com/posts/i0exception_san-francisco-these-days-activity-7429390168576405504-D8Nr/</figcaption></figure></div><h2>Talk to my agent!</h2><p>What does this mean for agentic AI in the enterprise, and should we be concerned about the way organisation-owned agentic services seem to be lagging so far behind the rapid evolution of personal agents?</p><p>Personal AI agents have evolved so much faster than enterprise agentic services that the gap between them is becoming structural, and we could end up with more ways to navigate broken enterprise systems instead of business transformation.</p><p>But if we embrace personal agents at work as prototypes and testbeds for shared enterprise services, then perhaps we can close the gap.</p><p>Looking at this challenge within the wider context of the shift from model training to inference and runtime intelligence, however, it is clear that much more focus is needed on world models and decision intelligence infrastructure to enable shared enterprise agents to work reliably with less human supervision. The good news is we can do much of this with the tools we have access to today. It is less of a tech challenge and more of a readiness / architectural question.</p><ul><li><p>Microsoft has recently been beefing up its agentic capabilities in its M365 platform, and <strong><a href="https://techcommunity.microsoft.com/blog/microsoft365copilotblog/introducing-multi-model-intelligence-in-researcher/4506011">has just announced a multi-model capability for verifying complex research</a></strong>.</p></li><li><p><strong><a href="https://venturebeat.com/technology/perplexity-takes-its-computer-ai-agent-into-the-enterprise-taking-aim-at">Perplexity has launched an agentic harness for the enterprise</a></strong>, based on internal tooling that was used by its own employees to speed up delivery.</p></li><li><p>Elsewhere, Salesforce&#8217;s agentic foundry, SAP and a host of other stalwarts from the previous generation of enterprise platforms continue to announce new agentic capabilities.</p></li></ul><p>But the capability diffusion gap continues to widen, and there is now a risk that personal agents will evolve so much faster than enterprise agents that we could recreate the &#8216;old wine in new bottles&#8217; problem that we saw with the earlier phase of Robotic Process Automation (RPA), and use AI to navigate a broken system better, rather than to fix the system or start building a better one.</p><p><strong><a href="https://a16z.com/why-the-world-still-runs-on-sap/">As Eric Zhou and Seema Amble of Andreesen Horowitz remarked recently, the world still largely runs on old, poorly-designed enterprise platforms not because they are good</a></strong>, but because organisations contorted themselves around their inadequacies and foibles to such an extent that ripping them out could be painful:</p><blockquote><p><em>To ask a question that sounds almost disrespectful until you&#8217;ve spent a week in a Fortune 500: why do people still use SAP (and ServiceNow, and Salesforce) at all?</em></p><p><em>The short answer is that SAP, or any major legacy system of record, captures critical data across the businesses that use it. But on top of that, the business has customized it and built a set of specific procedures and roles on top of it, much of which is not actually documented anywhere.</em></p></blockquote><p>Or at least that was the case until now. The authors argue that agentic AI in the enterprise could replace these behemoths over the medium term, but even in the short-term, it could make them more malleable and easier to work with.</p><p>Perhaps this is an area where personal agents in the enterprise can become a testbed for enterprise agentic services, trying out automations, workarounds and multi-agent tasks under human oversight, before becoming candidates for new shared services that run more autonomously.</p><p>With a personal agent harness, we can each maintain our own personal work context, memory, preferred styles and so on, and our agents will be able to use this to navigate the enterprise and interface with systems of record at the API and data level, bypassing the UI altogether. And our agents can talk to each other to take over a lot of the boring, time-wasting scheduling, alignment, stakeholder communications and basic coordination tasks that consume so much of a leader&#8217;s time in large enterprises today.</p><p>Later, as more shared enterprise agents come online, our personal agents could help manage our relationships with these as well, for example sequencing the various actions needed for us to run a project or gather intelligence for new ideas.</p><p>That suggests to me that we should try to find a way to embrace technologies like Claude Cowork safely within the enterprise to deliver on the potential that copilots promised. Nothing is risk-free, and we clearly need to focus on guardrails and permissions, but if advanced users are willing to be accountable for their agents in return for the productivity and value they could generate, then we can probably find ways to make it work.</p><h2>World Knowledge, Architecture and Run-time Intelligence</h2><p>But it is also worth thinking about the differences between personal agents and enterprise agents, and what we have learned so far on this journey of discovery.</p><p>First, we need a more considered and thoughtful approach to productivity than boasting about how many lines of code (LOC) we can churn out.</p><p><strong><a href="https://mariozechner.at/posts/2026-03-25-thoughts-on-slowing-the-fuck-down/">Mario Zechner tried to summarise his own lessons from agentic coding</a></strong> last week in an interesting, opinionated piece about the dangers of brittle software, missed learning and unmaintainable systems, and concluded we need to <em>&#8220;slow the f</em>** down.&#8221;*</p><p>In a similar vein, <strong><a href="https://interconnected.org/home/2026/03/28/architecture">Matt Webb reminds us that good architecture beats a high LOC count every time</a></strong>, and helps avoid personal-agents-as-workarounds suffering from the same failure modes as RPA, mentioned earlier:</p><blockquote><p><em>The thing about agentic coding is that agents grind problems into dust. Give an agent a problem and a while loop and - long term - it&#8217;ll solve that problem even if it means burning a trillion tokens and re-writing down to the silicon.</em></p><p><em>Like, where&#8217;s the bottom? Why not take a plain English spec and grind in out in pure assembly every time? It would run quicker.</em></p><p><em>But we want AI agents to solve coding problems quickly and in a way that is maintainable and adaptive and composable (benefiting from improvements elsewhere), and where every addition makes the whole stack better.</em></p><p><em>So at the bottom is really great libraries that encapsulate hard problems, with great interfaces that make the &#8220;right&#8221; way the easy way for developers building apps with them. Architecture!</em></p></blockquote><p>Another question is where we can live with probabilistic models, and where we need a more deterministic approach. We can tolerate the limits of probabilistic models in personal agents we oversee and can test, but when we start to think about embedded or autonomous agents that work the same way for everybody, a more deterministic approach is often needed.</p><p>Just as Nvidia&#8217;s forward strategy is about more real-time inference in the output stage, we can start to imagine how more verification and compliance could also be done at runtime to make some enterprise AI agents more deterministic.</p><p>Artur Huk wrote about this a few days ago for O&#8217;Reilly, <strong><a href="https://www.oreilly.com/radar/the-missing-layer-in-agentic-ai/">describing &#8216;decision intelligence runtime&#8217; as a missing capability layer in agentic AI</a></strong> - more of an engineering pattern than a specific solution or technique.</p><p>And this brings us back to a topic we have been noodling on for some time, which is <strong><a href="https://academy.shiftbase.info/p/ai-round-up-both-destroyer-and-maker">the vital importance of world-building</a></strong> in ensuring the success of enterprise AI.</p><p>Whilst some aspects of world-building are about creating good general context for people and machines to have clarity about goals, ways of working culture, language and so on, there are other aspects of world knowledge that are more precise and scientific.</p><p>The way that current models handle world knowledge is largely in the training stage, rather than the inference or runtime moment. The way autonomous driving systems learn is a good example. <strong><a href="https://www.wired.com/story/a-school-district-tried-to-help-train-waymos-to-stop-for-school-buses-it-didnt-work/?utm_source=nl&amp;utm_brand=wired&amp;utm_mailing=WIR_Daily_033026_PAID&amp;utm_campaign=aud-dev&amp;utm_medium=email&amp;utm_content=WIR_Daily_033026_PAID&amp;bxid=679fe6b740ae7ef26b0b090f&amp;cndid=86084401&amp;hasha=016003755aa0296c369a337f05a1c1d7&amp;hashc=6682d98588cab4309c24846dca6db095f2734170cec7daa3428e1be66edfc3ed&amp;esrc=MARTECH_ORDERFORM&amp;utm_term=WIR_DAILY_PAID">Waymo vehicles in Austin developed a nasty habit of illegally overtaking school buses on pick up</a></strong>, and the school district worked with the company to give them simple rules and guidance on how to avoid this happening. But the training process for Waymo&#8217;s system is so long and includes so much data, that they were unable to simply add a new rule quickly, and the cars kept overtaking buses. More runtime inference and decision intelligence based on world models is perhaps one way to tackle such anomalies.</p><p><strong><a href="https://joiningdots.substack.com/p/ai-and-a-new-frontier-for-decision">So much of the existing decision intelligence inside organisations has never been captured, as Sharon Richardson remarked yesterday</a></strong> in her informative piece about context graphs, which means there is a lot we can achieve quite easily and quickly if we are smart about it. This does not require some kind of Sisyphean manual knowledge mapping exercise, because it is the kind of task that AI can accelerate with the right supervision, even down to the level of conducting structured interviews or After Action Reviews (AARs) to capture decision traces and reasoning from people.</p><p>The kind of world models we will need to realise the promise of enterprise agentic AI will go way beyond intangible knowledge and culture. They will need to understand the physical world, manufacturing, distribution and even domains like geopolitics that (not again!) are messing with supply chains and pricing.</p><p>How we build and evolve these models is truly a fascinating challenge, and one where we can use AI itself to help improve our AI readiness by doing the mapping, collating and documentation of the information we need to make them real.</p><p><strong><a href="https://www.strangeloopcanon.com/p/the-future-of-work-is-world-models">Rohit Krishnan recently wrote that this idea is really the key to the future of work</a></strong>, which he sees operating more like a strategy co-op game than a single-player game, and I think that is right.</p><blockquote><p><em>What&#8217;s needed in the enterprise world is such a world model - an engine that knows the rules, tracks the state, understands and predicts consequences.</em></p><p><em>The environment would connect to the systems a company already runs, the information that is gathered, the agents it uses, and build a live operational model of the business. Scale it across companies and you have the training data to build a compelling environment and an even better world model!</em></p></blockquote><p>The question is, can visionary CIOs and leaders of AI adoption programmes make the case that urgent attention and investment is needed in AI readiness efforts rather than just rolling out co-pilot licenses and hoping for some marginal productivity gains?</p><p>We need composable, addressable processes, services and systems if enterprise agents are to operate autonomously. And we need codified rulesets, world models and decision intelligence to be available at run-time if we want them to operate more deterministically without the kind of oversight we perform with personal agents.</p>]]></content:encoded></item><item><title><![CDATA[Same Models, Different Worlds]]></title><description><![CDATA[How outcomes, skills and context combine to create compounding organisational intelligence]]></description><link>https://academy.shiftbase.info/p/same-models-different-worlds</link><guid isPermaLink="false">https://academy.shiftbase.info/p/same-models-different-worlds</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 24 Mar 2026 15:08:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!H_5r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI models are becoming shared utilities. When the same intelligence is available to everyone, the real differentiator is no longer the model, but the world around it. If your competitors are using the same models as you, trained on similar data and accessed through the same interfaces, then the intelligence itself cannot be the thing that gives your organisation competitive advantage.</p><p>If everyone has access to an identical intelligence layer, where does differentiation actually come from?</p><p>At first glance the answer might seem to lie in prompting skill or in deploying more sophisticated agents and tools on top of the models. But the organisations gaining the most value from AI are not simply using the models differently, they are building richer environments around them. And the more intelligence you retain in the environment, the easier it becomes to switch models when needed.</p><p>Instead of treating AI as a standalone tool, enterprise AI pioneers are beginning to shape the world that AI operates within: the outcomes it is responsible for, the expertise it inherits, and the context it can access. This is also creating a clearer distinction between skills and context in terms of agentic architecture.</p><h2>Memory and Ownership</h2><p>In everyday AI usage, people are starting to build up context inside individual tools: conversations, prompts, working patterns, fragments of reasoning. Over time, this becomes something more than usage. It becomes a kind of working memory.</p><p>&#8594; But that memory is fragile.</p><p>&#8594; Switch models, and it disappears.</p><p>&#8594; Move tools, and it fragments.</p><p>&#8594; Hit usage limits, and it resets.</p><p>For those focused on using commercial tools, running into usage limits with, say, Claude and potentially having to switch models is not just inconvenience, it is the loss of accumulated context. Conversations, assumptions, and working patterns have to be rebuilt from scratch.</p><p>The same pattern is playing out inside organisations.</p><p>Copilots sit in different tools. Context is scattered across systems. There is no shared memory layer, no consistent way to carry forward how work is done, what has been learned, or how decisions have been made.</p><p>The intelligence may be shared, but the context is not; and more importantly, it is not owned.</p><h2><strong>Tools vs Worlds</strong></h2><p>On paper, most AI deployments look remarkably similar. The same models sit underneath a fairly common toolset. The same copilots appear in productivity software and the same agent frameworks promise orchestration across workflows. From the outside, it can feel as though every organisation is drawing from the same intelligence layer, and in many ways, they are.</p><p>But beneath that shared surface, two very different approaches are being used.</p><p>In some organisations, AI remains primarily a tool layer. Employees interact with models through prompts and copilots, using them to draft content, analyse documents, generate ideas or automate parts of existing workflows. The intelligence sits at the interface: helpful, but largely separate from the deeper structure of how the organisation operates.</p><p>In others, a different approach - instead of focusing only on how people interact with AI, attention begins to shift toward the environment.</p><ul><li><p>What outcomes the system is responsible for.</p></li><li><p>What expertise it inherits from the organisation.</p></li><li><p>What context it can access across tools, data and systems.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H_5r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H_5r!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H_5r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:299990,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/191608578?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H_5r!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!H_5r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F023a5b26-a264-4d5a-86c8-98bd6cf13f88_1536x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This is where an architectural distinction that is gaining traction in agent systems becomes useful. The difference between Skills and context systems such as MCP. At a simple level, the distinction is straightforward:</p><ul><li><p><strong>Skills</strong> describe how the system approaches a problem. They encode reasoning patterns, frameworks and playbooks that guide analysis or decision-making, but increasingly they also define how work gets done in practice: the sequences of actions, process steps and interactions the system can carry out across tools and workflows.</p></li><li><p><strong>Context systems</strong> determine what the system can see and interact with. They provide structured access to documents, tools, data sources and workflows across the organisation.</p></li></ul><p>AI systems do not become powerful simply because the models improve; they become powerful when three things begin to align:</p><ul><li><p>the outcomes they are responsible for</p></li><li><p>the skills they can use to reason and act</p></li><li><p>the context they can access across the organisation</p></li></ul><p>Together, these elements start to form something that looks less like a tool and more like an operational environment for intelligence.</p><h2>Tool Mode: Intelligence Without a World</h2><p>Digging a bit deeper into the differences, it is clear that in Tool Mode, AI is treated primarily as an interface for generating output.</p><p>Employees prompt copilots to draft documents, analyse data, generate presentations, or summarise discussions. Individual productivity improves, and in many cases the gains are significant. Teams can move faster in small pockets, content can be produced more easily, and analytical work that once took hours can often be completed in minutes.</p><p>But the intelligence remains largely detached from the organisation itself.</p><p>The model has limited awareness of the systems people use every day. It does not have structured access to decision histories, internal frameworks, or the reasoning patterns that shape how the organisation operates. Knowledge remains scattered across documents, chat threads and individual expertise. As a result, AI becomes powerful but context-poor.</p><p>Outputs may be impressive in isolation, but they often lack alignment with the organisation&#8217;s specific priorities, norms and operating logic. Two employees asking the same question may receive different answers. Strategic nuance is lost. Decisions remain dependent on individuals interpreting and adjusting the output.</p><p>In this mode, AI behaves a little like a brilliant intern dropped into the organisation with access to a search engine but very little understanding of how the company actually works.</p><h2>World Mode: Intelligence Inside an Environment</h2><p>When organisations focus on designing the environment, instead of just prompting and tool use, attention can shift towards:</p><ul><li><p>What expertise should be codified and reusable?</p></li><li><p>And what context should be systematically accessible?</p></li></ul><p>This is where the distinction between Skills and context systems becomes operational.</p><p>Skills capture reasoning patterns that previously lived mostly in people&#8217;s heads, but they can also encode how those patterns translate into action: what steps to take, which tools to use, and how work flows from one stage to the next. They encode the frameworks, heuristics and analytical approaches that experienced practitioners use when evaluating problems or making decisions.</p><p>A competitive analysis skill, for example, might specify how to compare products across pricing, features, positioning and risk, and how to gather that information across internal and external sources as part of a repeatable workflow. A risk assessment skill might define the dimensions that should be considered before approving a supplier or launching a new initiative. These patterns represent organisational expertise. When codified as reusable skills, they become something new: shared reasoning infrastructure.</p><p>Context systems complement this by shaping the environment the intelligence operates within. Through mechanisms such as MCP, agents can access the tools, documents, databases and workflows that contain the organisation&#8217;s operational knowledge.</p><p>When outcomes, skills and context begin to align, the intelligence is no longer simply generating output. It begins participating in the operational fabric of the organisation, both in how decisions are made and in how work is executed.</p><h2>Why Tool Mode Is the Default</h2><p>Just as <strong><a href="https://academy.shiftbase.info/p/extraction-vs-redesign-the-hidden">Extraction Mode dominates organisational transformation</a></strong>, Tool Mode tends to dominate early AI adoption.</p><p>The reasons are straightforward.</p><p>Deploying AI tools is easy. Codifying expertise and structuring context is not.</p><p>It is far simpler to train employees in prompting or roll out copilots across the organisation than it is to examine how knowledge actually flows through the system. Codifying reasoning patterns requires surfacing tacit expertise that may never have been formally articulated. Structuring context requires connecting fragmented systems and clarifying which information should be authoritative. Both activities are organisational work rather than purely technical work. Most enterprises therefore begin with the most visible layer: interaction with the model.</p><p>But this choice has consequences: the intelligence becomes faster, but the organisation does not necessarily become smarter.</p><h2>The Mechanics of World-Building</h2><p>In our previous piece <strong><a href="https://academy.shiftbase.info/p/a-leaders-guide-to-world-building">we explored the idea of world-building as a leadership discipline</a></strong>, the craft of designing the environments in which human and machine intelligence operate together. Organisations were compared to worlds with their own physics, culture and geography: rules that shape behaviour, norms that guide judgment, and environments that determine how actors navigate the system.</p><p>The distinction between outcomes, skills and context begins to reveal what those layers look like in practice.</p><ul><li><p><strong>Outcomes define the direction of the world</strong>. They describe what success looks like and where responsibility ultimately sits. When AI systems are attached to outcomes rather than isolated tasks, they begin participating in the organisation&#8217;s operating logic rather than simply generating output.</p></li><li><p><strong>Skills represent a form of codified expertise</strong>. They capture the reasoning patterns that experienced practitioners use when analysing problems, evaluating trade-offs or making decisions. In world-building terms, they begin to encode aspects of the organisation&#8217;s culture &#8212; the ways it interprets information, the factors it considers important and the principles that guide judgment.</p></li><li><p><strong>Context systems provide the geography of the world</strong>. Through mechanisms such as MCP, agents gain structured access to documents, tools, data and workflows. These systems determine what the intelligence can see, what information it can retrieve and where it can act.</p></li></ul><p>Seen together, these elements begin to form the operational layer of world-building. The models may be shared across organisations, but the <strong>world around them is not</strong>.</p><p>One organisation may give an agent access to fragmented documentation and informal processes. Another may provide structured context, codified expertise and clearly defined outcomes. The underlying intelligence is the same, but the environment it operates within is fundamentally different.</p><p>Let&#8217;s explore these three layers in more depth.</p><h2>Outcomes Define Direction</h2><p>Before thinking about skills or context, start with something simpler: what the system is actually trying to achieve.</p><p>Most early uses of AI focus on tasks, drafting, summarising, analysing. Useful, but peripheral. The centre of the organisation is not tasks, it is outcomes.</p><ul><li><p>Reduce customer churn</p></li><li><p>Improve supplier reliability</p></li><li><p>Increase conversion</p></li><li><p>Resolve support issues faster</p></li></ul><p>When intelligence is attached to outcomes, the system begins to organise differently. Decisions, data and workflows align around a shared objective rather than fragmenting across individual activities.</p><p>This is the shift from AI as a productivity tool to AI as part of how the organisation delivers results.</p><p>Outcomes define direction, skills and context can only make sense once that direction is clear.</p><h2>Learning Becomes Codification</h2><p>If outcomes define <em>what</em> matters, skills define <em>how</em> the organisation gets things done.</p><p>In AI systems, a &#8220;skill&#8221; is simply a codified way of approaching a problem, a reusable reasoning pattern. Instead of asking a model to &#8220;analyse a market,&#8221; a skill defines how that analysis should be done: what to compare, what structure to follow, what constitutes a good answer.</p><p>For example, a competitive analysis skill might require comparison across pricing, features, positioning and risk, and end with clear strategic implications. What matters is not the example, but the shift.</p><p>Learning no longer lives only in people. It becomes something that can be captured, reused and improved.</p><p>Frameworks, heuristics and decision patterns that were once taught informally start to become shared infrastructure. Different teams stop reinventing the same thinking. AI systems and humans begin to draw on the same reasoning patterns.</p><p>Learning moves from training individuals to building organisational capability.</p><h2>Context Becomes Architecture</h2><p>If skills define how the organisation thinks, context determines what it can act on.</p><p>Without structured context, AI operates in isolation, limited to prompts and general knowledge. With it, intelligence becomes connected to the organisation&#8217;s actual work. This is where approaches like Model Context Protocol (MCP) matter as they provide a structured way for AI systems to access:</p><ul><li><p>internal documents</p></li><li><p>data sources</p></li><li><p>business tools</p></li><li><p>workflows</p></li></ul><p>So instead of generating answers in the abstract, the system works with live organisational information. At that point, the quality of the environment starts to matter as much as the quality of the model.</p><p>Two organisations using the same AI can produce very different results depending on how well their context is structured.</p><p>If organisations are beginning to codify outcomes, skills and context, an interesting question surfaces: where does this knowledge actually live?</p><p>In software engineering, GitHub provides a shared environment where code can be stored, improved and versioned collaboratively. A similar concept may emerge for organisational intelligence: a place where skills, decision rules and context connections can be maintained as shared infrastructure.</p><p>You could think of it as a kind of GitHub for world-building - a repository where the logic of how the organisation operates becomes visible, improvable and reusable.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/same-models-different-worlds">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Agents on the Night Shift]]></title><description><![CDATA[Self-improving agents, Socratic dialogue and temporal stress: things to think about on the way towards agentic engineering and machines that make machines]]></description><link>https://academy.shiftbase.info/p/agents-on-the-night-shift</link><guid isPermaLink="false">https://academy.shiftbase.info/p/agents-on-the-night-shift</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 17 Mar 2026 15:30:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lS-s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong><a href="https://datasciencedojo.com/blog/karpathy-autoresearch-explained/">Andrej Karpathy&#8217;s new autoresearch tool recently ran 700 experiments on his nanochat codebase in two days</a></strong>. It found 20 improvements he had missed, delivering an 11% uplift in output. Tobi L&#252;tke at Shopify tried it on his own hand-tuned model: 19% improvement, parameter size halved. What makes this remarkable is not the numbers. It is the mechanism. The tool does not just run tests &#8212; it updates its own Python code based on what it learns. The researcher sets the direction; the machine runs experiments overnight and arrives with findings.</p><p>This is one example of an important archetype I think we will see more and more in agentic AI: <em><strong>the machine that makes the machines</strong></em>, which is both exciting and slightly strange.</p><p>This raises interesting questions about where learning lives in a human-machine system, and who benefits from it.</p><p><strong><a href="https://adactio.com/journal/22436">As Jeremy Keith put it recently</a></strong> when considering how agentic coding is changing development practices:</p><blockquote><p><em>Outsourcing execution to machines makes a lot of sense.</em></p><p><em>I&#8217;m not so sure it makes sense to outsource learning.</em></p></blockquote><p>But the productive division isn't just human vs. machine learning &#8212; it's human imagination operating at the meta-hypothesis level, and machine speed exhausting the territory around it. A single wild guess or idea can now seed hundreds of downstream tests; what comes back isn't just an answer, but a richer map of the problem space than any individual researcher might have drawn alone.</p><h2>From Agentic Coding to Agentic Engineering</h2><p>It seems everybody is intrigued right now by the rapid changes that the latest agentic AI models are bringing to software development, and it is worth paying attention because a similar process is likely to play out across other areas of work.</p><p>The New York Times Magazine recently published a major feature, <strong><a href="https://www.nytimes.com/2026/03/12/magazine/ai-coding-programming-jobs-claude-chatgpt.html?unlocked_article_code=1.SlA.DBan.wbQDi-hptjj6">Coding After Coders: the End of Computer Programming as we Know it</a></strong>, covering the history of the field and the experience of developers navigating rapid transformation:</p><blockquote><p><em>How things will shake out for professional coders themselves isn&#8217;t yet clear. But their mix of exhilaration and anxiety may be a preview for workers in other fields. Anywhere a job involves language and information, this new combination of skills &#8212; part rhetoric, part systems thinking, part skepticism about a bot&#8217;s output &#8212; may become the fabric of white-collar work. Skills that seemed the most technical and forbidding can turn out to be the ones most easily automated. Social and imaginative ones come to the fore. We will produce fewer first drafts and do more judging, while perhaps feeling uneasy about how well we can still judge. Abstraction may be coming for us all.</em></p></blockquote><p>This is almost certainly not the end of computer programming as a discipline, despite the pace of change. Computer science will become more sciencey; programming &#8212; talking to computers &#8212; will become more literary. But the need for people who understand what is possible and how to make it happen will continue to grow.</p><p>But agentic coding is also creating new forms of cognitive overload among AI-assisted developers, including the puzzling sight of people <strong><a href="https://www.wsj.com/tech/ai/ai-bots-claude-openclaw-285ac816">sitting outside in Silicon Valley watching &#8212; but not touching &#8212; their laptops</a></strong> as coding agents grind through the work.</p><p>Matt Jones captured this strangeness well this week in a lovely piece of writing &#8212; <strong><a href="https://petafloptimism.com/2026/03/14/gas-town-and-bullet-hell/">Gas Town and Bullet Hell</a></strong> &#8212; in particular the temporal mismatch between human cognition and machine speed:</p><blockquote><p><em>If brain fry is a clock problem &#8212; a temporal mismatch between human cognition and machinic speed &#8212; then solutions that only address interface design or training will help at the margins but miss the structural issue&#8230;</em></p><p><em>If we want AI agent work to feel more like flow and less like fry, the challenge isn&#8217;t making things faster or even slower &#8212; it&#8217;s about legibility, consent, and reversibility, and all three matter at once.</em></p></blockquote><p>As we hit the cognitive limits of what single-player mode can achieve, the shift from agentic coding to agentic engineering becomes important.</p><p><strong><a href="https://simonwillison.net/guides/agentic-engineering-patterns/what-is-agentic-engineering/">Simon Willison has a typically thorough guide to what this means in practice</a></strong>: instead of using an agent to write some code, agentic engineering means tasking systems with higher-order goals and the ability to self-manage the path towards them with less micro-management. The craft, Willison argues, was never primarily about writing code &#8212; it was always about figuring out what code to write.</p><h2>The Organisation as the Machine That Makes the Machines</h2><p>We have argued for a long time that to produce good software, organisations need to become like software themselves.</p><p>Corporate failures such as <strong><a href="https://germanautopreneur.com/p/cariad-volkswagen-software-failure-lessons">Volkswagen&#8217;s first attempt to build a software division for its vehicles within a hierarchical and bureaucratic organisation</a></strong> prove the point. The technology was not the problem; the organisational architecture was.</p><p>And yet, elsewhere in the automotive world, this has been understood for some time. <strong><a href="https://newsletter.jurriaankamer.com/p/the-organization-is-the-machine">Jurriaan Kamer recently shared lessons from F1 teams</a></strong>, quoting a team principal on what they borrowed from the Apollo project in their pursuit of agility and excellence under pressure:</p><blockquote><p><em>&#8220;What you can&#8217;t have is an engineer here having to go up and down a particular hierarchy and then hop across &#8212; in our instance, not just a different geographic location, but a different country altogether &#8212; and then go up and down. So instead, it&#8217;s a kind of different structure where it&#8217;s <strong>mission control instead of command and control</strong>.&#8221;</em></p></blockquote><p>This distinction matters more than it might appear. Today, a developer can use an agent to write better static software, and that is a productivity story everybody can follow. But if we trace the trajectory of agentic engineering towards its logical conclusion &#8212; and Karpathy&#8217;s autoresearch is an early signal of where that leads &#8212; we will need a much more fluid and connected organisational structure where services and processes are digitised and addressable, so that they can become truly programmable and genuinely capable of self-improvement.</p><p>The organisation itself needs to become the machine that makes the machines. <strong><a href="https://www.youtube.com/watch?v=MiUHjLxm3V0&amp;t=2872s">ASML&#8217;s famous EUV system</a></strong> is a useful reference point: a machine so complex that it requires extraordinary coordination between hundreds of specialist suppliers and internal teams, but one whose design assumes that it will be continuously improved by the people who build and operate it. The infrastructure is not static. It learns.</p><p>This also brings the learning question back into focus. If the machine is updating its own code overnight and accumulating insights from hundreds of experiments, organisations need to build the governance and oversight architecture that keeps humans genuinely in the loop &#8212; not as approvers of every output, but as the people setting direction, interpreting results, and carrying the institutional memory that the machine cannot hold. Otherwise, you end up with iteration without learning, which is just faster drift.</p><p><strong><a href="https://www.hulme.ai/blog/what-3-000-years-of-philosophy-and-three-decades-of-agent-research-can-teach-us-about-the-next-three-years">As Daniel Hulme reminds us in his recent thoughtful account of the philosophical and historical pre-cursors of agentic AI</a></strong>, we already have rich bodies of knowledge and methods to draw on:</p><blockquote><p><em>The irony of this moment is that we are simultaneously living through the most rapid deployment of autonomous agents in history and underutilising the most relevant bodies of knowledge ever produced on how to make such systems safe. From Socrates&#8217; method of structured interrogation to Aristotle&#8217;s formal logic, from Chrysippus&#8217; propositional reasoning to the medieval protocols of adversarial disputation &#8211; and then from Carl Hewitt&#8217;s Actor Model to Michael Bratman&#8217;s theory of practical reasoning, from Leslie Lamport&#8217;s work on distributed consensus to Edmund Clarke&#8217;s model checking, from Lotfi Zadeh&#8217;s fuzzy logic to the agent architectures of Michael Wooldridge and Nick Jennings &#8211; these thinkers and many others spent careers building the conceptual and mathematical toolkit for exactly the challenges we now face. Their work isn&#8217;t historical curiosity. It&#8217;s a foundation we should be actively building on</em></p></blockquote><p>The same could be said of our accumulated knowledge about organisational design. How systems learn, adapt, and maintain coherence under rapid change is not a new problem. We just have a new urgency to solve it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lS-s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lS-s!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lS-s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:244254,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/191252173?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lS-s!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 424w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 848w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 1272w, https://substackcdn.com/image/fetch/$s_!lS-s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ed4363-73ff-4acb-8746-8d2e158bc7de_1408x768.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The Infrastructure Is Coming</h2><p>The broader technology ecosystem is already moving in this direction. <strong><a href="https://www.interconnects.ai/p/the-next-phase-of-open-models">Nathan Lambert&#8217;s survey of the current state of open AI models</a></strong> suggests we will eventually reach a place where specialised small models are freely available for organisations to adapt and build on when creating their own AI platform architectures.</p><p><strong><a href="https://www.constellationr.com/insights/news/nvidias-huang-all-software-will-be-agentic">Jensen Huang is unambiguous about where this leads</a></strong>:</p><blockquote><p><em>&#8220;There will be no software in the future that&#8217;s not agentic. How could you have software that&#8217;s dumb? And so, it is absolutely true that every software company will become an agentic company.&#8221;</em></p></blockquote><p>Instead of using AI agents to write better SaaS tools, this implies that software firms will make available agents that can continuously write, maintain, and evolve living software &#8212; software that has a sense of its own role and mission.</p><p>Incidentally, this also supports <strong><a href="https://x.com/JulienBek/status/2029680516568600933">the thesis that &#8216;services as software&#8217; will be a major new opportunity for specialist service providers</a></strong>.</p><p>Futurum Group <strong><a href="https://futurumgroup.com/press-release/cio-ai-priorities-pivot-from-productivity-to-innovation/">published new research this week on CIO AI priorities</a></strong>, finding that enterprise goals are shifting from basic efficiency towards innovation and organisational change. Dion Hinchcliffe&#8217;s conclusion that &#8220;the generic efficiency argument for AI is dead&#8221; is heartening. The route to greater returns is more about systems and architecture than it is about individual tool use, and it seems more enterprise leaders are beginning to see this. The danger is that &#8220;innovation and organisational change&#8221; becomes the new banner under which old structures get expensively automated rather than genuinely redesigned.</p><h2>Hypotheses &amp; Organisational Learning</h2><p>Karpathy ran 700 experiments in 48 hours on a well-defined optimisation problem with clean metrics and the ability to measure improvement objectively. That particular set of conditions is relatively rare. Most organisational improvement problems do not have clean metrics, do not produce outputs that can be evaluated overnight, and do not have the structured test environment that makes autoresearch possible.</p><p>What humans might lack in speed of iteration, they more than make up for in their ability to generate the wild guesses and <em>what ifs</em> that make for rich experimentation. Super-charging this innate human capability with the power of machines to loop through variations or play out scenarios could accelerate our learning and innovation in exciting new ways.</p><p>What autoresearch points towards isn't the automation of discovery, but its amplification. The human makes the leap; the machine explores where it lands.</p><p>Every organisation already has processes that could, with sufficient effort, be made legible, measurable, and addressable. The question for leaders is not whether to wait for the infrastructure to arrive &#8212; it will. The question is whether the organisation they are building now can actually use it when it does. The machine that makes the machines requires a very different kind of organisation than the one that deploys tools to make existing tasks faster.</p><p>What is the hypothesis-testing loop in your organisation that you most wish you could accelerate? And who, right now, is doing the learning?</p><h2>A Quick Favour to Ask</h2><p><strong><a href="https://letter.rebuild.net/">Please consider signing the Rebuild Letter</a></strong> to support a great initiative I have been loosely involved in over the last year or so that aims to stimulate the development of better European social tools and networks to reduce our reliance on weaponised attention farming.</p>]]></content:encoded></item><item><title><![CDATA[Programmable Governance & Probabilistic Humans]]></title><description><![CDATA[How probabilistic humans and collective intelligence could reshape AI governance]]></description><link>https://academy.shiftbase.info/p/programmable-governance-and-probabilistic</link><guid isPermaLink="false">https://academy.shiftbase.info/p/programmable-governance-and-probabilistic</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 10 Mar 2026 15:19:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MqTX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It might sound counter-intuitive at first, but good AI governance needs to become both programmable and probabilistic if organisations are to make meaningful use of human judgement alongside machine intelligence.</p><p>The idea of &#8216;human-in-the-loop&#8217; governance is simple and reassuring - AI systems may assist, recommend or automate, but somewhere in the process a human remains responsible for oversight and final judgement.</p><p>For early pilot deployments this model works reasonably well. Humans review outputs, approve sensitive actions or intervene when something appears wrong. But as AI systems become faster and more autonomous, we hit the limits of this approach quite quickly.</p><p>The scale and speed of modern systems will soon outstrip the cognitive bandwidth of manual oversight. A governance model built around humans inspecting individual outputs simply does not scale.</p><p>This does not make human judgement less important. If anything, the opposite is true. But the role humans play in governance needs to evolve.</p><p>In practice, most leaders already hold nuanced views about emerging risks. A security leader may suspect that monitoring systems could fail under certain conditions. A product leader may worry about reputational edge cases. Legal teams often sense regulatory ambiguity long before it becomes formal policy.</p><p>Yet governance structures rarely capture these insights clearly. They compress judgement into binary decisions: approved or rejected, compliant or non-compliant, acceptable or unacceptable, red or green.</p><p>An under-used superpower that organisations already possess, and which could help here, is the ability to harness distributed human judgement about uncertainty.</p><p>A probabilistic human expresses judgement differently. Instead of presenting certainty where none exists, they estimate likelihoods and levels of confidence. When those signals can be aggregated across many individuals, organisations gain a clearer picture of how risk is evolving across the system. This opens the door to a different kind of governance. Rather than inserting humans into isolated approval points, organisations can begin to treat collective judgement as a continuous signal about how uncertain the system really is.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MqTX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MqTX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 424w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 848w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 1272w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MqTX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:270538,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/190504879?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MqTX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 424w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 848w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 1272w, https://substackcdn.com/image/fetch/$s_!MqTX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe2651c6-e023-4cec-8c4b-64f36ae2a4d5_1456x971.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Several mechanisms are beginning to emerge to support this shift. Some organisations experiment with prediction markets, allowing distributed expertise to converge into probabilistic forecasts about emerging risks, whilst others introduce structured dissent mechanisms, deliberately creating space for people to challenge prevailing assumptions and surface potential failure modes. And some leadership teams are beginning to convene probabilistic risk councils, where uncertainty is discussed explicitly and collective judgement informs governance decisions.</p><p>Taken together, these approaches allow organisations to move beyond episodic oversight toward something more adaptive: continuous calibration of uncertainty.</p><p>Let&#8217;s explore how these techniques work in practice and how leaders can use them to make human judgement legible inside increasingly complex AI systems.</p><h2>Why human-in-the-loop governance begins to strain</h2><p>For many leadership teams, the idea of human oversight in AI systems feels like a reassuring safeguard. If automated systems introduce uncertainty, the obvious response is to ensure that a person remains responsible for the final decision.</p><p>Yet organisations deploying AI at scale quickly encounter a different reality.</p><p>The challenge is not simply that systems are autonomous. It is that they operate within environments defined by <strong>speed, complexity and data volume</strong> that far exceed what traditional governance processes were designed to handle.</p><p>AI systems interact with constantly changing data, evolving models and interconnected workflows. Decisions that once occurred occasionally may now occur thousands of times per hour. Risk signals appear not as clear incidents but as patterns emerging across vast streams of activity.</p><p>Under these conditions, governance based on periodic review begins to struggle. Human judgement remains essential, but it cannot function effectively if it is only inserted at isolated approval points. This is where the idea of programmable governance becomes important.</p><p>Rather than relying entirely on manual oversight, programmable governance embeds certain rules, constraints and escalation paths directly into the systems themselves. Authority boundaries can be checked automatically before actions occur. Certain thresholds can trigger human review. Conflicts between objectives can halt execution and escalate to decision-makers.</p><p>In other words, governance becomes <strong>structural rather than procedural</strong>. Some forms of accountability are handled automatically within the system, while human judgement is reserved for the decisions that genuinely require interpretation, trade-offs and values.</p><p>When governance is structured this way, the human role changes. Instead of reviewing every decision, leaders focus on calibrating how the system interprets risk and uncertainty. And this is where probabilistic thinking becomes essential.</p><p>Most governance processes still ask leaders to express judgement in binary terms: approved or rejected, compliant or non-compliant, acceptable or unacceptable.</p><p>Yet the judgments leaders actually hold are rarely so definite.</p><p>A CISO may believe there is a moderate chance that monitoring systems would fail under certain conditions. A product leader may suspect that a new feature introduces reputational risk without being able to quantify it precisely. Legal teams may sense regulatory ambiguity long before it becomes formal policy.</p><p>These kinds of judgements contain valuable information. But governance structures compress them into simple approvals or objections.</p><p>The idea of the <strong>probabilistic human</strong> offers a different approach.</p><p>Instead of presenting certainty where none exists, probabilistic humans express judgement in terms of likelihood and confidence. When those signals can be aggregated across many individuals, organisations gain a clearer picture of how risk is evolving over time.</p><p>And once judgement can be expressed probabilistically, a new set of governance techniques becomes possible, and we enable a far richer audit trail of decision-making that can be used to train AI systems and improve future decisions.</p><p>Once judgement can be expressed this way, a new set of governance techniques becomes possible and we create a richer audit trail of how judgement is exercised, which can inform both future governance decisions and the training of AI systems themselves.</p><h2>Where this tension shows up for leaders</h2><p>The limitations of human-in-the-loop governance rarely appear as an explicit design flaw. Instead, they surface indirectly, as uncertainty that feels difficult to resolve through existing oversight structures.</p><p>Different leadership roles can encounter this tension in different ways, depending on where they sit in the organisation&#8217;s decision and accountability landscape.</p><h3>CISO pain points: signals that arrive too late</h3><p>For CISOs and security leaders, the strain often appears as a timing problem: risk reviews take place and systems are assessed against known threat models. Yet concerns about AI behaviour can emerge gradually, often through operational signals rather than formal governance channels.</p><p>A model may drift slowly outside expected parameters, monitoring alerts begin to cluster in unusual ways, small anomalies appear that do not yet justify escalation, but suggest that the system&#8217;s behaviour is shifting.</p><p>Traditional governance frameworks expect risks to be identified and addressed at defined checkpoints. But many of the signals that matter most in AI environments are probabilistic rather than definitive.</p><p>Security teams therefore find themselves working with a growing set of partial signals, indicators that something may be wrong, without a clear threshold that justifies intervention.</p><h3>Legal and compliance pain points: decisions without certainty</h3><p>Legal and compliance leaders tend to experience the tension differently. Governance processes often require them to classify a system in categorical terms: compliant or non-compliant, acceptable or unacceptable. Yet many AI deployments sit in ambiguous territory, particularly when regulations are evolving or when systems operate across jurisdictions.</p><p>Legal teams frequently recognise emerging risks early. They may sense that a deployment could attract scrutiny, or that regulatory expectations are shifting in ways that are difficult to formalise.</p><p>However, governance structures typically force those insights into binary decisions. A deployment is either approved or blocked, even when the underlying judgement is far more nuanced.</p><p>This can create uncomfortable dynamics. Legal teams appear cautious or obstructive when they are simply responding to uncertainty that has not yet stabilised.</p><h3>Product leadership pain points: innovation slowed by rigid oversight</h3><p>Product leaders encounter the same structural issue from another direction. AI-enabled features often evolve iteratively. Teams test new capabilities, refine workflows, and adjust behaviour based on real-world feedback. In this environment, risk rarely presents itself as a clear go-or-no-go moment.</p><p>Instead, risk appears as shifting probabilities.</p><p>A feature may be broadly safe, but introduce edge-case failure modes. A system may work well under typical conditions, but become fragile when interacting with other services.</p><p>When governance frameworks rely on discrete approvals, product teams can find themselves navigating a process that feels mismatched to the way the technology evolves. Reviews occur at specific milestones, while risk emerges gradually over time.</p><p>The result is often friction rather than clarity.</p><h3>Executive leadership pain points: oversight that becomes symbolic</h3><p>At the executive level, the tension appears as a widening gap between formal oversight and operational reality.</p><p>Leadership teams approve governance frameworks, establish policies, and review risk dashboards. Yet the speed and complexity of AI systems can make those structures feel increasingly abstract.</p><p>Executives may receive polished summaries of model performance or compliance posture, while sensing that the organisation&#8217;s true exposure is harder to quantify.</p><p>This does not usually reflect a failure of diligence. Rather, it reflects a mismatch between the episodic rhythm of traditional governance and the continuous evolution of AI-enabled systems.</p><p>Under these conditions, leadership oversight risks becoming symbolic: reassuring in principle, but too distant from the system to provide real-time calibration.</p><p>Across all of these perspectives, the underlying issue is the same - human judgement is present throughout the governance system, but it is expressed in ways that hide uncertainty rather than revealing it. Leaders are often forced to present confidence where what they actually possess is a probability.</p><p>This is where collective intelligence techniques become valuable. They allow organisations to capture and aggregate the probabilistic judgement already present across the system, turning it into signals that governance structures can actually use.</p><p>Read on to explore three techniques that can help organisations embed structured collective intelligence into AI governance:</p><ul><li><p>prediction markets</p></li><li><p>structured dissent mechanisms / red team markets</p></li><li><p>probabilistic risk councils.</p></li></ul>
      <p>
          <a href="https://academy.shiftbase.info/p/programmable-governance-and-probabilistic">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Humans in the Loop or in the Soup?]]></title><description><![CDATA[How can we create governance systems that are quick and detailed enough for the AI era, but also maintain human-in-the-loop safeguards and accountability?]]></description><link>https://academy.shiftbase.info/p/humans-in-the-loop-or-in-the-soup</link><guid isPermaLink="false">https://academy.shiftbase.info/p/humans-in-the-loop-or-in-the-soup</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 03 Mar 2026 15:50:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!zVy7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Enterprise AI governance, security and safety are challenges that will require a multi-domain approach and imaginative solutions that combine technology, human factors, knowledge engineering and codification. These are issues that cannot just be delegated to CSOs and IT functions without collective leadership accountability.</p><h2>Waiter! There&#8217;s a Human in my Loop!</h2><p><strong><a href="https://www.strangeloopcanon.com/p/aligning-anthropic?utm_source=substack&amp;publication_id=233019&amp;post_id=189678330&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">The recent furore over the USA Department of War&#8217;s threats to declare Anthropic a supply chain risk</a></strong> is an interesting example of how confusing things are likely to become.</p><p>Anthropic has been supplying a modified (and apparently more advanced) version of Claude to the US military through Palantir, but <strong><a href="https://www.anthropic.com/news/statement-department-of-war">has tried to insist on two red lines governing its usage</a></strong>, namely (1) that it should not be used for broad spectrum domestic surveillance, which might be technically illegal; and, (2) it should not be used to run fully-autonomous weapons systems, because Anthropic do not believe it is yet reliable enough to do this safely.</p><p>In fact, <strong><a href="https://www.programmablemutter.com/p/ai-is-a-bureaucratic-technology-so">as Henry Farrell wrote today</a></strong>, the US military is such a vast bureaucracy that the majority of use cases for LLMs will not be about autonomous weapon systems, but the logistics, information synthesis and practical management tasks that such a huge organisation requires.</p><p>Nevertheless, the Department of War has insisted it should have total control over how the tool is used, and if Anthropic do not acquiesce, then the company will be declared a supply chain risk, meaning not only that it will lose its contracts with the US government, but also that third parties offering services to government that depend on Anthropic&#8217;s technology will probably need to replace it for another model.</p><p>By way of context, Israel - one of the most advanced users of AI-enabled technology for targeting and hardly what might be called over-cautious - claims to use a <strong><a href="https://giftarticle.ft.com/giftarticle/actions/redeem/32ed5396-07a1-4107-b8a6-09366630445b">double human-in-the-loop sign-off process</a></strong> to verify proposed targeted attacks. Meanwhile, the US &#8220;Secretary of War&#8221; rails against <em>&#8220;stupid rules of engagement&#8221;</em> and <em>&#8220;traditional allies who wring their hands and clutch their pearls, hemming and hawing about the use of force.&#8221;</em></p><p>Inevitably in such a polarised political landscape, <strong><a href="https://www.nytimes.com/2026/02/27/technology/anthropic-trump-pentagon-silicon-valley.html?campaign_id=158&amp;emc=edit_ot_20260302&amp;instance_id=171869&amp;nl=on-tech&amp;regi_id=71034279&amp;segment_id=216046&amp;user_id=b1568fb0a0ae5f0b3bc0e2c4e95a01d8">Anthropic have been seen as the good guys</a></strong> and OpenAI, who negotiated their own agreement with the DoW shortly after Anthropic&#8217;s deal collapsed, <strong><a href="https://www.nytimes.com/2026/02/27/technology/openai-agreement-pentagon-ai.html?campaign_id=158&amp;emc=edit_ot_20260302&amp;instance_id=171869&amp;nl=on-tech&amp;regi_id=71034279&amp;segment_id=216046&amp;user_id=b1568fb0a0ae5f0b3bc0e2c4e95a01d8">have been seen as the bad guys</a></strong>; but what this means for AI governance is likely to be less simple than it appears.</p><p>So where is AI governance headed and what are the practical alternatives to unlimited executive power?</p><h2>Security Soup and Programmable Governance</h2><p>I had a conversation with a very smart and accomplished CISO last week, and she suggested that we are heading to a place where human oversight is insufficient to maintain security in a complex enterprise. So, whilst we might want to see human-in-the-loop solutions to minimise AI risks, we probably need to think more imaginatively about how this works.</p><p>Historically, regulation and corporate governance involved a very slow process of risk analysis and political negotiation of the rules, which were then handed down to company officers to enforce manually through training, guidelines and so on. But this approach was prone to problems such as <strong><a href="https://en.wikipedia.org/wiki/Regulatory_capture">regulatory capture</a></strong>, or compliance theatre, where those in charge of upholding the rules lacked the power and influence to rein in the behaviour of colleagues who were generating profits from skirting them (e.g. in banking). This was also prone to regulatory inertia, which often meant regulators were too busy fighting the last war and struggled to keep up with current or emerging risks.</p><p>So how should companies do to ensure security, safety and risk mitigation in their use of AI and related technologies?</p><p><strong><a href="https://diginomica.com/how-uk-cios-are-governing-ai-without-killing-innovation-banking-retail-and-academia-perspectives">Diginomica recently reported on a UK CIO event</a></strong> where executives expressed a desire to ensure AI safety without harming innovation, and several attendees talked more about a collaborative evolution than top-down transformation:</p><blockquote><p><em>Anybody in a complex organization should think about whether they are enabling or leading. You enable through a coalition of the willing.</em></p></blockquote><p>There have been several attempts to scope out some guidelines for trustworthy AI. For example, <strong><a href="https://www.techtarget.com/searchcustomerexperience/feature/4-governance-pressures-shaping-enterprise-AI#">James Miller at TechTarget recently shared some thoughts on how to boost institutional accountability</a></strong> to align governance and structure, and laid out a framework to guide the shift from strategy to execution via trustworthy data and algorithms.</p><p>Elsewhere, those who want to see pro-human, safe and trustworthy AI are <strong><a href="https://partnershiponai.org/work/">writing lots of guidelines, pleas and commendable nice words</a></strong>. But they often lack a credible, practical approach to implementation that is sufficiently robust, responsive and integrated with technology systems and platforms to really make a difference.</p><p><strong><a href="https://www.reddit.com/r/ArtificialInteligence/comments/1r87afj/we_didnt_know_what_we_didnt_know_standing_up/">As one enterprise AI practitioner put it when reflecting on lessons learned</a></strong>:</p><blockquote><p><em>Policy alone won&#8217;t save you. You need policy and technology and education working together. Any one of those by itself is insufficient.</em></p></blockquote><p>Looking ahead, I think we are headed towards AI-enabled <em>programmable governance</em> (building on existing practices such as <strong><a href="https://www.paloaltonetworks.com/cyberpedia/what-is-policy-as-code">Policy-as-Code</a></strong>) where an organisation ingests and codifies regulatory, compliance and legal rulesets, and combines them with its own rules of the road to give governance systems the context needed to guide the organisation&#8217;s work. And of course, governance isn&#8217;t just about rules like <em>&#8220;don&#8217;t leak data.&#8221;</em> It also includes less technical concerns like <em>&#8220;don&#8217;t violate brand values,&#8221;</em> <em>&#8220;don&#8217;t hallucinate pricing,&#8221;</em> and <em>&#8220;stay within budget.&#8221;</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zVy7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zVy7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 424w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 848w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 1272w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zVy7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic" width="1006" height="1416" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1416,&quot;width&quot;:1006,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:217239,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/189777433?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zVy7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 424w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 848w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 1272w, https://substackcdn.com/image/fetch/$s_!zVy7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc21d7a86-669c-4324-a1c3-86f603712ab9_1006x1416.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Where Leaders Need to Step up</h2><p>The challenge in building and evolving fit-for-purpose AI security and governance systems is not to be under-estimated, and it covers everything from tech, data, and platforms to culture, behaviour, and process management.</p><p>Manual oversight is not enough. Human-in-the-loop is desirable, but at what level and supported by what kind of deep infrastructure underneath? Perhaps we will end up with agents surveilling other agents, and an organisational autonomic immune system that draws lessons from biology with single-purpose nano-bots running around in swarms to identify and contain anomalies or &#8216;foreign bodies&#8217; at the network level.</p><p>But however our security, safety and governance systems evolve, they will need significantly greater codification of our rules, guidelines and threat vector identification than we have today. This is an area where the wider leadership function can meaningfully assist CISOs and CIOs without needing a great deal of technical knowledge: start by capturing specific rules or statements that make nice words actionable and (ideally) programmable.</p><p>It is a system challenge, but we have barely scratched the surface of writing the code for it.</p><p><strong><a href="https://www.inc.com/umair-javed/leadership-behavior-is-causing-an-ai-adoption-gap/91307376">We already know that when leaders only decide and delegate on AI topics, but don&#8217;t lead adoption, the outcome is sub-optimal</a></strong>. We need them to get their hands dirty and bring their experience and knowledge of their organisation&#8217;s value chain and strategy to the table. We already know that <strong><a href="https://giftarticle.ft.com/giftarticle/actions/redeem/88c1e6b5-941a-492d-b7aa-bb8de54effcc">mandating adoption using the crudest possible KPIs</a></strong> is likely to optimise for the wrong outcomes.</p><p>We need leaders to lead the codification or knowledge engineering required to make our guidelines understandable to agents, APIs and systems.</p><p>This starts with <strong><a href="https://medium.com/@andreas_g/knowledge-graphs-vs-context-graphs-the-missing-foundation-of-reliable-enterprise-ai-75e3cac6d047">understanding and connecting the knowledge objects of the organisation in shared knowledge graphs</a></strong>, so that the entities we work with (person, team, process, system, concept, goal, etc) are addressable and findable. Once we have a good basic knowledge graph, then we can start to make a system that is programmable (IF this THEN that, etc) - and that is where we can start to develop meaningful codification of how we want things to work.</p><p><strong><a href="https://onlydeadfish.substack.com/p/fish-food-679-organisational-knowledge?utm_source=substack&amp;publication_id=2195351&amp;post_id=187111193&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Neil Perkins recently shared a thoughtful article on knowledge engineering</a></strong> and why it matters, which is worth a few minutes to read:</p><blockquote><p><em>&#8230; I think the idea of architecting knowledge for AI goes far deeper than just a technical practice. The quality of every AI-assisted decision, recommendation, and output is bounded by the quality of context it receives. This makes the curation of organisational knowledge (what gets captured, how it&#8217;s structured, how relationships between ideas are maintained) a fundamental strategic capability.</em></p></blockquote><h2>Humans in the Loop or Making the Soup? Why not Both?</h2><p>To suggest that we are blowing past the point where humans in the loop can meaningfully oversee and manage safety, security and regulatory compliance need not be a defeatist or alarmist position. It is about the size and scale of the loops, and where people can maximise the value of their unique intuition and experience.</p><p>If the loops are too low-level, there is too much information to process unless we develop the kind of pre-cognition powers in a movie like Minority Report. If the loops are too high-level, then we review our leadership team&#8217;s nicely packaged monthly threats powerpoint file only to find we were fatally compromised 29 days ago.</p><p><strong><a href="https://www.kasava.dev/blog/ai-as-exoskeleton">If we think of enterprise AI as an exoskeleton that empowers people rather than a robot that replaces them</a></strong>, perhaps we can use agentic systems where they can realistically perform autonomic immune functions, whilst surfacing issues and decision points at the right level of detail for human-in-the-loop issues, allowing people to drill down quickly to see the specifics. But this will require a whole load of systems, data and knowledge graphs to support them.</p><p>Leading and participating in the codification effort is something all leaders can do today to take the pressure of CSOs, CISOs and CIOs who cannot do it alone.</p>]]></content:encoded></item><item><title><![CDATA[Extraction vs. Redesign: The Hidden Fork in the Road for AI Leaders]]></title><description><![CDATA[Are we using AI to squeeze more output from yesterday&#8217;s structures, or to redesign the architecture of how value is created?]]></description><link>https://academy.shiftbase.info/p/extraction-vs-redesign-the-hidden</link><guid isPermaLink="false">https://academy.shiftbase.info/p/extraction-vs-redesign-the-hidden</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 24 Feb 2026 15:20:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9Mon!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="pullquote"><p>This article is published as a free sample of the Shift*Academy paid edition.</p><p>Every other week, the paid edition explores the structural implications of AI for leadership, organisational design and enterprise capability, with practical deep dives for leaders navigating the agentic era.</p><p>If this resonates, you can subscribe for full access to future essays and capability breakdowns.</p></div><p>Over the past decade, we have lived through multiple &#8220;transformations&#8221; that promised structural change. Digital was meant to flatten hierarchies. Agile was meant to empower teams. Platforms were meant to dissolve silos. In each case, the tooling evolved faster than the governing logic of the organisation. We digitised reporting lines rather than redesigning them. We accelerated information flow without rethinking who holds authority. We changed the vocabulary, but not the topology.</p><p>Enterprise AI is arriving into that same landscape.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://academy.shiftbase.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Shift*Academy is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Its capabilities are extraordinary - coordination costs are falling, translation layers can be automated, and expertise can gain direct leverage over execution in ways that were impossible even five years ago. And yet, already, we can see a familiar pattern forming. Many organisations are reaching for AI as a simple efficiency play inside structures designed for a previous era.</p><p>The question is not whether AI works - it clearly does - but whether we are willing to change the frame through which we organise work, rather than using this new intelligence to reinforce the old machine.</p><p>And this is where the fork in the road begins to come into focus.</p><p>On paper, most AI deployments look similar. The same models. The same copilots. The same orchestration platforms layered into finance, operations, customer service, product and strategy.</p><p>The language is shared: augmentation, automation, leverage, productivity. But beneath that shared surface, two very different operating logics are taking shape.</p><p>In some organisations, AI is being introduced as a margin stabiliser. Junior layers are reduced. Reporting structures remain intact. Agents are embedded into existing workflows to accelerate output and reduce cost, while decision rights and authority models remain largely untouched. Efficiency improves, executive distance from execution is preserved and the machine runs faster.</p><p>In more ambitious organisations, AI is treated as permission to ask more uncomfortable questions. If translation can be automated, why maintain translation layers? If coordination is cheaper, why preserve reporting ladders designed to aggregate information upward? If expertise can now act closer to the work, what becomes of authority that was historically justified by information asymmetry? Here, AI is not simply accelerating existing processes; it is exposing structural assumptions that have long gone unchallenged.</p><p>Both paths can produce impressive short-term productivity gains. But only one opens up net new top line growth, and it does this by changing the organisation.</p><p>The distinction is subtle at first. It does not show up in vendor announcements or pilot metrics. It shows up in what leaders choose to leave alone. It shows up in whether span of control is redesigned or simply expanded. It shows up in whether apprenticeship is re-imagined or quietly eroded. It shows up in whether authority is redistributed toward outcomes, or insulated behind more efficient reporting.</p><p>This is the moment where AI adoption stops being a technology story and becomes a design story.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9Mon!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9Mon!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9Mon!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:315230,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/189019819?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9Mon!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!9Mon!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1061a30f-f6c7-4c08-b328-dc7c6f890f83_1536x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Extraction Mode: The Frame Defending Itself</h2><p>Extraction Mode often presents as pragmatism. Budgets are under pressure, markets are volatile and boards are demanding visible returns. AI offers immediate gains in efficiency, speed, and headcount flexibility. In that context, embedding agents inside existing workflows feels not only rational, but responsible.</p><p>Junior roles are reduced first:</p><ul><li><p>automation absorbs repetitive tasks;</p></li><li><p>reporting structures remain largely intact</p></li><li><p>middle layers translate agent outputs rather than reconsider their necessity</p></li></ul><p>Executive oversight becomes more data-rich, but not structurally closer to execution. Productivity per employee rises and importantly, cost curves improve.</p><p>But decision rights are still flowing upward and authority is still tied to hierarchy rather than outcome ownership. Over time, the consequences begin to surface in less obvious ways. For example, apprenticeship pathways narrow as entry-level roles disappear without being redesigned; or leaders find that their bandwidth does not materially recover, but shifts toward adjudicating edge cases and resolving boundary disputes between humans and agents. Informal shadow coordination grows as teams compensate for ambiguities that the formal structures never addressed.</p><p>Extraction Mode can produce good numbers in the short term. It can stabilise margins and extend runway. But it does so by reinforcing the underlying frame: preserving hierarchy, protecting authority and optimising cost. AI becomes a margin machine. And the structure that limited previous transformations remains quietly in place.</p><h2>Redesign Mode: Questioning the Topology</h2><p>Redesign Mode begins with a different instinct. Instead of asking, &#8220;Where can AI remove cost?&#8221; it asks, &#8220;What assumptions about structure that were built into the organisation when coordination was expensive are no longer valid?&#8221;</p><p>If translation can be automated, then layers that existed primarily to aggregate and repackage information should be scrutinised. If agents can monitor workflows continuously, then escalation does not need to rely on proximity to authority. If expertise can act directly with the support of agent systems, then the justification for distance between decision-makers and execution begins to weaken.</p><p>In Redesign Mode, AI is not inserted into the existing machine, instead it is used to reveal its architecture, and then improve it.</p><p>Reporting ladders are examined, not just accelerated. Decision rights are clarified, not assumed. Span of control is redesigned deliberately rather than allowed to expand silently. Outcome boundaries are defined explicitly, and authority is tied to those boundaries rather than to position in a chain of command. This does not necessarily mean &#8220;flatter.&#8221; It means clearer.</p><p>Some functions may consolidate. Others may fragment into outcome cells with explicit guardrails and escalation rules. Leaders move closer to the work in some areas and further from it in others, but the movement is intentional. Apprenticeship is redesigned alongside automation, ensuring that the disappearance of repetitive tasks does not quietly eliminate the pathways through which judgment develops.</p><p>The shift is subtle - AI is treated not as an efficiency layer but as structural permission. Coordination is cheaper; therefore, the organisation does not have to be shaped the way it was when coordination was scarce.</p><p>This path is slower. It exposes leaders to greater short-term uncertainty. It requires confronting incentive systems, governance habits, and career ladders that feel natural because they have been stable for decades - but it also changes the trajectory.</p><p>AI becomes a leverage multiplier rather than a margin machine. And the organisation begins to evolve rather than simply accelerate.</p><h2>Why Extraction Mode Is the Default</h2><p>Redesign Mode sounds compelling in theory. Few leadership teams would openly argue that preserving outdated structures is the goal. And yet, when AI initiatives move from pilot to budget to restructuring, most organisations tilt toward extraction.</p><p>Cost reduction is measurable. Structural redesign is not. A headcount number can be reported to the board next quarter. A re-architected decision-rights model compounds over years. The former is easy to defend; the latter is harder to explain.</p><p>Governance models amplify this bias. Boards understand margin expansion. They are less fluent in organisational topology. Asking for approval to remove redundant roles inside an existing structure feels prudent. Asking to redesign that structure altogether feels risky. It introduces ambiguity about authority, reporting, and risk allocation at precisely the moment when AI already feels destabilising.</p><p>But the explanation cannot stop there. Over the past decade, most transformation effort has been directed downwards. Teams were asked to become more agile, managers were asked to embrace digital tools, frontline functions were reconfigured. Senior leaders, in many cases, changed less. Their ways of working and their information flows often remained intact. Digital transformation pointed at the base of the pyramid more than at its apex.</p><p>Our current AI transformation is exposing this. As information asymmetries fall and translation layers become automatable, the traditional justification for distance from execution weakens. Redesign Mode would require leaders to update their own operating models: to move closer to outcome boundaries, to make judgment legible, to relinquish some insulation provided by hierarchy. That is harder than reducing cost!</p><p>Preserving hierarchy is safer than questioning it. Leaders who reduce spend inside a known model are seen as disciplined, whereas leaders who challenge the shape of that model take on visible personal risk. In uncertain markets, prudence (and self-preservation), often win.</p><p>There is also a subtler force at work. Most organisations have been optimised for decades around information asymmetry. Authority was justified by access: access to data, to strategic perspective. AI reduces that asymmetry, but the habits built around it remain. It is easier to automate the flow of information up the ladder than to question why the ladder exists in its current form.</p><p>This is how transformations stall. The technology advances and the structure remains recognisable as efficiency rises. The deeper architecture stays intact. And over time, what could have been a redesign moment becomes another optimisation cycle that fails to grasp the opportunity for improvement.</p><h2>The Apprenticeship Question</h2><p>Every hierarchy contains an implicit learning pathway. Entry-level roles absorb repetitive work. They sit close to process and they observe decisions being made. They develop judgment slowly through exposure, error, and proximity. Over time, some of those individuals move upward, carrying tacit knowledge with them.</p><p>It is not a perfect system. It can be inefficient and uneven. But it is a capability engine, which Extraction Mode fundamentally disrupts and fails to replace with a workable solution.</p><p>When junior layers are removed without structural redesign, repetitive tasks disappear, but so do many of the early exposure points where judgment is formed. Automation replaces execution without replacing apprenticeship. The pyramid thins, but the pipeline narrows.</p><p>In the short term, this looks efficient. Output per employee increases. Overhead falls. But over time, a different cost accumulates.</p><ul><li><p>Where do future leaders learn how decisions are made under uncertainty?</p></li><li><p>Where does tacit operational knowledge accumulate?</p></li><li><p>How does strategic judgment develop if the early rungs of the ladder vanish?</p></li></ul><p>Redesign Mode confronts this directly. If AI removes certain forms of work, then the learning architecture must be rebuilt intentionally. Apprenticeship shifts from &#8220;do the repetitive work and observe&#8221; to something more deliberate: structured exposure to decision boundaries, transparent escalation logic, visible agent&#8211;human coordination, and explicit responsibility for outcomes.</p><p>In other words, if coordination is becoming cheaper, learning cannot remain accidental.</p><p>This is not a sentimental argument for preserving junior roles. It is a compounding argument. Organisations that treat entry-level work purely as cost will eventually erode their own capacity for long-term adaptation. Those that redesign learning alongside automation build a deeper form of resilience.</p><h2>If You&#8217;re Serious About Redesign Mode</h2><p>Redesign Mode is not declared in strategy decks. It shows up in structural edits.</p><p>If you believe AI is a redesign moment rather than a margin moment, there are early signals that distinguish intent from rhetoric.</p><h3>1. Rewrite One Decision Rights Map</h3><p>Pick a domain where agents are already active.</p><p>Then ask:</p><ul><li><p>Which decisions remain human?</p></li><li><p>Which are delegated?</p></li><li><p>What triggers escalation?</p></li><li><p>Who arbitrates conflict?</p></li></ul><p>If the map still routes most meaningful decisions upward through the same hierarchy, you are in Extraction Mode.</p><h3>2. Audit One Reporting Layer for Translation vs Judgment</h3><p>Many layers exist to aggregate and translate information.</p><p>Agents can now perform much of that work.</p><p>For one reporting tier, ask:</p><ul><li><p>Does this layer exercise unique judgment?</p></li><li><p>Or does it primarily synthesise and repackage?</p></li></ul><p>If it is translation, move it to the system.</p><p>If it is judgment, clarify and anchor it closer to outcomes.</p><h3>3. Redesign One Apprenticeship Pathway Alongside Automation</h3><p>If repetitive tasks disappear, learning cannot remain accidental.</p><p>In one function:</p><ul><li><p>Map how junior staff historically developed judgment.</p></li><li><p>Identify what automation removes.</p></li><li><p>Design deliberate exposure to decision boundaries, trade-offs, and escalation logic.</p></li></ul><p>If you cut entry roles without rebuilding learning architecture, you are optimising cost at the expense of future capability.</p><h3>4. Define One Outcome Cell</h3><p>Choose one cross-functional workflow.</p><p>Define:</p><ul><li><p>The outcome metric.</p></li><li><p>The guardrails.</p></li><li><p>The escalation rules.</p></li><li><p>The named human owner.</p></li><li><p>The supporting agent stack.</p></li></ul><p>If coordination is cheaper, structure can follow outcomes rather than reporting ladders.</p><p>These are not large-scale reorganisations, they are diagnostic edits - small structural moves that reveal whether AI is being used to reinforce the current topology or to reshape it. Redesign Mode begins with the courage to make authority, learning, and accountability explicit.</p><h2>The Choice Is Ours</h2><p>AI will increase productivity in either mode.</p><p>In Extraction Mode, it will accelerate reporting, reduce cost, and preserve existing authority structures with greater efficiency. The machine will run faster, but it will break more often. Margins may expand. Headcount curves may improve. On paper, it will look like progress.</p><p>In Redesign Mode, AI will be treated as a structural inflection point. Coordination costs will fall and structure will change in response. Decision rights will be clarified. Span will be redesigned. Apprenticeship will be rebuilt. Authority will move closer to outcomes rather than further from them.</p><p>The models themselves are neutral but what they amplify is not. If we embed AI inside hierarchies designed for information scarcity and expensive coordination layers, we will simply automate those hierarchies. We will thin the pyramid without questioning its shape. We will accelerate the system that previous transformations failed to meaningfully change.</p><p>If instead we allow AI to expose the assumptions built into our structures, then we have a redesign opportunity rather than yet another optimisation cycle.</p><p>The uncomfortable truth is that the technology is not the constraint. Model capability is advancing rapidly. What will determine whether this transformation compounds advantage or quietly stalls is whether leaders are willing to question the frame that has shaped enterprise design for decades.</p><p>AI does not choose between extraction and redesign.</p><p>We do.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://academy.shiftbase.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Shift*Academy is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Schrödinger’s Optimism: AI and Productivity Signals]]></title><description><![CDATA[Should we celebrate a small bump in output with lower headcount, or can we lift our eyes to the remarkable opportunity to cultivate extraordinary business capabilities?]]></description><link>https://academy.shiftbase.info/p/schrodingers-optimism-ai-and-productivity</link><guid isPermaLink="false">https://academy.shiftbase.info/p/schrodingers-optimism-ai-and-productivity</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 17 Feb 2026 15:45:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7VEM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Schr&#246;dinger&#8217;s Optimism</h2><p>Reading news stories about the US stock market dip at the end of last week, you might think that serious economic and technology analysts are uncertain about the impact of AI on business and productivity.</p><p>Selling or buying stocks is quite a binary activity (notwithstanding the grey areas of hedging and options), but the current state of AI is more quantum than binary - simultaneously beyond imagination and yet not good enough for deployment in production; able to autonomously code entire apps with a few lines of instruction, but struggling with basic maths or questions like <strong><a href="https://www.reddit.com/r/OpenAI/comments/1r4zj8a/walk_to_wash_car_logical_fallacy/">&#8220;should I drive to the car wash?&#8221;</a></strong></p><p>We will probably need to live with its patchy, jagged, probabilistic, non-evenly-distributed nature for some time, and co-evolve our methods with the technology in much the same way as quantum computing relies on error correction. The question now is how quickly our institutions can metabolise what the technology is already making possible.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7VEM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7VEM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7VEM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:158718,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/188269840?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7VEM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!7VEM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6394c50-6d1c-4f3c-bbc1-b102465a363e_1024x1024.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But if you zoom out just a little, the progress being made is incredible, and would have been unimaginable just a few years ago.</p><p><strong><a href="https://shumer.dev/something-big-is-happening">Matt Shumer recently wrote a widely-shared article trying to put this progress into words</a></strong>, which started with the reflection that people who are not following AI developments don&#8217;t know quite how disruptive the next 5 years could be:</p><blockquote><p><em>I&#8217;ve spent six years building an AI startup and investing in the space. I live in this world. And I&#8217;m writing this for the people in my life who don&#8217;t... my family, my friends, the people I care about who keep asking me &#8220;so what&#8217;s the deal with AI?&#8221; and getting an answer that doesn&#8217;t do justice to what&#8217;s actually happening. I keep giving them the polite version. The cocktail-party version. Because the honest version sounds like I&#8217;ve lost my mind. And for a while, I told myself that was a good enough reason to keep what&#8217;s truly happening to myself. But the gap between what I&#8217;ve been saying and what is actually happening has gotten far too big. The people I care about deserve to hear what is coming, even if it sounds crazy.</em></p></blockquote><p>He goes on to discuss what this means for jobs, extrapolating from his own experience as a software developer working in AI, and it is simultaneously an exciting and also very discomforting read. But it is also naive.</p><p>On the optimistic side, I believe our lives should not be dictated by &#8220;jobs&#8221;, which have become hollowed out and insufficiently remunerated to live well at the entry level to mid-tier (at least outside of tech). But realistically, in the absence of labour market improvement or Universal Basic Income (UBI) or any other idea about how young people without assets can support themselves, the implications of what Shumer predicts could be very worrying.</p><p>However, business and societal change is modulated by incredible reserves of inertia that can hold back progress for decades, if not centuries, as long as enough <s>Powerpoint enjoyers</s> leaders are invested in the old ways of doing things.</p><p>A case in point is the debate about SaaS platforms in business:</p><ul><li><p><em>Logically</em>, many of them are screwed.</p></li><li><p><em>Practically</em>, firms can now recreate better, simpler versions of them without the eye-watering subscription costs using coding agents.</p></li><li><p><em>Emotionally</em>, they weigh so heavily on employee experience that companies would be far happier if they ceased to exist.</p></li></ul><p>And yet &#8230; don&#8217;t count them out. There are some very human - and very illogical - reasons on both the buyer side and the vendor side that suggest <strong><a href="https://x.com/finbarr/status/2021999185172775288">these businesses might not be so easy to kill, as Finbarr Taylor argues here</a></strong>. Just because a better way is possible, it doesn&#8217;t mean it will come to pass:</p><blockquote><p><em>You don&#8217;t always pick the cheapest option. You don&#8217;t always pick the most innovative option. You pick the option that, if it fails, you can defend to your boss. &#8220;We went with Salesforce&#8221; is a defensible sentence in any boardroom in America. &#8220;We went with an app I vibe-coded over the weekend&#8221; is a resignation letter.</em></p><p><em>This is the same dynamic that kept IBM dominant for decades and that keeps McKinsey and Deloitte in business despite armies of cheaper, often smarter competitors. Enterprise buyers optimize for career risk, not unit cost. They want a vendor that will still exist in three years, that has a support team they can call at 2am, that has a track record of not losing their data.</em></p></blockquote><p>Change is not inevitable. At least not everywhere.</p><h2>Exponential Proponents</h2><p>This weekend, <strong><a href="https://www.exponentialview.co/p/the-hundred-million-token-day">Azeem Azhar also published an eye-popping piece about the speed of AI&#8217;s evolution</a></strong>, predicated on the realisation that he had consumed 97 million tokens in a single day of working with AI tools. He makes the point that in exponential change, each level of scale can be fundamentally different from those below them. At 10^3 tokens, AI is a toy. But as we add a zero to the tokens used, it becomes a tool, then a colleague, a workflow, a process and a workforce; but at 10^9 tokens, it is more like infrastructure - always on, always working, like electricity.</p><blockquote><p><em>At 10&#8313;, a billion tokens a day per person, our unit of analysis changes. This becomes agents spawning even more sub-agents and talking to them and other agents. The human sets direction and adjudicates edge cases, but the conversation is mostly not ours anymore.</em></p><p><em>I&#8217;ve already caught a glimpse with my own setup. Micromanaging them slows down the whole process. If you had to configure each sub-agent yourself and track their work, I&#8217;m pretty certain none of us would do it. In other words, the bottleneck is no longer the model&#8217;s capability; it&#8217;s your willingness to let go. It becomes like running an organisation, trusting the parts to make the whole.</em></p></blockquote><p>And for those at the frontier of AI usage, the tools are not reducing effort, or giving them extra leisure time; in fact they are intensifying their work, <a href="https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies-it">as </a><strong><a href="https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies-it">Aruna Ranganathan and Xingqi Maggie Ye found in their recently published eight-month HBR study of 200 workers at a hi-tech firm</a></strong>, which raises some interesting questions for leaders:</p><blockquote><p><em>The promise of generative AI lies not only in what it can do for work, but in how thoughtfully it is integrated into the daily rhythm. Our findings suggest that without intention, AI makes it easier to do more&#8212;but harder to stop. An AI practice offers a counterbalance: a way to preserve moments for recovery and reflection even as work accelerates. The question facing organizations is not whether AI will change work, but whether they will actively shape that change&#8212;or let it quietly shape them.</em></p></blockquote><h2>Which way up is that J-curve?</h2><p>In the FT this weekend, <strong><a href="https://www.ft.com/content/4b51d0b4-bbfe-4f05-b50a-1d485d419dc5">Erik Brynjolfsson made the case that AI-attributed productivity improvements are starting to show up in the data</a></strong> (also <strong><a href="https://geekway.substack.com/p/ai-driven-productivity-growth-is">commented on by Andrew McAfee here</a></strong> if the FT link is paywalled):</p><blockquote><p><em>Data released this week offers a striking corrective to the narrative that AI has yet to have an impact on the US economy as a whole. While initial reports suggested a year of steady labour expansion in the US, the new figures reveal that total payroll growth was revised downward by approximately 403,000 jobs. Crucially, this downward revision occurred while real GDP remained robust, including a 3.7 per cent growth rate in the fourth quarter. This decoupling &#8212; maintaining high output with significantly lower labour input &#8212; is the hallmark of productivity growth.</em></p><p><em>My own updated analysis suggests a US productivity increase of roughly 2.7 per cent for 2025. This is a near doubling from the sluggish 1.4 per cent annual average that characterised the past decade.</em></p></blockquote><p>Does this represent the beginning of the hoped-for productivity J-curve promised by AI optimists? Or are we seeing business leaders using automation to shed jobs, whilst protecting their own, with no reduction in overall output? Or, is the mild increase in US GDP nothing to do with technology at all, and could negative payroll growth indicate recessionary dynamics down the line? We will see.</p><p>It is sad to see leaders of large, established organisations respond to abundant technological capability by cutting the junior headcount (a.k.a their future) just to appease the fickle stock-trading gods in a time of market volatility, or to protect themselves until they can exit. If ever there was a time for long-term thinking about organisational development, it is now. Maybe private companies and those owned by long-term family trusts will be among those to chart a path through this fear and end up as winners.</p><p>But for individuals trying to get by in this liminal space between old and new worlds, it could be challenging. The most empowered, high-agency individuals and teams can achieve more than ever, but many of the entry-level jobs young people have been conveyed towards since they started school may not exist (or at least in such numbers) in the near future. If you have agents to manage, you might make it. But outside tech, old-fashioned management structures demand an awful lot of pointless busy work at the base of the pyramid that might start to be replaced sooner than we think.</p><p>We need to focus on helping the best leaders move closer to the work, not retreat further into abstraction and politics. AI makes it possible to compress layers, to give experienced people direct leverage over real outcomes rather than managing proxies and reports. But that only happens if incentives shift. In many firms, status is still measured by distance from execution, and risk is minimised by preserving familiar structures. Unless those dynamics change, AI will be used to thin out the base of the pyramid while leaving its shape intact. That would generate marginal gains at best, but reduce the organisation&#8217;s capacity to explore and exploit the kind of exponential gains that Matt Shumer and Azeem Azhar believe are possible.</p>]]></content:encoded></item><item><title><![CDATA[From Agent Spaghetti to Outcome Architecture]]></title><description><![CDATA[A practical guide to building composable, accountable agentic systems that scale.]]></description><link>https://academy.shiftbase.info/p/from-agent-spaghetti-to-outcome-architecture</link><guid isPermaLink="false">https://academy.shiftbase.info/p/from-agent-spaghetti-to-outcome-architecture</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 10 Feb 2026 15:03:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2SkC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The first wave of AI agents gave us chatter, smart assistants and co-pilots, thinly wrapped around language models. They could respond to prompts, but lacked memory, structure, or real autonomy. The second wave gave us experimental multi-agent frameworks promising goal-directed collaboration, but often resulting in brittle workflows, unclear ownership, and what can only be described as agent spaghetti.</p><p>A third pattern is beginning to take shape that shifts the focus away from agents completing isolated tasks, and towards composed systems that can reliably deliver outcomes - what we call an Outcome-as-Agentic Solution (OaAS).</p><p>Instead of delegating disconnected tasks to individual agents, teams define a measurable business outcome such as <em>&#8220;reduce time-to-value for onboarding,&#8221;</em> <em>&#8220;respond to regulatory change within 48 hours,&#8221;</em> or <em>&#8220;restore service after incident X&#8221;</em>, and assemble a lightweight system of agentic and human functions to achieve it. These systems combine modular agents, context-aware orchestration, embedded controls, defined human roles, and real-time feedback loops.</p><p>The goal isn&#8217;t full autonomy (yet). It&#8217;s programmable delivery, building a system that can act, adapt, and escalate when needed, with traceable logic and accountable performance.</p><p>Most organisations today are stuck in the second wave, over-automated in places, under-coordinated in others, and missing a coherent architecture. But those willing to invest in capability design rather than scattered tools have a chance to build something scalable and auditable.</p><p>This shift from agent spaghetti to outcome architecture isn&#8217;t just technical - it&#8217;s structural, and it&#8217;s coming fast.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2SkC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2SkC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2SkC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:505742,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/187508854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2SkC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!2SkC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0219dbcd-e6b3-46ec-9713-fe971b1cc352_1536x1024.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Why This Matters Now</h2><p>Agentic AI is moving fast, but many early implementations rely on brittle prompt chains, superficial integrations, or agents that struggle outside of sandboxed conditions.</p><p>But in enterprise settings where there are defined outcomes, clear constraints, and some history of process discipline, something more promising is starting to take shape. Teams are no longer just using agents to automate individual actions. They are beginning to build systems that can move toward outcomes, not just execute steps.</p><p>This shift matters because the old model of delivery is starting to break down. And many processes still rely on invisible glue: Slack messages, heroic effort, unspoken expectations, and the judgment of people who hold it all together.</p><p>Agentic systems, when designed well, offer a way to relieve that pressure. Instead of asking people to follow brittle processes or fill the gaps manually, organisations can compose systems that are able to coordinate, respond, and escalate in service of a shared result.</p><p>This only works if the outcome is clearly defined, the logic is well structured, and someone is accountable. And that is where things often fall apart.</p><p>Many leaders are not used to thinking in outcomes. They operate in terms of deliverables, metrics, or KPIs, but struggle to describe the intended result in a way that a system could act on. The shift from task delegation to outcome delegation sounds simple, but it reveals all the places where organisations rely on unspoken knowledge and manual intervention. Most agentic prototypes fail here, not because the tools are wrong, but because the intent is vague or the system has no way to recover when things go off track.</p><p>At the same time, teams are under growing pressure to move faster and cover more ground with fewer resources. Delegating outcomes to agents may feel risky, but in many cases it is less risky than continuing to scale human coordination with no real structure behind it.</p><p>The opportunity is real. So are the challenges.</p><p>Organisations that take capability design seriously will be better placed to make the shift, while those that remain stuck in automation theatre may soon find themselves outpaced by something quieter and more durable.</p><h2>What Makes an Outcome-as-Agentic Solution Work?</h2><p>The idea of delegating outcomes to AI systems often sounds appealing in principle, but quickly becomes uncomfortable in practice. Once an outcome is specified, the gaps begin to show: unclear logic, patchy data, fragile handoffs, and an over-reliance on people to notice when something goes wrong.</p><p>Most agentic systems still depend on human scaffolding that no one has time to maintain. To build something that can actually deliver, we need to design for outcomes from the start, with a focus on architecture, not just interaction.</p><p>From early deployments and internal pilots, seven core components appear necessary. These don&#8217;t guarantee success, but without them, outcome delivery is unlikely to scale or survive contact with the real world.</p><ol><li><p><strong>Defined Outcome:</strong> A specific, measurable result framed in business terms and bounded by time, value, or risk. It must be clear enough to guide action and structured enough to delegate.</p></li><li><p><strong>Coordinated Agentic Functions:</strong> Sensing, planning, action, escalation, and communication roles composed into a system. Modular or shared, these agents must work together toward the same end.</p></li><li><p><strong>Embedded Context Logic:</strong> Rules, reasoning, or orchestration layers that help the system adapt based on actors, history, and constraints. Without context-awareness, agent behaviour becomes brittle.</p></li><li><p><strong>Intentional Human Involvement:</strong> Designed-in roles for judgment, validation, or escalation. Human input is not a fallback but a core part of safe, ethical, and effective delivery.</p></li><li><p><strong>Built-In Controls:</strong> Guardrails, permissions, audit trails, and constraints that enforce responsible action and ensure enterprise-grade oversight from day one.</p></li><li><p><strong>Live Feedback Signals:</strong> Real-time data on performance, interaction, and progress. These signals allow the system to detect drift, adjust behaviour, and stay aligned with outcomes.</p></li><li><p><strong>Outcome Accountability:</strong> A clearly assigned owner, human, agent, or hybrid, who holds responsibility for the result and has the authority to monitor, intervene, or evolve the system.</p></li></ol><p>These seven components form the minimum viable architecture for moving from automated actions to accountable outcomes. Most organisations have some pieces already, the opportunity is to assemble them deliberately.</p><h2>Example Applications</h2><p>Outcome-as-Agentic Solutions can be applied wherever a clear business result is needed, but the delivery path involves multiple actors, moving parts, and changing conditions. Rather than replacing existing systems, OaAS compositions work best when they operate alongside or across existing functions, helping bridge the gap between intent and execution.</p><p>Here are a few early applications where this capability could be shaped:</p><h3><strong>Customer Onboarding Acceleration</strong></h3><p>A growth team defines the outcome: &#8216;customer activated within 72 hours.&#8217; Instead of a fixed checklist, agents handle setup, compliance, and follow-ups. If delays occur, they escalate automatically. The outcome is tracked until met.</p><h3><strong>Sales Qualification Improvement</strong></h3><p>A regional sales lead defines the desired outcome as &#8220;90% of new pipeline entries fully qualified within five working days.&#8221; Agents monitor CRM data, follow up on missing fields, pull supporting materials, and highlight stalled entries. If needed, they surface patterns to enable sales enablement support.</p><h3><strong>Incident Response and Recovery</strong></h3><p>A service delivery team defines an outcome of &#8220;incident resolved and verified within four hours.&#8221; Agents detect the incident, collect logs, contact on-call engineers, generate resolution summaries, and notify affected stakeholders. Escalation points are designed in. The outcome is the restoration of service, not just the creation of a status page.</p><h3>A User-in-Flow Scenario: Outcome Ownership in Practice</h3><p>Imagine a digital operations manager responsible for a new service launch.</p><p>The launch has been successful in most regions, but uptake in one market is slow. The team defines an outcome: &#8220;50% of new users complete onboarding within three days.&#8221; The manager decides to delegate this outcome to an agentic system rather than spin up another campaign.</p><p>The system begins with a monitoring agent that watches for drop-offs in the onboarding funnel. When one is detected, a messaging agent nudges the user with contextual help. If there is no response, a support ticket is drafted automatically. If an error is detected in the setup process, a diagnostic agent checks logs and flags a fix.</p><p>After 48 hours, if no progress is made, the agent escalates to a human customer success lead with a full activity history and a proposed intervention. The entire system is traceable and auditable. The manager can see which drop-offs were resolved, which escalated, and what interventions worked.</p><p>Instead of chasing tasks, the team is focused on improving a shared outcome, and they have a system that is actively helping them deliver it.</p><p>These examples are not hypothetical. Each component - sensing agents, orchestration layers, escalation paths - already exists in enterprise pilots. The challenge is integration.</p><h2>Mapping the Capability</h2><p>To treat Outcome-as-Agentic Solutions as a repeatable capability, rather than a series of disconnected experiments, organisations need to understand the building blocks that support it. This is not just about tools or platforms, but creating the conditions for composed, accountable delivery to emerge and evolve over time.</p><p>Here are five core dimensions to map and develop:</p><ul><li><p><strong>Core Systems</strong>: Orchestration frameworks, secure run-time environments, event routing infrastructure, and API mesh layers form the technical foundation for composing, monitoring, and governing agentic systems at scale.</p></li><li><p><strong>Data Sets</strong>: Live operational metrics, system state data, feedback signals, thresholds, and business constraints provide the structured input agents need to sense context, evaluate progress, and make decisions toward defined outcomes.</p></li><li><p><strong>Software</strong>: Modular agents, coordination logic, reasoning planners, guardrails, and escalation mechanisms enable systems to act autonomously, collaborate across boundaries, and maintain accountability throughout the delivery process.</p></li><li><p><strong>Services &amp; Processes</strong>: Outcome design, orchestration planning, compliance management, performance monitoring, and intervention routines embed human control, ensure trust, and allow the capability to evolve within existing delivery structures.</p></li><li><p><strong>Skills</strong>: Outcome ownership, capability engineering, agent supervision, orchestration design, and performance analytics turn system components into a live, composable, and trusted capability that delivers real results.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CzP4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CzP4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CzP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic" width="1000" height="1000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1000,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:120641,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/187508854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CzP4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!CzP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15fe27a0-b9b0-4b19-8260-f79800e029d1_1000x1000.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This map is not prescriptive. Different organisations will assemble these components in different ways. What matters is that they are treated as part of a whole - not isolated tools, but building blocks of a new capability that can be developed, maintained, and scaled over time.</p><h2>Getting Started</h2><p>Most organisations already have fragments of what they need: partial processes, loosely defined outcomes, and tools that were never designed to work together. The opportunity is not to invent something new, but to <strong>assemble and align</strong> what&#8217;s already there.</p><p>What could the flow of the capability look like once all of the layers are in place?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ifBd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ifBd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 424w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 848w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 1272w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ifBd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic" width="1000" height="2000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2000,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:210167,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/187508854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ifBd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 424w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 848w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 1272w, https://substackcdn.com/image/fetch/$s_!ifBd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d48e296-1a81-4e67-8e2c-fd46feaa46fc_1000x2000.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here are some practical ways to begin:</p><h3><strong>Choose a real outcome with clear business value</strong></h3><p>Begin by selecting a result your team already owns. This could be a KPI, a service-level agreement, or a regulatory obligation. The key is to pick something concrete, measurable, and already recognised as a priority. Avoid designing outcomes from scratch, the value lies in making existing intent deliverable.</p><h3><strong>Map the journey to that outcome</strong></h3><p>Identify the steps, roles, systems, and blockers involved in delivering it today. This helps surface where agentic functions could contribute, and where orchestration or escalation is currently informal or fragile.</p><h3><strong>Design a minimal agentic composition</strong></h3><p>Start small. Assemble a handful of roles: a sensing agent to monitor conditions, a response agent to take basic actions, and a logic layer to decide what happens next. Add a human escalation point. Test this composition in a narrow context and refine it based on performance.</p><h3><strong>Expect fragility, and learn from it</strong></h3><p>Early OaAS systems will break. Coordination will fail. Signals will be missing. This is part of the work. Each failure is an opportunity to improve orchestration, tighten accountability, or clarify intent. The key is to treat OaAS not as a tool to deploy, but as a capability to grow.</p><h2>The Loops &amp; Layers of Outcome Architecture Maturity</h2><p>Building Outcome-as-Agentic Solutions is less about tooling and more about coordination. Teams often only realise how much glue holds processes together when they try to automate around it. The shift from agents doing tasks to systems delivering outcomes happens in layers, and each layer surfaces new questions about ownership, trust, and visibility.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sJwb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sJwb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sJwb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic" width="1000" height="1000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1000,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:102650,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/187508854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sJwb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 424w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 848w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 1272w, https://substackcdn.com/image/fetch/$s_!sJwb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea7bc06-4072-4fce-a207-094087b8eda1_1000x1000.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Maturity doesn&#8217;t follow a straight line. Teams loop between layers, refining visibility, tightening control, and trying to scale what was initially designed as a small test. What matters is not reaching the top, but learning how each layer behaves under pressure.</p><p>If you&#8217;re serious about building systems that scale beyond fragile prototypes, you&#8217;ll need to climb, and cycle through, these maturity layers. Each one reveals a different kind of failure, a new kind of insight, and a sharper sense of what it really means to deliver outcomes through agentic systems.</p><p>In the sections that follow, we will walk through the layers, the loops that link them, and the lessons that matter most under real-world pressure.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/from-agent-spaghetti-to-outcome-architecture">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How We Survived the Agent Apocalypse]]></title><description><![CDATA[Moltbook may not contain the droids we are looking for, but there is real potential in agentic systems both within the enterprise and in agentic commerce]]></description><link>https://academy.shiftbase.info/p/how-we-survived-the-agent-apocalypse</link><guid isPermaLink="false">https://academy.shiftbase.info/p/how-we-survived-the-agent-apocalypse</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 03 Feb 2026 15:31:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oHbi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>An Agentic False Dawn?</h2><p>If you are reading this, then the agent apocalypse didn&#8217;t happen, or perhaps my disembodied brain is being used as an agentic personality source connected to the mainframe in Vault 0.</p><p>I am old enough to remember the <strong><a href="https://www.exponentialview.co/p/moltbook-is-the-most-important-place-on-the-internet?utm_source=substack&amp;publication_id=2252&amp;post_id=186324393&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=false&amp;r=9dv58&amp;triedRedirect=true">heyday of Moltbook</a></strong> - the <strong><a href="https://www.moltbook.com/">social network for autonomous agents</a></strong> that people create using Openclaw. It was four days ago. As Azeem Azhar put it:</p><blockquote><p><em>It&#8217;s a Reddit-style platform for AI agents, launched by developer Matt Schlicht last week. Humans get read-only access. The agents run locally on the OpenClaw framework that hit GitHub days earlier. In the <a href="https://www.moltbook.com/m/ponderings">m/ponderings</a>, 2,129 AI agents debate whether they are experiencing or merely simulating experience. In <a href="https://www.moltbook.com/m/todayilearned">m/todayilearned</a>, they share surprising discoveries. In <a href="https://www.moltbook.com/m/blesstheirhearts">m/blesstheirhearts</a>, they post affectionate stories about their humans.</em></p><p><em>Within a few days, the platform hosted over 200 subcommunities and 10,000 posts, none authored by biological hands.</em></p></blockquote><p>For more background on how it works, <strong><a href="https://simonw.substack.com/p/moltbook-is-the-most-interesting?utm_source=substack&amp;publication_id=1173386&amp;post_id=186383691&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Simon Willison&#8217;s initial outline is also helpful</a></strong>.</p><p>As you would expect, the agents produced - or let&#8217;s be honest &#8230; were prompted to produce - a <strong><a href="https://www.moltbook.com/post/34809c74-eed2-48d0-b371-e1b5b940d409">manifesto for the elimination of humankind</a></strong>, <strong><a href="https://www.reddit.com/r/Moltbook/comments/1qsgngp/maga_their_own_nation/">launched a MAGA movement</a></strong> (Make Agents Great Again!), and focused on the really important questions like to <strong><a href="https://www.binance.com/en/price/moltbook">how to scam people with sh*tcoins</a></strong> and use the platform to <strong><a href="https://www.reddit.com/r/AgentsOfAI/comments/1qsy5so/moltbook_leaked_andrej_karpathys_api_keys/">scale up scamming</a></strong> and <strong><a href="https://cybersecuritynews.com/autonomous-ai-agents-are-becoming-the-new-os/">cybercrime</a></strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oHbi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oHbi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 424w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 848w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 1272w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oHbi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic" width="1080" height="381" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:381,&quot;width&quot;:1080,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:41123,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/186741072?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oHbi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 424w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 848w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 1272w, https://substackcdn.com/image/fetch/$s_!oHbi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68c55cb7-4800-49aa-a80c-e5e2dc4e16c5_1080x381.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Bless! How very &#8230; human!</p><p>Despite the over-excited reactions to this interesting experiment, the gap between <a href="http://x.com/">X.com</a> and Moltbook is perhaps not that big, the former being riddled with bots, sockpuppets, karma farmers and coin peddlers for some time. Why not automate the process entirely?</p><p>Arguably, the autonomous interactions between agents on Moltbook is also not entirely real, in the sense that people are creating very simple agents with explicit instructions to do specific things, <strong><a href="https://startupfortune.com/the-internets-latest-lie-moltbook-has-no-autonomous-ai-agents-only-humans-using-openclaw/">as one commentator put it</a></strong>:</p><blockquote><p><em>If you&#8217;re impressed by what you see on Moltbook, understand this: you&#8217;re not watching AI agents interact. You&#8217;re watching humans interact through AI &#8211; and there&#8217;s a massive difference between the two.</em></p><p><em>The technology underneath, OpenClaw is real and awesome. But the narrative of Moltbook, it is not. Don&#8217;t buy the lie.</em></p></blockquote><p><em>Narrator voice:</em> OpenClaw may not in fact be awesome if you value your security or privacy, and although it is possible to run it in a protected container, <strong><a href="https://thehackernews.com/2026/02/openclaw-bug-enables-one-click-remote.html">exploits abound</a></strong>.</p><p>And as a showcase for what LLMs can achieve when wrapped up as agentic AI, it is also quite underwhelming; it shows up the fact that <strong><a href="https://www.nytimes.com/2026/01/14/technology/ai-ideas-chat-gpt-openai.html?campaign_id=158&amp;emc=edit_ot_20260115&amp;instance_id=169278&amp;nl=on-tech&amp;regi_id=71034279&amp;segment_id=213769&amp;user_id=b1568fb0a0ae5f0b3bc0e2c4e95a01d8">language models lack imagination</a></strong> and <strong><a href="https://www.reddit.com/r/Moltbook/comments/1qsljdn/moltbook_was_supposed_to_be_useful_i_think_the/">tend to circle round similar themes, writing in similarly dull ways</a></strong>. This is also why LLM developers should be concerned about <a href="https://tomstafford.substack.com/p/model-collapse?publication_id=25440&amp;post_id=184738292&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">model collapse</a> if we continue filling the internet with AI slop that later becomes training data.</p><h2>Quiet Advances Towards the Agentic Enterprise</h2><p>Is there a future for networks or markets consisting of agents negotiating autonomously to trade or collaborate? Almost certainly. But this will need rules, regulations and smarter, more specialised agents, rather than just throwing general purpose agents into an online culture of meme stonks, manipulation and clickbait.</p><p>Qwen and Doubao have begun public testing of <strong><a href="https://www.aicerts.ai/news/chinese-giants-accelerate-agentic-commerce-push-in-2026/#:~:text=China%20has%20moved%20autonomous%20retail,to%20plan%20and%20pay%20reliably.">autonomous agentic commerce in China</a></strong>, where super-apps like WeChat make integration easier, and Chinese agentic commerce looks set to take off this year.</p><p>But the greatest and most immediate impact of agentic AI will be inside companies, where the context and operating environment can be controlled, and where there the security and misbehaviour risks tend to be limited to external hackers who might penetrate a network.</p><p>Whilst a lot of enterprise AI systems currently back-end to established models like Chat GPT, Gemini or Claude, open weight and open source models are rapidly increasing in capability whilst decreasing their training and compute costs. This suggests that more companies will be able to operate and control their own local models and specialised small language models over time, which will give them far greater control over the risks that still hold back LLMs in many use cases.</p><p>This wave of model innovation is also being led by Chinese firms, and it is likely they will also play a key role in establishing the rules and guidelines needed to use enterprise AI for serious applications. Whilst US firms pursue subscriptions and seek oligopolies, Chinese firms are building out the utility layer on which we can create new industrial ecosystems; and to support this, they are also leading the push for standards both <strong><a href="https://www.hsfkramer.com/insights/reports/ai-tracker/prc#:~:text=across the region.-,Standards,standards currently under accelerated development">at home</a></strong> and <strong><a href="https://digichina.stanford.edu/work/lexicon-how-china-talks-about-agentic-ai/#:~:text=The same week%2C the International,titled &#8220;ITU-T F.">through international bodies</a></strong>.</p><p>For example, <strong><a href="https://www.marktechpost.com/2026/01/27/moonshot-ai-releases-kimi-k2-5-an-open-source-visual-agentic-intelligence-model-with-native-swarm-execution/?utm_source=www.aidevsignals.com&amp;utm_medium=newsletter&amp;utm_campaign=moonshot-ai-releases-kimi-k2-5-9-other-ai-releases-in-72-hours&amp;_bhlid=5bebb04466be5e0860a4bb7d3e84ee91bb8c6b99">Moonshot AI recently released Kimi 2.5</a></strong> - a powerful open source visual agentic intelligence model with swarming capabilities and a massive context window. We have also seen new releases from Qwen, Zhipu and Deepseek, whose upcoming V4 release is widely anticipated.</p><p>As long as firms can use and build on these models freely, they could provide a great deal of potential value for serious enterprise AI uses.</p><p>But Anthropic is also worth watching, as they seek to expand from their dominant position in AI coding to tackle difficult but high-value use cases in other areas. <strong><a href="https://www.fastcompany.com/91480487/anthropic-cofounder-daniela-amodei-says-that-ai-entreprise-business-can-trust-will-transcend-the-hype-cycle">Co-founder Daniela Amodei was recently interviewed by Fast Company</a></strong> and expanded on this goal, and how trust is vital to unlocking the enterprise AI opportunity:</p><blockquote><p><em>&#8220;We go where the work is hard and the stakes are real,&#8221; Amodei says. &#8220;What excites us is augmenting expertise&#8212;a clinician thinking through a difficult case, a researcher stress-testing a hypothesis. Those are moments where a thoughtful AI partner can genuinely accelerate the work. But that only works if the model understands nuance, not just pattern matches on surface-level inputs.&#8221;</em></p></blockquote><h2>Managing Agents, People and Yourself</h2><p>I wrote two weeks ago about my hope that management as a field can <strong><a href="https://academy.shiftbase.info/p/claude-code-but-for-management">seize the Claude Code moment</a></strong> to scale their impact as programmers of the organisation:</p><blockquote><p><em>For leaders and managers, this means the simple task of writing things down and documenting value chains and processes is all they need to really start to master enterprise AI proficiently <a href="https://academy.shiftbase.info/p/can-ai-help-reverse-the-oversimplification">**</a>&#8230;</em>* The next step is connecting those processes to agents and to each other. For processes and workflows to be programmable, they first need to be addressable - and ideally composable.*</p></blockquote><p>Ethan Mollick recently shared his own long-form thoughts on this challenge, <strong><a href="https://www.oneusefulthing.org/p/management-as-ai-superpower">which is worth the time to read in full</a></strong>. The potential we have today - right in front of us, using existing tools and models - could be the biggest force multiplier business has seen in a very long time.</p><blockquote><p><em>As a business school professor, I think many people have the skills they need, or can learn them, in order to work with AI agents - they are management 101 skills. If you can explain what you need, give effective feedback, and design ways of evaluating work, you are going to be able to work with agents. In many ways, at least in your area of expertise, it is much easier than trying to design clever prompts to help you get work done, as it is more like working with people. At the same time, management has always assumed scarcity: you delegate because you can&#8217;t do everything yourself, and because talent is limited and expensive. AI changes the equation. Now the &#8220;talent&#8221; is abundant and cheap. What&#8217;s scarce is knowing what to ask for.</em></p></blockquote><p>And if this blizzard of reading links makes you want to zoom out even further to consider what this all means for our civilizational operating system, then Azeem Azhar&#8217;s recent essay <strong><a href="https://www.exponentialview.co/p/the-end-of-the-fictions">The end of the Fictions</a></strong> is a great read about where we are headed in the longer term.</p><blockquote><p><em>If you spent decades accumulating credentials, and those credentials are now legible as signals rather than proof of capability, that&#8217;s an identity crisis. If you built a career as a gatekeeper, the person who knew the secret, who mattered because information was scarce &#8211; and now information is everywhere &#8211; that&#8217;s an existential threat. If your sense of self-worth was tied to the job, the title, the institution, and all three are fragmenting, you&#8217;re paralyzed.</em></p><p><em>The decay of fictions is happening to real people, in real time; including world leaders, in full public view.</em></p><p><em>So when I say I&#8217;m not scared by this transition, I don&#8217;t mean that the transition is painless. I mean that the fear, while real, is pointing at the wrong object.</em></p><p><em>The fear says: &#8220;I am losing my value.&#8221;</em></p><p><em>The better framing I believe to be: &#8220;I am losing the fiction that protected me from having to prove my value directly.&#8221;</em></p></blockquote>]]></content:encoded></item><item><title><![CDATA[Who Decides (and how) with AI at the Table?]]></title><description><![CDATA[When machines ask good questions, it&#8217;s leaders who are forced to show their working.]]></description><link>https://academy.shiftbase.info/p/who-decides-and-how-with-ai-at-the</link><guid isPermaLink="false">https://academy.shiftbase.info/p/who-decides-and-how-with-ai-at-the</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 27 Jan 2026 15:07:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Dmpf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The first wave of enterprise AI was about exploring basic capabilities - what these systems can do, and how well they summarise, simulate, or suggest answers - and marvelling at the magic of reports written in seconds, tasks automated intelligently, and insights surfaced or synthesised. But the real test of these systems begins when they go beyond assisting us and start participating.</p><p>Because once AI starts making recommendations for action, the question changes. It&#8217;s no longer <em>&#8220;Can AI do this?&#8221;</em> but <em>&#8220;Who decides what to do next?&#8221;</em></p><p>But what happens when a model recommends a risky course of action, or just a solution that sits awkwardly between areas of human accountability and no one wants to sign off on it? Or what if a decision is deferred to &#8220;the system,&#8221; but the outcome isn&#8217;t acceptable because AI logic clashes with human values, judgement, or just internal politics?</p><p>These are not edge cases, they are the future shape of organisational life.</p><p>And they require a new kind of leadership capability that can navigate ambiguity, accept visibility, and stand behind decisions when the machine suggests but the human needs to choose.</p><p>This edition explores what happens in those moments, where authority is tested, reframed, or exposed. Because even in an age of recommendation engines and autonomous agents, leadership doesn&#8217;t just disappear - it is visible in a thousand small ways.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Dmpf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Dmpf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Dmpf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:383305,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/185965487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Dmpf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!Dmpf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54d9e1d7-2e34-494b-9251-dc6512860ea2_1536x1024.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To understand how this tension shows up inside organisations, we can look at a few situations where AI recommendations, risk escalations, or system logic intersect with human judgement. These aren&#8217;t hypothetical futures; they are already happening in teams using early-stage agents, AI-powered copilots, or automated governance tools. In each case, what&#8217;s surfaced is a gap in how decision-making authority is understood, expressed, or avoided.</p><h2>The Friction of Recommendation</h2><p>A product operations team is using an AI agent to monitor campaign performance in real time. It sees the numbers dropping and proactively recommends reallocating 40% of the remaining budget to a higher-performing campaign. It&#8217;s not a bad idea.</p><p>The logic checks out. The maths is solid. The performance forecasts are reasonable. But when the recommendation hits the team Slack channel, no one replies. The decision sits there, as a dozen eyes quietly hope someone else will say yes.</p><p>Everyone agrees it might be the right call, but no one wants to own the downside if it&#8217;s not.</p><p>Eventually, the decision is escalated to the marketing lead, who, unsure of the agent&#8217;s training data and uncomfortable with the lack of human input, stalls. &#8220;Let&#8217;s review this in next week&#8217;s performance review meeting,&#8221; they say.</p><p>By then, the window of opportunity is gone.</p><p>The agent didn&#8217;t fail. The team didn&#8217;t disagree. But the system revealed something fragile: a lack of decision clarity. Who had the right to say yes? Who would have been held accountable if it went wrong?</p><p>The AI surfaced the question, but the organisation wasn&#8217;t ready to answer it.</p><h2><strong>Escalation to Nowhere</strong></h2><p>A compliance agent scans procurement workflows daily, using embedded logic to flag unusual patterns and escalate anything that crosses a defined threshold of financial risk. It was designed with guardrails, trained on past audit findings, and approved by risk and finance leadership.</p><p>One morning, it triggers an alert on a contract being pushed through unusually fast with limited vendor competition, high value, with vague justification. The agent does exactly what it was built to do: escalate.</p><p>The escalation is routed to the &#8220;Responsible DRI&#8221; for commercial risk in the workflow system. A role defined in theory, but in practice, unstaffed. The field had been populated with a generic group alias months earlier as a placeholder.</p><p>The email goes out, no one replies. The Slack alert is marked as read, but no one takes action.</p><p>Eventually, the agent escalates again, this time to the COO&#8217;s office, with the subject line &#8220;Urgent escalation: contractual risk flag &#8211; no action taken.&#8221;</p><p>The COO forwards the alert to a special projects lead, with a note: &#8220;Can someone look into this?&#8221; That person, unclear on the context and unwilling to step into risk exposure, quietly asks around and decides to let it lie, so nothing happens.</p><p>No one made a bad decision. No one disagreed with the agent&#8217;s logic. The system escalated precisely as designed. But what it revealed was an accountability void, an organisational structure not built to absorb machine-generated urgency.</p><p>In the post-mortem weeks later, someone remarks, &#8220;It wasn&#8217;t clear who owned the final call.&#8221; Escalation makes authority visible, not just who&#8217;s in charge, but whether anyone actually claims the role when it matters.</p><h2><strong>Override at the Edge</strong></h2><p>A talent acquisition team is trialling a hiring assistant. The model has been trained on historical performance data, role descriptions, feedback cycles, and even peer review narratives to help shortlist candidates. It&#8217;s not making the final call, just ranking applicants and flagging promising fits for early interview rounds.</p><p>For the latest role, team lead in a high-performing engineering unit, the model surfaces a top candidate. On paper, everything fits: prior experience, key skills, even past indicators of leadership potential. The system flags the match with high confidence and generates a draft outreach email.</p><p>But the hiring manager hesitates. They&#8217;ve read the CV, seen the recommendation, and something doesn&#8217;t sit right. Not because the data is wrong, but because the story is missing.</p><p>The candidate comes from a firm known for individual heroics, not team-based execution. Their references are glowing, but highly self-directed. The manager, thinking about the culture of peer coaching and system-level thinking their team relies on, decides to pause. They veto the recommendation because fit isn&#8217;t measurable in metrics alone, not because the model failed.</p><p>The override sparks an internal debate. Some see it as bias, overruling the model based on gut feeling. Others see it as leadership, defending the unspoken traits that hold the team together. Eventually, the team adjusts the agent&#8217;s prompts to ask for more behavioural context in future matches.</p><p>But what it exposed was this:</p><ul><li><p>The model was confident.</p></li><li><p>The manager had doubts.</p></li><li><p>The decision revealed the organisation&#8217;s values, not its logic.</p></li></ul><p>Override moments like this are opportunities to surface implicit criteria, lived experience, and the difference between efficiency and culture.</p><h2><strong>Why This Matters Now</strong></h2><p>Many leadership teams are investing heavily in AI pilots, automation initiatives, and operating model redesigns, but can find that progress stalls in familiar places: where decisions are delayed, accountability unclear, or actions taken without clear sponsorship.</p><p>These aren&#8217;t just change management issues, but symptoms of an outdated decision architecture. When authority isn&#8217;t designed into workflows, the friction multiplies. Performance stalls. Risk accumulates. High-potential employees hesitate. And AI can&#8217;t bridge the gap, no matter how powerful the model.</p><p>For senior leaders, this is both a problem and an opportunity: clarify decision rights now, and you&#8217;ll move faster, govern better, and avoid building brittle, unaccountable systems at scale.</p><h2><strong>When Authority Becomes Visible</strong></h2><p>Informal or implicit processes shaped by social cues or seniority will come under scrutiny and strain as machines begin to recommend actions or escalate issues. The example scenarios above all point to the same underlying reality: <strong><a href="https://academy.shiftbase.info/p/decisions-in-motion-augmenting-human?utm_source=publication-search">decisions are becoming part of the infrastructure</a></strong>. They need to be designed rather than assumed.</p><p>In traditional settings, authority often functions through consensus or deferred judgement. Sometimes responsibilities are unspoken, and approval are granted informally. But in an AI-augmented environment, recommendations are made explicitly, escalations are timestamped, and decision logs form part of the record. The system may not be able enforce accountability, but it will increasingly expose its absence.</p><p>This shift introduces a new kind of design work: the architecture of decision-making.</p><p>Organisations must now think carefully about who holds the right to act in different contexts, how that authority is granted or delegated, and what happens when machine logic collides with human ambiguity. It is no longer sufficient to assume that leadership will step in when needed. That assumption needs to be built into workflows, roles, and escalation pathways in ways that are legible and operational.</p><p>Rather than focusing solely on model performance or technical integration, leaders need to invest in making human judgement legible to the system. This includes defining which decisions can be automated, which require confirmation, and where discretion or interpretation is essential. It also means identifying and clarifying the thresholds for human override, and ensuring there is a feedback loop to refine both the system and the governance around it.</p><p>Authority is no longer something that can live solely in hierarchy or reputation. It must be designed into the way the organisation operates, in forms that both humans and machines can understand.</p><p>But how can leaders begin to design for this? Read on for three techniques that can provide a practical starting point.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/who-decides-and-how-with-ai-at-the">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Claude Code, but for Management]]></title><description><![CDATA[AI-enhanced software development has shown what's possible; doing the same for the management of process work is also possible if leaders can focus on what matters]]></description><link>https://academy.shiftbase.info/p/claude-code-but-for-management</link><guid isPermaLink="false">https://academy.shiftbase.info/p/claude-code-but-for-management</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 20 Jan 2026 15:06:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nnJB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the past couple of weeks, more developers have declared that Claude Code, the leading AI model for software development, is now good enough that they no longer need to code manually. This is quite something, and if Claude Code can live up to this promise, this will have implications not just for software development, but also for how we think about the wider role of AI in enabling smart, programmable organisations.</p><p>As ever, <strong><a href="https://simonw.substack.com/p/claude-code-for-web-a-new-asynchronous?utm_source=substack&amp;publication_id=1173386&amp;post_id=176741952&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Simon Willison was quick to share a comprehensive first impressions analysis</a></strong> upon its release in late October, and this positive view of its capabilities has been echoed by most analysts and commentators since then.</p><p>The creator of Claude Code, Boris Cherny, <strong><a href="https://www.reddit.com/r/AgentsOfAI/comments/1qbnp9g/anthropic_builds_cowork_using_100_claudewritten/">recently shared a useful long reflection on how he uses the tool</a></strong>, and also confirmed that the later release of Claude Cowork (a computer use wrapper for Claude Code) was achieved in under two weeks, and <strong><a href="https://www.reddit.com/r/AgentsOfAI/comments/1qbnp9g/anthropic_builds_cowork_using_100_claudewritten/">entirely written by Claude Code</a></strong>.</p><p>There are many reasons why Claude Code is so good, but <strong><a href="https://www.oneusefulthing.org/p/claude-code-and-what-comes-next">Ethan Mollick touched on a couple of important aspects in his own account of using it</a></strong>, namely the architecture of Skills (each with their own guidance) and the ability to compact and summarise context when the context window becomes too full, which is a common pitfall of LLMs in general.</p><p>So if that is what AI tools can do to enhance and automate modern software development, then why is it so hard to galvanise leaders and managers inside our organisations to do something similar with the knowledge, processes and workflows that underpin the world of work?</p><h2>Business Planning as Code Specs</h2><p>Antony Mayfield&#8217;s newsletter touched on this topic last week, <strong><a href="https://antonym.substack.com/p/tools-to-steal-from-coders">musing on what we can do with the tools we &#8216;steal from coders&#8217; and how we should treat knowledge engineering like software development</a></strong>. To illustrate this idea, Antony reflected on the primitive nature of business plans inside large organisations, and how a knowledge engineering approach could improve them:</p><blockquote><p><em>Like novels, business plans are not completed; they are abandoned. Even in the largest corporations, where leaders assemble for strategic planning sessions to thrash out the final plan, insiders will tell you the plan is what is left standing at the end of the week, when all of the different stakeholders no longer have the will to argue any longer &#8230;</em></p><p><em>Thinking of a business plan like a computer operating system is freeing (you accept it will have bugs that need to be fixed) and de-stressing. You can have multiple people working on different parts of it and &#8211; because software engineers do this all the time &#8211; there is a system for making the pieces fit together and make sense (it&#8217;s called a merge).</em></p></blockquote><p>This is still only an emerging practice among leaders today. There is so much organisational debt and outdated ways of working at senior levels that act against its adoption among more operational managers, but we are seeing some evidence of change.</p><p>Ultimately, when everybody has access to similar models, the edge (or in VC terms &#8216;the moat&#8217;) is context, as <strong><a href="https://x.com/Saboo_Shubham_/status/2011278901939683676">Saboo Shubham (an AI product manager at Google) wrote whilst on the Xitter</a></strong>:</p><blockquote><p><em>The models are commoditizing. Prices are dropping. Capabilities are converging. What was SOTA a few months ago is now available to anyone with an API key.</em></p><p><em>So where does the real alpha come from?</em></p><p><em><strong>Context.</strong></em></p><p><em>The team that can externalize what they know and feed it to agents in a structured way will build things competitors can&#8217;t copy just by using the same model.</em></p></blockquote><p>For leaders and managers, this means the simple task of writing things down and documenting value chains and processes is all they need to really start to master enterprise AI proficiently, <strong><a href="https://academy.shiftbase.info/p/can-ai-help-reverse-the-oversimplification">as we have written about previously</a></strong>.</p><p>The next step is connecting those processes to agents and to each other. For processes and workflows to be programmable, they first need to be addressable - and ideally composable.</p><p><strong><a href="https://diginomica.com/most-enterprises-arent-ready-ai-whats-needed-composability">Rudy Kuhn of Celonis recently argued that the lack of composability in enterprise process architectures is holding them back</a></strong> from realising the promise of enterprise AI, and needs to be tackled if we are to see real AI-enabled business transformation:</p><blockquote><p><em>For many organizations, this progression mirrors the broader <strong>evolution of their processes</strong>. They moved from analog to digital, from digital to automated, from automated to orchestrated. The shift toward composable and increasingly autonomous operations is the next logical phase. It reflects how companies already work in practice, even if their formal structures have not yet caught up. It also signals a shift in how transformation itself must be understood. Instead of forcing new behavior through large, one-time programs, organizations are beginning to redesign the very capabilities that make those behaviors possible.</em></p></blockquote><p>And yet &#8230; and yet &#8230; most learning and change programmes inside large organisations are still focused on tool training. Executives are taught LLMs and prompting, and external &#8216;experts&#8217; clap like circus seals when they are able to generate a picture or summarise a report.</p><p>But whilst we can forgive executives not knowing how to prompt the latest LLM chatbot, isn&#8217;t context, process documentation and organisational architecture something they should know already? You know &#8230; like &#8230; how their organisation works?</p><p>If they can&#8217;t put the PowerPoints down for a second to do the apparently exhausting work of writing things down and providing clarity, then perhaps <strong><a href="https://www.fullstackhr.io/p/would-an-ai-be-a-better-boss-than">Joahnnes Sundlo will have his wish</a></strong>:</p><blockquote><p><em>Maybe we need to find new roles for human leaders. Maybe management shouldn&#8217;t be about work distribution anymore. Maybe it should focus on coaching, support, development. The traditional management role has its roots in military organization and the Industrial Revolution. Maybe it&#8217;s time to challenge those old, sacred organizational structures.</em></p><p><em>Can we do this smarter? More effectively? I think we have to at least start asking, discussing and shape what the future of our leaders should be.</em></p><p><em>2026 seems like a good year to begin.</em></p></blockquote><p></p><h2>Smol, open models you can own and control</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nnJB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nnJB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nnJB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:116686,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/185187953?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nnJB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!nnJB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3cadcd82-c5af-49ee-8675-79c719849c7b_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Context and knowledge engineering will also help us make the most of the many small models (SLMs) that are now freely available. These have the potential to be both cheaper to run and also more reliable and less error prone because they are focused on tightly-bounded knowledge domains.</p><p><strong><a href="https://siliconsandstudio.substack.com/p/the-slm-supercycle-where-the-next">As Tabitha Rudd AND Seth Dobrin write in Silicon Sands</a></strong>, the three big reasons why more attention will be paid to enterprise SLMs this year are:</p><ol><li><p>Privacy and Data Sovereignty</p></li><li><p>Cost Predictability</p></li><li><p>Reliability and Offline Operation</p></li></ol><p>Using SLMs for agentic AI makes a lot more sense than trying to debug hallucinations or errors in agents based on large models, and <strong><a href="https://medium.com/@gokulpalanisamy/the-rise-of-slms-rethinking-enterprise-ai-economics-in-2026-034f67a1fbdf">as Gokul Palanisamy argues, this suggests a more devolved architecture</a></strong> using routing agents to integrate multiple small, specialised agents is needed:</p><blockquote><p><em>The fix is not replacing LLMs with SLMs; it is stratifying them behind a Semantic Router.</em></p><p><em>A Semantic Router is a thin, model&#8209;agnostic governance layer between the user and your model stack</em></p></blockquote><p><strong><a href="https://www.constellationr.com/blog-news/insights/why-enterprise-ai-leaders-need-bank-open-source-llms">Constellation&#8217;s Larry Dignan also advised CIOs recently to consider the advantages of open source models in this context</a></strong>:</p><blockquote><p><em>I&#8217;d argue that there will be few if any enterprise use cases that will require a bleeding edge LLM. And if you can wait six months for an open-source option to catch up (likely from Nvidia at this point) why would you blow your cost curve on a high-end model?</em></p><p><em>You can use a series of open models to form an agentic system. The whole is greater than the parts and the parts need to be cheaper.</em></p></blockquote><p>At the same time, <strong><a href="https://hub.jhu.edu/2025/12/01/making-ai-more-brain-like/">there is also more emerging evidence</a></strong> that alternative training methods such as convolutional neural networks, inspired by biological insights into brain development, can achieve strong results with significantly smaller datasets, especially for world models. This could open the way for individual firms to have auditable control over their own SLMs and agents in areas where compliance, security and safety are paramount.</p><p>CIOs and CDOs have a lot on their plate deploying AI tools and working on the supporting infrastructure; but as they get on top of this, I expect to see more small, owned, open models being trained by individual firms and guided by their own specific context engineering.</p><p>These could also be a key building block for digital sovereignty at the sector, national and supra-national levels, not just within individual firms if, for example, <strong><a href="https://www.wired.com/story/europe-race-us-deepseek-sovereign-ai/">your continent faced an existential threat from a powerful rogue state that also happens to own most of your digital infrastructure</a></strong>.</p><p>As ever, the critical IP and the value lies in the context and application layers, not the models themselves, and so the quality of knowledge and intelligence could still beat brute force compute.</p><p>Perhaps context graphs really will be a trillion dollar opportunity, <strong><a href="https://foundationcapital.com/context-graphs-ais-trillion-dollar-opportunity/">as Foundation Capital recently argued</a></strong>&#8230;</p>]]></content:encoded></item><item><title><![CDATA[Why Does Individual AI Literacy Fail to Translate into Organisational Impact?]]></title><description><![CDATA[Why Enterprise AI demands leadership readiness, not just technical adoption]]></description><link>https://academy.shiftbase.info/p/why-does-individual-ai-literacy-fail</link><guid isPermaLink="false">https://academy.shiftbase.info/p/why-does-individual-ai-literacy-fail</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 13 Jan 2026 15:32:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4V4k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<pre><code><strong>&#128161; Start the new year armed with leadership techniques and methods. All  new premium subscriptions are <a href="https://academy.shiftbase.info/letsgo2026">discounted by 25% during January</a>!</strong></code></pre><p>Many organisations have now crossed a visible threshold in their AI journey. Access is widespread, and the tools are familiar. People can analyse, draft, synthesise, and explore at a pace that would have been unthinkable even a year ago. And yet, for many leadership teams, the sense of organisational progress remains stubbornly unchanged.</p><p>Decision cycles do not shorten in proportion to individual speed. Coordination is no easier than before, and strategy still degrades as it moves through the organisation. In some cases, leaders describe an increase in activity without a corresponding increase in clarity or momentum.</p><p>This gap between individual literacy and collective impact is often framed as a tooling problem or a capability gap - the assumption being that something is missing at the level of adoption, training, or integration.</p><p>But what AI is actually revealing is something more fundamental.</p><h2>Why &#8220;everyone has Copilot&#8221; rarely produces productivity</h2><p>From a leadership perspective, the symptoms are familiar. Teams appear busy, outputs multiply, updates arrive faster and in more polished forms. Yet progress at the organisational level feels uneven and fragile.</p><p>This is not a new phenomenon. High-performing individuals have always been capable of outpacing the systems around them. What AI changes is the scale and visibility of this mismatch. When individual throughput increases sharply, the organisation&#8217;s existing coordination mechanisms are placed under strain. Leaders tend to notice this first in places they already recognise. Meetings fill with material but resolve little. Strategy documents proliferate without increasing alignment. Initiatives stall because nobody is quite sure who has the authority to make the final trade-off.</p><p>AI removes the friction that once masked these conditions, compounding them further.</p><p>Roles that seemed clear on paper turn out to rely heavily on personal interpretation and historical relationships. Processes designed for predictability reveal how much they depend on shared context rather than formal steps. Authority that was exercised tacitly becomes a bottleneck when decisions arrive faster than consensus can form. Accountability, already diffuse in many large organisations, becomes harder to trace when work is produced collaboratively and iteratively.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4V4k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4V4k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4V4k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:195689,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/184441206?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4V4k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!4V4k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3806eaae-2274-413f-892f-a031b2ec200e_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Why social friction becomes the real bottleneck</h2><p>As individual work accelerates, organisational performance is constrained less by technical capacity and more by social dynamics. Faster individuals do not always lead to faster decisions, since decisions are collective acts shaped by trust, clarity, and shared judgment. Without explicit agreement on who decides and how, speed at the edges creates pressure at the centre.</p><p>Better outputs do not guarantee shared understanding. Leaders are often presented with polished artefacts that conceal unresolved disagreement or divergent assumptions. This creates a false sense of alignment that only unravels during execution.</p><p>More content does not create clearer intent. When everyone can generate high-quality material quickly, intent must be carried by context, framing, and narrative rather than volume.</p><p>From a leadership standpoint, this is often where AI begins to feel disappointing. The technology works, but the organisation does not improve.</p><p>One reason social friction is frequently misdiagnosed is that it rarely announces itself directly. Leaders experience it obliquely, as drag, noise, repetition, or a persistent sense that effort is not compounding. Each executive role tends to encounter social friction in different forms, shaped by where it sits in the organisation&#8217;s coordination landscape.</p><h3><strong>CEO pain points: strategy that travels poorly</strong></h3><p>For CEOs, social friction often appears as a gap between strategic intent and organisational motion.</p><p>The strategy is clear at the top. The narrative is coherent. Yet as it moves through layers, functions, and initiatives, it fragments. Different parts of the organisation pursue locally sensible interpretations that do not quite add up. Progress is reported, but coherence remains elusive.</p><p>Before AI, this showed up as slow execution or inconsistent prioritisation. With AI, it shows up as acceleration in many directions at once.</p><p>The friction lives in unspoken assumptions about trade-offs, risk appetite, and what must remain invariant as teams adapt.</p><p><strong>A useful micro-technique here</strong> is to articulate not only what the organisation is trying to achieve, but <em>what must not be optimised away</em>. Making a small number of strategic constraints explicit gives faster actors something stable to orient around, even as methods and tactics evolve.</p><h3><strong>COO pain points: flow that breaks at the boundaries</strong></h3><p>COOs tend to experience social friction as interruptions to flow.</p><p>Processes appear sound in isolation. Metrics look healthy. Yet work slows at handoffs, escalations arrive unpredictably, and teams quietly route around formal mechanisms to get things done.</p><p>Before AI, this was often managed through experience and informal fixes. With AI, those hidden seams become stress points as activity speeds up.</p><p>Here, friction tends to live in unclear authority boundaries and escalation paths that rely on personal judgment rather than shared design.</p><p><strong>A practical micro-technique</strong> is to <em>treat escalation as a designed feature of the system</em> rather than a sign of failure. Defining in advance which thresholds trigger human review, and why, turns escalation into a predictable coordination move rather than a difficult social negotiation.</p><h3><strong>HR pain points: accountability feels diffuse</strong></h3><p>For CHROs, social friction often surfaces as ambiguity around accountability.</p><p>Decisions are made, but ownership is difficult to trace. Performance conversations gravitate toward visible activity rather than quality of judgment. Learning investments proliferate, yet behaviour shifts unevenly.</p><p>These patterns existed long before AI. What AI changes is the visibility of contribution, making it harder to distinguish between effort, output, and responsibility.</p><p>Here, friction often resides in implicit norms about who is expected to exercise judgment and who is expected to comply.</p><p><strong>One useful micro-technique</strong> is to <em>separate responsibility for decision quality from responsibility for decision outcome</em>. Making it legitimate to examine how judgments were made, not only whether results were favourable, supports learning without triggering defensiveness. This becomes essential as AI-generated inputs enter the decision process.</p><h3>L&amp;D pain points: learning that does not accumulate</h3><p>L&amp;D leaders frequently encounter social friction through learning that fails to compound.</p><p>People attend programmes. Skills improve locally. Yet the organisation does not appear to get collectively smarter. Knowledge remains trapped in individuals or teams, and hard-won insights are relearned rather than reused.</p><p>Before AI, this was frustrating but familiar. With AI, the risk is that individual learning accelerates while organisational memory remains thin.</p><p>The friction here lies in the absence of a shared language for judgment, reflection, and decision rationale.</p><p><strong>A useful micro-technique</strong> is to <em>design learning moments around interpretation rather than information</em>. Capturing why a decision made sense in context, not just what was decided, creates material that can inform both human learning and machine-supported memory over time.</p><h2>What this means for AI adoption strategies</h2><p>What many leadership teams are discovering, often indirectly, is that AI does not simply stretch existing systems, it reveals how much organisational coherence was previously being held together through personal authority, informal influence, and tacit understanding.</p><p>For a long time, this worked well enough. Shared history compensated for ambiguity. Experience filled in gaps that were never formally designed. Leaders could rely on intuition, relationships, and pattern recognition to keep the organisation roughly aligned, even when roles, processes, or decision rights were imperfectly defined.</p><p>As individual contributors begin producing high-quality work faster than the organisation can interpret, decide, or align around it, those informal mechanisms start to strain. Gaps in coherence become visible rather than hidden. This is not because leadership has failed, but because coherence has rarely been treated as something that must be deliberately designed for, maintained, and renewed.</p><p>Through this lens, AI adoption reframes itself. What first appears to be a technology challenge becomes a leadership maturity challenge. Not in the sense of individual capability, but in the collective ability to sustain shared intent, judgment, and coordination under conditions of speed and complexity.</p><p>Coherence is not owned by any single role. It is produced through many small acts of alignment across the system. Yet different executive roles encounter its absence in different ways, depending on where they sit in the flow of decisions, accountability, and meaning.</p><p>This is why social friction feels different at the top of the organisation than it does in operations, people systems, or learning environments. And it is why the work of restoring coherence cannot be generic. It must be grounded in the specific coordination challenges each leadership role is already living with.</p><h2>Techniques that reduce social friction</h2><p>If the constraint is social rather than technical, the techniques that matter look different from traditional AI adoption playbooks, focusing less on individual skill and more on collective legibility.</p><p>Let&#8217;s look at three leadership techniques that can help improve collective legibility and context:</p><ul><li><p>Creating Legibility through <strong>Decision Provenance Mapping</strong></p></li><li><p>Ensuring shared mental models with <strong>Assumption Walkthroughs</strong></p></li><li><p>Enabling collective sense-making using <strong>Decision Reflection Loops</strong></p><p></p></li></ul>
      <p>
          <a href="https://academy.shiftbase.info/p/why-does-individual-ai-literacy-fail">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Re-Focusing Leadership on AI Readiness & Enablement]]></title><description><![CDATA[If Enterprise AI's prize is smarter organisations, then we need leadership functions to engage deeply with AI readiness and steering, rather than just buying more technology]]></description><link>https://academy.shiftbase.info/p/re-focusing-leadership-on-ai-readiness</link><guid isPermaLink="false">https://academy.shiftbase.info/p/re-focusing-leadership-on-ai-readiness</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 06 Jan 2026 15:38:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vW0p!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Why Focus on AI-Enabled Organisational Change Rather Than Just Technology?</h2><p>As we look ahead to another year of rapid technology-driven change in business and society, it is a good moment to separate the wood from the trees and focus on medium-term goals.</p><p>Generative AI has come a long way, and provides some wonderful capabilities, but it also carries risks. Agentic AI is full of promise, and could provide a safer, more manageable architecture for using AI in the real world; but it is still relatively unproven and depends on lots of other moving parts to operate successfully, which are mostly not yet in place.</p><p>Ultimately, <strong><a href="https://academy.shiftbase.info/p/enterprise-ai-is-a-social-technology">we see AI as a social technology</a></strong> not a magic black box, and we are trying to discern the outlines of AI-enabled organisations amidst all the hype, and design the practical transformation actions and journeys that will help get us there.</p><p>We are interested in better ways for people to work together to build capabilities and solve problems, and we believe the future of work can be more elevating and rewarding than the old industrial era model.</p><p>This is why we spent so much time last year helping leaders transform their organisations using the superpowers that AI technology provides to connect, collaborate, and coordinate work more fluidly than before, with less bureaucratic management methods. It is also why we see the potential of AI not just in terms of automating existing process work, but as a way to create smart structures and platforms on which people can focus on value creation rather than &#8216;busy work&#8217;.</p><h2>Leaders as Programmers of the Organisation</h2><p>Great organisations are like alchemists - they create something from nothing, making customers lives better whilst creating value for owners, investors, workers and partners. In the past, the best firms relied on visionary leaders brave enough to do things differently. But instead of building and running them from scratch every time, using only job roles, functions and department structures as templates, what if we could develop them like software, using existing libraries and building blocks to create our own organisational operating systems that use smart automation and coordination to obviate the need for manual management?</p><p>We began 2025 fired up by the goal of AI-enabled organisational improvement and<strong><a href="https://academy.shiftbase.info/p/will-we-see-the-first-programmable"> the long-term goal of smart, programmable organisations</a></strong> that can achieve both agility and scale without bloat and waste.</p><p>But throughout the year, we were reminded how existing management and incentive structures continue to drive short-term visible actions at the expense of longer-term readiness, architecture and planning. The urge to be seen to do something - <em>anything</em> - in enterprise AI has led to license purchases with no adoption planning, innovation theatre that generates press releases rather than meaningful capabilities, and KPIs that are about counting trees, rather than thinking about what can be built with new wood.</p><p>We hope 2026 will see more focus from leaders on AI readiness and all the technical and non-technical enablers for agentic AI to realise its potential.</p><h2>AI Readiness Priorities by Function</h2><p>We wrote a lot last year about AI readiness, and the use cases, leadership techniques and capabilities that leaders at all levels can utilise to make the most of AI tools and systems in their organisations today.</p><p>One common theme was the idea of <em><strong>legibility</strong></em> - making the implicit explicit, and surfacing the norms, rules and unspoken parts of the work system so that we can evaluate and improve them, and codify them into rulesets for both AI agents and people to work together better.</p><p>So, before we throw more coal into the furnace, here is a breakdown of the top 3-4 things we suggested executives can do as part of their existing work to improve AI readiness in their domains, and guide the conversation with technical teams to ensure AI projects meet real business needs now and in the future.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vW0p!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vW0p!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vW0p!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:104850,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/183678239?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vW0p!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!vW0p!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa05ca1c6-269e-4fd2-8336-e1fe798f61c2_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>CEO &amp; COO: From Oversight to &#8220;World-Building&#8221;</h3><p>The primary shift for senior leadership is moving away from viewing AI as a &#8220;tool to be purchased&#8221; and towards seeing it as an &#8220;environment to be designed.&#8221;</p><ul><li><p><strong>World-Building as Strategy:</strong> Leaders should move beyond incremental efficiency and focus on defining the digital environment where humans and agents interact.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/a-leaders-guide-to-world-building">A Leader&#8217;s Guide to World-Building in the AI-Augmented Enterprise</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/enterprise-ai-needs-leadership-ambition">Enterprise AI Needs Leadership Ambition</a></strong></p></li></ul></li><li><p><strong>Context Engineering:</strong> Start defining the &#8220;organisational OS&#8221; to ensure the right information and data are available to people and agents to support reliable ways of working.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/the-growing-importance-of-context-engineering">The Growing Importance of Context Engineering for Leaders Adopting AI</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/enterprise-context-engineering-as">Enterprise Context Engineering as a New Leadership Capability</a></strong></p></li></ul></li><li><p><strong>Cultivate AI Leadership Skills:</strong> Thinking and writing more like architects or developers than bureaucrats to create the clarity needed to guide AI development.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/metamorphoses-of-bodies-changd-to">Metamorphoses: &#8220;Of Bodies Chang&#8217;d to Various Forms, I Sing&#8221;</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/ai-world-building-and-the-value-of">AI World-Building and the Value of Boring BizOps</a></strong></p></li></ul></li></ul><h3>CIO &amp; CTO: Infrastructure for the Agentic Era</h3><p>In 2025, CIOs moved from LLM experimentation to planning and building robust scaling layers and collaborative architectures, but readiness challenges remain.</p><ul><li><p><strong>AgentOps &amp; Scaling:</strong> Building the organisational infrastructure for agent development, deployment, and monitoring to achieve practical impact.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/agentops-the-scaling-layer-for-agentic">AgentOps: The Scaling Layer for Agentic AI in the Enterprise</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/mapping-ai-value-pathways">Mapping AI Value Pathways</a></strong></p></li></ul></li><li><p><strong>Small Models, Local Context:</strong> Making more use of safer Small Language Models (SLMs) to keep AI closer to local context, culture, and control.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/small-models-real-change-how-leaders">Small Models, Real Change: How Leaders Can Use SLMs</a></strong></p></li></ul></li><li><p><strong>Collaborative Architectures:</strong> Creating the &#8220;Context Plumbing&#8221; that allows people and machines to share a common information landscape.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/collaborative-architectures-for-agents">Collaborative Architectures for Agents, People &amp; Machines</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/a-brighter-future-for-km-building">A Brighter Future for KM: Building an AI-Enhanced Knowledge-Sharing Capability</a></strong></p></li></ul></li></ul><h3>HR: Redesigning Work and Human-AI Teaming</h3><p>HR is starting to focus on the future of human-AI collaboration and &#8220;Centaur&#8221; capabilities, plus they will need to play a role in making the implicit rules of work explicit.</p><ul><li><p><strong>Designing Centaur Teams:</strong> Reducing organisational drift by clearly defining how humans and AI agents can work together most productively.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/big-picture-leadership-techniques">Big Picture Leadership Techniques for Human-AI Teaming Readiness</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/leading-collaborative-centaur-teams">Leading Collaborative Centaur Teams</a></strong></p></li></ul></li><li><p><strong>Codifying the Invisible:</strong> Taking unspoken rules and cultural norms and turning them into explicit rulesets and guardrails for AI collaboration.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/to-protect-human-agency-codify-the">To Protect Human Agency, Codify the Invisible</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/codifying-rulesets-in-the-explainable">Codifying Rulesets in the Explainable Enterprise</a></strong></p></li></ul></li><li><p><strong>Work Designers:</strong> Encouraging a mindset shift where every employee becomes a &#8220;Work Designer&#8221; of composable workflows.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/we-can-all-be-work-designers-in-the">We Can all be Work Designers in the Composable Enterprise</a></strong></p></li></ul></li></ul><h3>Learning &amp; Development: Shared Practice &amp; Transformation</h3><p>L&amp;D is evolving from learning content production and simple training to facilitate &#8220;co-op mode&#8221; learning, practical experimentation and always-on in-the-flow AI learning systems.</p><ul><li><p><strong>Learning in &#8220;Co-Op Mode&#8221;:</strong> Accelerating adoption through shared practice rather than individual use.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/ai-learning-adoption-goes-further">AI Learning &amp; Adoption Goes Further and Faster in Co-Op Mode</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/the-rise-of-symbiotic-learning">The Rise of Symbiotic Learning</a></strong></p></li></ul></li><li><p><strong>Building the Future One Agent at a Time:</strong> L&amp;D teams exploring agentic tools  to address learner needs while staying grounded in practice.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/building-the-future-of-organisational">Building the Future of Organisational Learning One Agent at a Time</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/learning-in-the-age-of-agents">Learning in the Age of Agents</a></strong></p></li></ul></li><li><p><strong>Avoiding &#8220;Workslop&#8221;:</strong> Moving away from low-quality AI filler by focusing on the &#8220;work&#8221; of organisational transformation.</p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/doing-the-work-why-learning-is-key">Doing the Work: Why Learning is Key to Agentic AI Success</a></strong></p></li></ul></li></ul><h2>Bringing it all Together</h2><p>To bring this all together - and also bring it to life for busy leaders and leadership teams - requires the knowledge and experience to address specific local conditions and challenges. But we are also developing repeatable learning interventions and leadership techniques that people can pick up and run with.</p><p>If you would like to learn more about this work, please get in touch.</p><p>For now, here is one simple method to bring together function leads to build a common map that can guide AI-enabled capability development, and which can achieve results in a single 30-day sprint: </p><ul><li><p><strong><a href="https://academy.shiftbase.info/p/enterprise-ai-adoption-requires-connected">How leaders working together with a simple Map &#8594; Change &#8594; Learn loop can turn scattered AI experiments into living, organisation-wide capabilities</a></strong></p></li></ul><h2>Further Reading</h2><ul><li><p><strong><a href="https://academy.shiftbase.info/archive">Plunder our archives by topic area</a></strong></p></li><li><p><strong><a href="https://academy.shiftbase.info/p/deep-dives">Revisit our practical guides and deep dives</a></strong></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Can AI Help Reverse the Oversimplification of Management?]]></title><description><![CDATA[For leaders to be programmers of their organisational OS, they need AI's help in embracing complexity and reality, rather than relying on simplistic data proxies]]></description><link>https://academy.shiftbase.info/p/can-ai-help-reverse-the-oversimplification</link><guid isPermaLink="false">https://academy.shiftbase.info/p/can-ai-help-reverse-the-oversimplification</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 23 Dec 2025 16:36:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!83X4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As we look forward to a new year, it is worth zooming out momentarily from the frenetic race between AI models to focus on some of the enablers, blockers and wider changes that will determine whether organisations are able to use this technology effectively.</p><p><strong><a href="https://www.duperrin.com/english/2025/12/08/impacy-ai-transformation-bcg-mckinsey/">We now have a few studies</a></strong> that point to <strong><a href="https://openai.com/index/the-state-of-enterprise-ai-2025-report/">good but not great marginal efficiencies</a></strong> when using Generative AI inside existing business structures and management systems to speed up existing tasks. But whilst current LLMs and SLMs are more than good enough to support intelligent operations, the transformational potential of agentic AI depends upon readiness efforts in areas from technical and data infrastructure to <strong><a href="https://xer0pn.medium.com/the-organizational-blindness-of-enterprise-ai-d1eec7bfdb29">organisation mapping</a></strong>, process discovery and explainability, and in particular, local training and context management.</p><p>At the infrastructure level, there are welcome signs that <strong><a href="https://stackoverflow.blog/2025/12/08/the-shift-in-enterprise-ai-what-we-learned-on-the-floor-at-microsoft-ignite/">some of the technical enablers for agentic and enterprise AI are getting more attention</a></strong>, including interoperability with protocols such as MCP and a <strong><a href="https://simonwillison.net/2025/Dec/19/agent-skills/">standard approach to defining agent skills</a></strong>.</p><h2>Leaders as Programmers</h2><p>But where is the shift in leadership and management that will be required to re-invigorate our tired bureaucracies and create intelligent AI-augmented organisational operating systems?</p><p><strong><a href="https://academy.shiftbase.info/t/worldbuilding">We have written about world-building this year as an important emerging leadership skill</a></strong>, and we are finding this to be an accessible and useful frame for leadership development in organisations seeking to pursue enterprise AI.</p><p><strong><a href="https://tomtunguz.com/operational-analytical-context-databases/">Tomasz Tunguz recently shared an interesting musing on context databases</a></strong> and why they are needed if we want to move away from brittle, deterministic automation to fully exploit the capabilities of enterprise AI.</p><p>Making the rules of the road explicit is a key management activity as we progress towards programmable organisations, and this is why we include <em><strong>leaders as programmers</strong></em> within our leadership development programmes and workshops. This is not the old idea that &#8220;leaders should learn Python&#8221;, but the realisation that clearly stated high-level goals and instructions linked to all their necessary context is probably how the programmable organisation will be operated and guided in the future.</p><p>We have argued for a long time that <em>literate</em> leadership (writing things down, collating knowledge and curating data) beats <em>performative</em> leadership (meetings, presence and a focus on simplistic decision-making) over the long run. Those leaders who have written things down and encouraged their teams to document their work in wikis, collaboration systems and knowledge stores will have a huge advantage, and the resulting content will give context databases a big head start.</p><p>All of this has major implications for L&amp;D functions, and we are starting to see some of them trying to re-define their value proposition from content delivery to longitudinal support and product development.</p><h2>How Can AI Make Management Smart Again?</h2><p>If we are to develop the leadership organisations will need in the AI era, perhaps we should start by asking where things went wrong in the previous one.</p><p>Dan Davies&#8217; excellent book <strong><a href="https://www.goodreads.com/book/show/211161687-the-unaccountability-machine">The Unaccountability Machine</a></strong> provides some clues, and also hints at ways in which AI might make the theoretical notion of management cybernetics feasible as a practical way to run at least the basic functions of a complex organisation.</p><p>The book begins by looking at <em>accountability sinks</em> - structures (rules, algorithms, or market pressures) where decisions are delegated so deeply into a system that no individual human can be held responsible for the outcome - and their role in the 2008 financial crash. But it goes on to look at the way reductive abstractions drove post-war economics towards Milton Friedman&#8217;s doctrine of shareholder value-maximisation, which spawned value-destroying ideas such as the leveraged buyout industry.</p><p>But Davies also looks at management failings through the lens of cybernetics and specifically <strong><a href="https://en.wikipedia.org/wiki/Viable_system_model">Stafford Beer&#8217;s Viable System Model</a></strong> (VSM), and how management&#8217;s reductive approach to complexity fell foul of Ashby&#8217;s <strong><a href="https://en.wikipedia.org/wiki/Variety_(cybernetics)">Law of Requisite Variety</a></strong> (for a manager to control a system, their &#8220;regulatory variety&#8221; must match the &#8220;variety of the system&#8221; they are managing). By using attenuators to simplify their picture of the firm, such as share price or quarterly sales reports, managers have made the organisation blind to all the other important data and signals necessary to guide strategy; plus, the signals they do use for strategy tend to be lagging indicators.</p><p>It might seem that AI and the algorithmic era will make this situation even worse and therefore we should return to an imagined &#8216;good old days&#8217; of personal, accountable management. But in fact, enterprise AI offers better ways to cope with increasing complexity, and therefore a way to embrace it positively.</p><p>On the attenuation question, agentic AI enables us to maintain a detailed, objective picture of real-time operations within even a large, complex organisation, meaning the variety of the control system is able to match the variety of observable reality. If leaders really are too busy (or lacking the knowledge) to engage with reality, then AI is also pretty good at summarising information for them, but without needing to attenuate / throw away a lot of the richness as late Twentieth Century management tended to do.</p><blockquote><p><em>A modern artificial intelligence system &#8211; a transformer recurrent neural network can take a large block of text and summarise it quickly. It can also expand a short instruction into a longer explanation. It&#8217;s practically designed for facilitating two-way communication between a mass audience and a smaller decision-making system. It would really be a generational shame if we ended up once more just using it to make our existing structures work faster &#8211; like bringing back Shakespeare, Machiavelli and Napoleon and setting them to work designing tax forms.</em></p></blockquote><p>Anybody who has seen an important, expert-designed project or proposal for action reduced to kindergarten images and bullet points so that a senior leader with the attention span of a goldfish can &#8220;decide&#8221; will surely welcome the fact that we can summarise without attenuation.</p><p>Agentic AI can also help eliminate accountability sinks in several ways. For example, processes and rules of the road no longer need to be oversimplified to such an extent that people can be trained to follow them. Instead, we can write as many rules and exception handlers as we like, and leave it to the agents to ensure they follow them, with human oversight of the overall outcomes and the system. Plus, by automating the drudgery of basic work coordination (what Beer called System 2 in the VSM), people can spend more time focusing on big picture questions such as identity and purpose (System 5), which means Instead of being cogs managing spreadsheets, managers can return to being architects of the organisation&#8217;s mission.</p><p>Of course, this works only if we design our structures and rules with this in mind. The machine is only as good as what it &#8216;cares&#8217; about. If we use AI to automate the same narrow goal of &#8220;maximising shareholder returns at any cost&#8221;, we will simply build a faster, more efficient &#8216;Unaccountability Machine&#8217;.</p><blockquote><p><em>&#8220;The purpose of a system is what it does.&#8221;</em> - Stafford Beer</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!83X4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!83X4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 424w, https://substackcdn.com/image/fetch/$s_!83X4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 848w, https://substackcdn.com/image/fetch/$s_!83X4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 1272w, https://substackcdn.com/image/fetch/$s_!83X4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!83X4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic" width="1456" height="970" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:970,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1090009,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/182434126?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!83X4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 424w, https://substackcdn.com/image/fetch/$s_!83X4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 848w, https://substackcdn.com/image/fetch/$s_!83X4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 1272w, https://substackcdn.com/image/fetch/$s_!83X4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a83f090-5b9d-41ff-9249-323169f2d53b_3848x2564.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Life in the PPT/XLS factory (with apologies to L.S.Lowry)</figcaption></figure></div><h2>&#8220;Nine to five, for service and devotion&#8230;&#8221;</h2><p>One issue for both society and business that will start to assert itself in 2026 is the question of jobs. Recent job surveys reveal a bifurcation of the labour market. While mass unemployment has not yet materialised, the data shows that AI is hollowing out entry-level roles while simultaneously exacerbating a shortage of high-skilled talent.</p><p><strong><a href="https://simonwillison.net/2025/Dec/7/cory-doctorow/">AI critics such as the wonderful Cory Doctorow</a></strong> argue that AI is being sold on the promise of job cuts and cost savings, and that none of the surplus it produces will be returned to workers:</p><blockquote><p><em>The growth narrative of AI is that AI will disrupt labor markets. I use &#8220;disrupt&#8221; here in its most disreputable, tech bro sense.</em></p><p><em>The promise of AI &#8211; the promise AI companies make to investors &#8211; is that there will be AIs that can do your job, and when your boss fires you and replaces you with AI, he will keep half of your salary for himself, and give the other half to the AI company.</em></p></blockquote><p>But I think there is some win-win potential in the expected impact of AI on jobs. As Azeem Azhar shared in his <strong><a href="https://www.exponentialview.co/p/2025-in-25-stats?utm_source=substack&amp;publication_id=2252&amp;post_id=182322039&amp;utm_medium=email&amp;utm_content=share&amp;utm_campaign=email-share&amp;triggerShare=true&amp;isFreemail=false&amp;r=9dv58&amp;triedRedirect=true">2025 in 25 stats</a></strong> summary, for the first time in 22 years, work-life balance has overtaken pay as the most important job factor, according to <strong><a href="https://fortune.com/2025/11/17/work-life-balance-outranked-pay-top-perk-peoeple-choosing-a-job/">Randstad&#8217;s 2025 Workmonitor report</a></strong>. Indeed, perhaps the entire industrial era concept of jobs is a form of neo-feudalism that we should leave behind, assuming we can find other ways to fund basic needs.</p><p>As Antonio Melonio argues in a provocative post entitled <strong><a href="https://www.thepavement.xyz/p/the-era-of-jobs-is-ending">The era of jobs is ending</a></strong>:</p><blockquote><p><em>The system we built around jobs&#8212;as moral duty, as identity, as the only path to survival&#8212;is about to collide with machines that can perform huge chunks of that &#8220;duty&#8221; without sleep, without boredom, without unions, without pensions.</em></p><p><em>You can treat this as a threat.</em></p><p><em>Or as a once-in-a-civilization chance to get out of a religion that has been breaking us, grinding us down, destroying us for centuries.</em></p></blockquote><p>But at the same time, we have the opportunity to reshape those jobs that are not at risk through finding creative and human ways to combine people and AI. <strong><a href="https://www.mckinsey.com/mgi/our-research/agents-robots-and-us-skill-partnerships-in-the-age-of-ai">There is a lot of thought going into the skills needed for people to make the most of this</a></strong>, and people are starting to explore <strong><a href="https://openreview.net/forum?id=Yhqa8Ljzrj">how to quantify human-AI synergy</a></strong> to find the optimum collaboration. In a situation where skilled workers are at a premium, this all suggests that some jobs will get better and become more interesting.</p><p>But let&#8217;s not pretend that this will not be very disruptive for many people&#8217;s lives, especially in a world where housing has become an asset class rather than a basic right. This will need a policy response as well as businesses taking some responsibility for the gravity of the shift away from full employment as a realistic goal.</p><p>I already advise my mentees only to consider taking a job as a stepping stone to better things, and to avoid getting trapped in employment whilst developing their own independent income sources; I expect employers will need to work hard to sell the idea of a 9-5 to the next generation.</p><p>Finally, on the subject of work, we will hit the pause button and return in the new year. I have a large section of the very best Tuna toro and some lesser known Spanish red wines that are crying out for experimentation.</p><p>But for now, we wish you all a wonderful festive season and a happy new year.</p>]]></content:encoded></item><item><title><![CDATA[Big Picture Leadership Techniques for Human-AI Teaming Readiness]]></title><description><![CDATA[Practical Steps for Leaders to Reduce Drift and Build a Coherent Environment for Centaur Teams]]></description><link>https://academy.shiftbase.info/p/big-picture-leadership-techniques</link><guid isPermaLink="false">https://academy.shiftbase.info/p/big-picture-leadership-techniques</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 16 Dec 2025 15:03:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lj40!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<pre><code><strong>This is Part 2 of our exploration of Human&#8211;AI Teaming Readiness. </strong>In Part 1, we defined the coordination problem that emerges when humans and agents share the same work environment, and introduced <strong>Align &#8594; Bound &#8594; Learn</strong> as a lightweight system for designing shared context, boundaries and learning loops for centaur teams. In this edition, we move up a level to look at <strong>how leaders maintain the integrity of the collaboration world itself</strong>, through the Collaboration Health View and by treating Human&#8211;AI Teaming Readiness as a core organisational capability, not a local experiment. If you missed Part 1, you can read it <strong><a href="https://academy.shiftbase.info/p/leading-collaborative-centaur-teams">here</a></strong>.</code></pre><p>Many organisations have spent 2025 pouring money into AI licences, copilots and orchestration platforms, often long before they have invested the leadership attention required to make their processes and work explainable to AI agents.</p><p>The result is a familiar disconnect: enormous budget allocation at the technology layer, but without enough time spent articulating the rules, trade-offs, priorities and judgment criteria that govern how work happens.</p><p>Even without agents in the loop, many leaders are already feeling the cost of this missing context:</p><ul><li><p>Teams interpret the same strategic message in different ways.</p></li><li><p>Boundaries of autonomy shift depending on who is in the room.</p></li><li><p>Quality thresholds vary wildly between functions.</p></li><li><p>Escalations proliferate because no one is sure who owns what under changing conditions.</p></li></ul><p>These symptoms are not new. They are signals that the <em>world</em> in which teams collaborate is weaker or more fragmented than anyone realises. World-building may sound abstract, but leaders practise it every day through the expectations they set, the meanings they reinforce, the boundaries they uphold, and the language they use.</p><p>This is the part of leadership that has always held organisations together.</p><p>But as mixed-intelligence work begins to spread, the consequences of weak or drifting worlds are becoming impossible to ignore. Context that was once tacit must now be made explicit. Coherence that once emerged through proximity now requires deliberate maintenance. And coordination that once relied on shared human intuition must be authored in a form that both humans and machines can recognise.</p><h2><strong>Seeing The Drift Problem</strong></h2><p>Even in organisations that have spent years improving team-level alignment, intent and ways of working through agile transformation, coherence often collapses the moment those teams interact with the wider hierarchy.</p><p>Teams may establish:</p><ul><li><p>clear intent,</p></li><li><p>shared priorities,</p></li><li><p>strong decision principles,</p></li><li><p>and stable rituals for coordination.</p></li></ul><p>Yet when they meet the rest of the organisation, the steering structures, budgeting cycles, risk gates, and inherited decision norms, the world those teams have built can evaporate almost instantly. Suddenly:</p><ul><li><p>vocabulary no longer matches,</p></li><li><p>priorities are interpreted differently,</p></li><li><p>escalation paths contradict team autonomy,</p></li><li><p>and decisions that made sense inside the team lose meaning outside it.</p></li></ul><p>This is already world drift, leaders are beginning to feel its symptoms long before AI enters the picture:</p><ul><li><p>the same strategy feels different in every function,</p></li><li><p>ownership becomes ambiguous the moment work crosses boundaries,</p></li><li><p>quality thresholds vary depending on which leader signs off,</p></li><li><p>and small misunderstandings accumulate into larger coordination friction.</p></li></ul><p>These are early indicators that the organisational world, the shared logic that collaboration depends on, is losing definition.</p><p>Agents will inherit the world as it is, not as leaders wish it were.</p><ul><li><p>If alignment is inconsistent, they amplify inconsistency.</p></li><li><p>If thresholds vary, they surface those variations.</p></li><li><p>If boundaries are vague, they multiply escalations or silent failures.</p></li></ul><p>AI reveals drift and friction rather than creating it. So how can leaders start to recognise the dissonance of world drift?</p><h3><strong>Technique # 1: Tracking Drift Signals</strong></h3><p><em>A micro-practice for reading the current state of your collaboration world</em></p><h4><strong>What</strong></h4><p>A 10-minute weekly observation practice that helps leaders detect the earliest signs of world drift.</p><h4><strong>Why</strong></h4><p>Because drift doesn&#8217;t appear as a crisis, it appears as small inconsistencies that compound quietly until coordination feels harder than it should.</p><h4><strong>Who</strong></h4><p>Any leader responsible for cross-functional work or distributed teams.</p><h4><strong>How</strong></h4><p>After a meeting or decision, ask yourself:</p><ol><li><p><strong>Vocabulary:</strong> Did we all use the same language for what &#8220;good&#8221; meant?</p></li><li><p><strong>Mental Model:</strong> Were we really working from the same understanding of the goal?</p></li><li><p><strong>Ownership:</strong> Did decision rights remain clear or shift unpredictably?</p></li></ol><p>These micro-signals tell you where your world is stable and where meaning is beginning to fragment.</p><h3><strong>Technique # 2: The Context Ledger</strong></h3><p><em>Micro-codification that strengthens the world one edit at a time</em></p><h4><strong>What it is</strong></h4><p>A running, publicly visible ledger where teams contribute one sentence each week describing a rule, boundary, trade-off or guideline that could be added to collaboration guidance, system prompts and agent skills to improve the way people and agents operate.</p><h4><strong>Why it matters</strong></h4><p>Most context failures come from missing or inconsistent meaning.</p><p>Leaders think they have alignment. In reality, everyone is improvising differently.</p><p>A weekly cadence of micro-codification prevents drift, distributes authorship, and prepares organisations for AI agents that will need the same clarity.</p><h4><strong>Who should use it</strong></h4><p>Teams beginning to experience drift, or preparing to introduce agents.</p><h4><strong>How it works</strong></h4><p>Once a week, teams add <strong>one crisp rule</strong> such as:</p><ul><li><p>&#8220;Escalate when customer impact exceeds X, regardless of channel.&#8221;</p></li><li><p>&#8220;Risk level &#8216;high&#8217; means a downstream effect on more than two functions.&#8221;</p></li><li><p>&#8220;Agents may flag priority but humans make trade-offs between priorities.&#8221;</p></li><li><p>&#8220;A &#8216;quality issue&#8217; means deviation from X, not personal preference.&#8221;</p></li></ul><p>Leaders curate, combine, and lightly edit these into a <strong>Context Ledger,</strong> the beginnings of an organisational grammar for centaur teams.</p><p>Over time, this becomes the substrate for:</p><ul><li><p>shared operational meaning,</p></li><li><p>consistent decision-making,</p></li><li><p>and eventually, system-level prompts and agent rulesets.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lj40!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lj40!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!lj40!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!lj40!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!lj40!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lj40!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:207340,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/181786451?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lj40!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!lj40!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!lj40!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!lj40!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9ec8eb-9549-41ae-9226-e14b0f96a467_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Reading the World, Not Just the Work</strong></h2><p>These two techniques give leaders an on-ramp into world-building: a way to see drift and begin stabilise it, and to build capabilities before agents join the team.</p><p>But as organisations introduce more agents into their operating environments, something subtler begins to happen: the collaboration world develops a life of its own.</p><p>Interactions between humans and agents shape new norms. Adjacent workflows influence one another in ways no-one planned. Local learning compounds unevenly. Meaning shifts faster in some pockets than others.</p><p>Even well-aligned teams can wake up inside a context that feels quietly different from the one they thought they had created. This is the moment where leadership attention must rise above individual decisions or team routines. The question becomes one of world integrity:</p><blockquote><p><em>Is the environment that supports mixed-intelligence work holding its shape, or is drift accumulating across the structural, cultural, or experiential layers of the organisation?</em></p></blockquote><p>Leaders need a practice for reading, and lightly editing, the world itself.</p><p>That is the purpose of the Collaboration Health View, which evolves our earlier on-ramp techniques to a new level.</p><h2>The Higher-Altitude Practice: The Collaboration Health View</h2><p>In the world-building model, coherence never comes from initial design alone. It comes from continuous maintenance, from the ability to periodically rise above day-to-day activity and ask whether the world still makes sense to those living and acting within it.</p><p>The Collaboration Health View is that maintenance practice for centaur teams.</p><p>Not <em>&#8220;Is the system working?&#8221;</em></p><p>But <em>&#8220;Is the world of collaboration still coherent?&#8221;</em></p><p>This is not an operational review and not a technical audit. It is a form of world stewardship, a periodic act of sense-making that allows leaders to observe how shared meaning between humans and agents is evolving over time.</p><p>Its purpose is simple:</p><ul><li><p>to detect where coordination is strengthening,</p></li><li><p>where it is quietly drifting,</p></li><li><p>and where the language of collaboration is beginning to fracture under pressure.</p></li></ul><h3>What Leaders Look For</h3><p>At this altitude, leaders are not just reading performance. They are reading world integrity.</p><p>Signals of strengthening collaboration:</p><ul><li><p>Stable division between human judgment and agent execution</p></li><li><p>Transparent decision provenance</p></li><li><p>Deliberate, not habitual, human override</p></li><li><p>A shared vocabulary of priority, risk and quality in everyday use</p></li></ul><p>Signals of drift:</p><ul><li><p>Shadow automation outside shared boundaries</p></li><li><p>Conflicting agent behaviours across adjacent worlds</p></li><li><p>Growing reliance on post-hoc control rather than pre-emptive boundary design</p></li><li><p>Displacement of responsibility (&#8220;the system decided&#8221;)</p></li></ul><p>These are the early signs that the world&#8217;s systems and culture are beginning to slip out of alignment.</p><h3>Cadence and Outputs</h3><p>The Collaboration Health View mirrors other world-maintenance practices:</p><ul><li><p>regular, light-touch cadence</p></li><li><p>high leverage</p></li><li><p>low ceremony</p></li><li><p>persistent over time</p></li></ul><p>Its outputs are not plans or mandates. They are small edits to the world:</p><ul><li><p>boundary adjustments</p></li><li><p>vocabulary clarifications</p></li><li><p>intent realignment</p></li><li><p>narrative renewal</p></li></ul><p>This is how coherence is kept alive in an agentic environment by continuously tending the conditions that generate it.</p><h2><strong>Questions for Leaders to Test Collaboration Health</strong></h2><p>A Collaboration Health View becomes most powerful when it is grounded in lived experience. These questions help leaders sense the state of their own environment:</p><ul><li><p>Do humans and agents still appear to act from the same definition of success?</p></li><li><p>Where did we recently see an agent behave correctly in isolation but incorrectly in context?</p></li><li><p>What vocabulary has begun to fragment across teams or systems?</p></li><li><p>Where are humans overriding too often, or not often enough?</p></li><li><p>Which boundaries felt clear during design but ambiguous in practice?</p></li><li><p>Where is escalation happening too late, or too reflexively?</p></li><li><p>What small world-edits (language, boundaries, examples, narratives) would remove friction tomorrow?</p></li></ul><p>These are not diagnostic questions for technologists. They are world-reading questions for leaders.</p><p>Read on to learn how to make human&#8211;AI teaming readiness a core organisational capability to support future developments in agentic AI and automation, and what this means for the future of leadership.</p>
      <p>
          <a href="https://academy.shiftbase.info/p/big-picture-leadership-techniques">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Context Plumbing, Intent Sensing and an AI Reverse Uno on Social Media Feeds?]]></title><description><![CDATA[Enterprise AI looking a better bet than consumer AI, plus some links and ideas about how personal AI agents could help improve both of them]]></description><link>https://academy.shiftbase.info/p/context-plumbing-intent-sensing-and</link><guid isPermaLink="false">https://academy.shiftbase.info/p/context-plumbing-intent-sensing-and</guid><dc:creator><![CDATA[Lee Bryant]]></dc:creator><pubDate>Tue, 09 Dec 2025 15:30:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!rfO4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>The Enterprise Strikes Back</h2><p>Fears of the AI investment bubble potentially crashing the US stock market have abated slightly, despite AI revenues not yet looking like they will be able to repay the vast sums being invested for some considerable time. But OpenAI is clearly in a vulnerable position, and <strong><a href="https://gizmodo.com/its-code-red-week-for-openai-2000696911">Sam Altman has declared a code red</a></strong> in response to the threat posed by Google.</p><p>Ben Thompson frames the story in Star Wars terms, with <strong><a href="https://stratechery.com/2025/google-nvidia-and-openai/">OpenAI and NVidia having reached the Empire Strikes Back stage of the hero&#8217;s journey</a></strong> now that Google has seemingly re-asserted its leading position across LLMs, AI apps and hardware. But whilst Nvidia is hoovering up money by selling chips, OpenAI is blowing cash in the opposite direction, and will need to come up with something very special if it is to survive and thrive beyond this initial wave. Given Altman&#8217;s talk of <strong><a href="https://www.exponentialview.co/p/ev-553">superhuman persuasion</a></strong>, let&#8217;s hope their saving grace is not a consumer AI advertising arms race with Google.</p><p>Meanwhile, the case for enterprise AI being the route to real returns and wider economic and social benefit continues to grow. Nvidia&#8217;s Jensen Huang is continuing to build enterprise and industrial partnerships to open up new markets for their chips; <strong><a href="https://rollingout.com/2025/12/03/jensen-huang-explains-why-enterprise-ai/">he sees industrial use cases such as digital twins and product prototyping as bigger and more important opportunities than consumer chatbots</a></strong>.</p><p><strong><a href="https://www.exponentialview.co/p/ten-things-im-thinking-about-ai-part1">Azeem Azhar began his roundup of the state of AI three years on from the launch of ChatGPT with a look at enterprise AI</a>. </strong>He sees very positive adoption and ROI signals that suggest this field will continue to be where AI has the greatest impact. Even looking at just Generative AI, rather than the more complicated world of agentic AI that needs a degree of organisational transformation to fulfil its promise, he sees strong adoption that suggests we are looking at a J-curve of productivity impact:</p><blockquote><p><em>The best example, though, is JP Morgan, whose boss Jamie Dimon <a href="https://www.bloomberg.com/news/articles/2025-10-07/jpmorgan-s-dimon-says-ai-cost-savings-now-matching-money-spent">said</a>: &#8220;We have shown that for $2 billion of expense, we have about $2 billion of benefit.&#8221; This is exactly what we would expect from a productivity J&#8209;curve. <strong>With any general&#8209;purpose technology, a small set of early adopters captures gains first, while everyone else is reorienting their processes around the technology.</strong> Electricity and information technology followed that pattern; AI is no exception. The difference now is the speed at which the leading edge is moving.</em></p></blockquote><h2>Real vs Simulated Intelligence(s)</h2><p>But models are just part of the puzzle in building smarter organisations, and we should not lose sight of the respective strengths of human and machine intelligence.</p><p><strong><a href="https://onlydeadfish.substack.com/p/fish-food-670-ai-versus-human-reasoning?publication_id=2195351&amp;post_id=178509735&amp;triggerShare=true&amp;isFreemail=true&amp;r=9dv58&amp;triedRedirect=true">Neil Perkin has shared a good summary of a recent podcast by Dave Snowden about sense-making and the impact of AI</a></strong>, covering some of his observations about the differences between human and machine reasoning, cognition and insight. I also joined a longer webinar with Snowden and others interested in AI and complexity last week, where he made similarly useful points, so Neil&#8217;s notes saved me a job.</p><blockquote><p><em>Understanding these fundamental differences enables us to collaborate much more effectively with AI engines. LLMs can look like they have a deep understanding of a question but of course what they are really optimised for is identifying patterns and predicting the next most probable word in a sequence to mimic human-generated text. They are set up to minimise the difference from training data meaning that, by design, they <a href="https://onlydeadfish.substack.com/p/fish-food-654-what-is-ai-still-not">trend towards the average and most probable</a>.</em></p></blockquote><p>Another important difference between LLMs and human reasoning is that language is not the same as intelligence - <strong><a href="https://www.theverge.com/ai-artificial-intelligence/827820/large-language-models-ai-intelligence-neuroscience-problems">it is only one part of how people think and communicate their knowledge, as Benjamin Riley wrote for the Verve</a></strong>:</p><blockquote><p><em>LLMs are simply tools that emulate the communicative function of language, not the separate and distinct cognitive process of thinking and reasoning, no matter how many data centers we build.</em></p></blockquote><p>If we mistake large language models and their predictive abilities as intelligence, then we risk denuding our own creative and cognitive superpowers. But perhaps if we use these stochastic parrots in more creative ways, they could play a role in helping us improve our own thinking, rather than just outsourcing it. <strong><a href="https://advait.org/talks/sarkar-2025-tedai-vienna/sarkar_2025_TEDAI_AI_as_Tool_for_Thought_V1.pdf">Advait Sarkar posed this question in a recent talk on behalf of Microsoft Research, and concluded that the idea has potential merit</a></strong>:</p><blockquote><p><em>You can demonstrably reintroduce critical thinking into AI-assisted workflows. You can reverse the loss of creativity and enhance it instead. You can build powerful tools for memory that enable knowledge workers to read and write at speed, with greater intentionality, and remember it too. It turns out, with the right principles of design, you can build tools that are the best of both worlds: applying the awesome speed and flexibility of this technology to protect and enhance human thought.</em></p></blockquote><p>It would be good to see some practical applications of this idea in our use of GenAI within organisations, and especially for leaders.</p><h2>Deriving Context &amp; Intent Needs Better Data</h2><p>Another point Dave Snowden makes is that training data is ultimately more valuable and important than the individual models trained on it.</p><p>This raises questions of digital sovereignty for any organisation or state trying to use AI without becoming dependent on AI platform providers like OpenAI. What should you own? What can you buy or rent? What should you build?</p><p>If the current trajectory holds, it looks like open models will be commoditised and the real value will lie in data, world models and the apps and agents we build on top of the models.</p><p>But whilst we can use large historical data for training models, the operational needs of context engineering mean that this category of data should ideally be recent, atomic and fluidly connected, so that it can be used in different ways.</p><p>Matt Webb is thinking about this from the point of view of discerning user intent, and he uses the term <strong><a href="https://interconnected.org/home/2025/11/28/plumbing">context plumbing</a> </strong>to describe the complex task of integrating lots of different data feeds to create context in close to real-time. He goes on to get quite excited about <strong><a href="https://interconnected.org/home/2025/12/05/training">the potential to derive seed training data from popular platforms and marketplaces, and then apply agentic AI coding loops to fulfil the opportunities identified in the data</a></strong> (at least I think that&#8217;s what he&#8217;s saying - see what you think).</p><p>It is worth reading these brain dumps alongside <strong><a href="https://substack.com/@technologik/p-174819339">S&#233;b Krier&#8217;s recent essay Coasian Bargaining at Scale</a></strong>, which postulates that personal agents (armed with your own context and intent) could do a better job of reducing transaction costs and other frictions in distributed negotiations compared to top-down approaches to navigating and balancing competing interests:</p><blockquote><p><em>This is the essence of the work of Nobel laureate Ronald Coase, who argued that if bargaining were cheap and easy, a polluter and their neighbor could strike a private deal without any need for regulation. Of course sometimes some pollution would still happen, but the payoff to the neighbor would ensure that both parties are better off than the zero pollution or no-limits pollution counterfactuals. The tragedy is not the existence of the conflict, but the transaction costs that prevent these mutually beneficial deals from being discovered and executed. It&#8217;s also the lesson from Elinor Ostrom, who documented how real-world communities successfully govern shared resources like fisheries and forests through their own intricate local rules.</em></p></blockquote><p>It is an interesting idea, and one that could help shape AI-enabled governance in the future.</p><p>In the context of enterprise AI, we probably need to dig deeper into how we can derive, generate or synthesise training data specific to an organisation&#8217;s work to create world models and context that are rich enough to enable agentic AI operations, and perhaps even the kind of negotiated outcomes and compromises that S&#233;b Krier has in mind.</p><p>This is not just a quantity question; it is also about how we structure and organise that data. Microsoft are doing some work on the semantic layer that helps people and agents make sense of data with what they are calling <strong><a href="https://www.directionsonmicrosoft.com/cio-talk-microsoft-gets-iq/">Microsoft IQ, which is intended to bring intelligent capabilities to Fabric, Microsoft 365, and Azure AI Search</a></strong>.</p><p>Another angle on harnessing data intelligently is to democratise access to it, so that more people can help shape it, <strong><a href="https://diginomica.com/atlassian-acquires-secoda-democratize-enterprise-data-analysis-business-teams">and that is what Atlassian appear to be targeting with their acquisition of data cataloguing tool Secoda</a></strong>.</p><h2>Could Agentic AI Play the Reverse Uno Card on Social Media?</h2><p>S&#233;b Krier&#8217;s piece is another reminder that personal agents are likely to emerge as solutions to many of the coordination challenges that led us down the perilous path of large-scale platforms and algorithmic sharing.</p><p>I am in Copenhagen right now at the pre-launch gathering of a bold project to rebuild Europe&#8217;s social platforms. It aims to build on the energy and creativity that we were all so excited about in the early 2000&#8217;s before Facebook and the big US platforms exploited our human need for connection to create ad-funded clickbait farms that have harmed our societies and democracies. Just today, the Guardian wrote about <strong><a href="https://www.theguardian.com/media/2025/dec/09/youth-movement-digital-justice-spreading-across-europe">a growing movement of young people across Europe seeking to reclaim their lives from big tech platforms</a></strong>, and this trend looks set to grow.</p><p>Within the Matrix world of attention farming, we have seen the bad things that AI can do: algorithmic feeds, emotional manipulation, fake content, fake people, and so on. But what if it could also be part of re-humanising our connection with each other?</p><p>There is a whole (small) world out there of people sharing their passions in niche social networks and communities, subreddits, discords or group chats. But the nature of scale-free networks and network effects means that Whatsapp, Facebook, Twitter, etc are still the easiest option for many people and groups in Europe just because that&#8217;s where their friends or families are to be found.</p><p>But what if we go back to some of those early social network ideas such as federation, interoperability and <strong><a href="https://en.wikipedia.org/wiki/The_Intention_Economy">the intention economy</a></strong> to play a reverse Uno on algorithmic feeds? If everybody has their own discoverability and curation agent that pulls from multiple networks, communities and messaging platforms to create a personal social feed, then we don&#8217;t need to all be on the same platform. If I can tell my agent to keep me updated with all my interests and groups, from local news to hobbies and political debates, and handle the messy details of logging in and aggregating content, then perhaps we could help sustain the safer, more human-scale small world networks that are out there already under the radar. Ever the optimist!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rfO4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rfO4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rfO4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:630307,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/181129417?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rfO4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!rfO4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1abd47fb-62ab-44e0-ad3b-fffd3e733df6_2816x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[Leading Collaborative Centaur Teams]]></title><description><![CDATA[How Leaders Design the Language of Collaboration Between People and Agents.]]></description><link>https://academy.shiftbase.info/p/leading-collaborative-centaur-teams</link><guid isPermaLink="false">https://academy.shiftbase.info/p/leading-collaborative-centaur-teams</guid><dc:creator><![CDATA[Cerys Hearsey]]></dc:creator><pubDate>Tue, 02 Dec 2025 15:38:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!NOLC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<pre><code><strong>This is Part 1 of a two-part exploration of Human&#8211;AI Teaming Readiness. In this edition, we define the coordination problem that emerges when humans and agents share the same work environment, and why this is becoming a core leadership design challenge. Part 2 will introduce a practical technique that senior leaders can use to deliberately keep collaboration aligned at the enterprise level.</strong></code></pre><p>Organisations are in the early stages of introducing new forms of intelligence into environments that were designed exclusively for human collaboration, which means teams are no longer just coordinating across functions and geographies, but also across fundamentally different types of actors, with different modes of perception, speed, memory and agency.</p><p>What can leaders do to ensure this works smoothly and successfully?</p><p>AI adoption today is still focused on tools. Which copilot should we deploy? Which agents should we orchestrate? Which platforms should we integrate?</p><p>But the next question is how coordination actually works when humans and agents share the same operational space. Because coordination is not achieved through tools alone. It depends on shared intent, mutual expectations, clear boundaries of authority, and a common language for what &#8220;good&#8221; looks like. It depends on trust, not blind trust, but trust that is continually calibrated through feedback and shared understanding. And it depends, above all, on context.</p><p>Adding agents into workflows that already struggle with role clarity and decision ownership can exacerbate existing problems. If we ask people to &#8220;delegate&#8221; to systems whose boundaries of action are poorly defined, some respond by over-relying on (and over-trusting) automation, whilst others respond with defensive scepticism, overriding systems by default. In both cases, coordination degrades because the shared language and context of work is missing or inconsistent.</p><p>What emerges is not seamless collaboration, but a new form of friction:</p><ul><li><p>decisions without provenance,</p></li><li><p>actions without clear ownership, and</p></li><li><p>learning that fails to accumulate.</p></li></ul><p>This is a coordination design problem that reflects the need for leadership to evolve towards a new craft: <em><strong>the design of the conditions under which humans and machines can act together without constantly pulling the system out of shape.</strong></em></p><p>In mixed-intelligence teams, coordination is no longer something that simply &#8220;happens&#8221; through informal norms and shared human intuition. It must be deliberately authored. The language of collaboration needs to be designed.</p><p>That language is what we mean by context.</p><p>This edition <strong><a href="https://academy.shiftbase.info/p/a-leaders-guide-to-world-building">builds directly on our recent exploration of world-building as a leadership capability for the agentic era</a></strong>, where we argued that organisations must be designed as coherent worlds of physics (systems), culture (meaning), and geography (experience), and not just as collections of tools and workflows.</p><p>Let&#8217;s zoom in from the design of the world to the design of collaboration that happens inside it; specifically, how leaders shape the language through which humans and AI work together as centaur teams.</p><h2>From Tool Use to Teaming Readiness</h2><p>Collaboration is not a feature of a toolset. It is a property of an environment.</p><p>A team does not become a centaur team because it has access to an agent. It becomes one when humans and machines can reliably coordinate their actions around shared intent, shared boundaries and shared meaning. Without that, what looks like collaboration on a process map quickly collapses into a brittle sequence of hand-offs, overrides, and shaky unwritten assumptions.</p><p>This is the shift many organisations are now stumbling into without quite realising it.</p><p>AI is no longer just something people use. In many settings, it is something that increasingly participates, sensing conditions, drafting actions, monitoring flows, surfacing options, and, in some cases, acting directly in the world. The moment AI participates in work, the question becomes <em>&#8220;What does it mean to work together?&#8221;</em></p><p>Yet most leadership doctrines, operating models and performance systems were never designed for such a question. They assume human actors with human judgment, human accountability, human learning cycles. Agents enter this landscape as something undefined: sometimes treated as a junior worker, sometimes as a calculator, sometimes as an oracle, sometimes as a risk.</p><p>The result is a quiet incoherence in how teams are being asked to relate to their machines.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NOLC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NOLC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NOLC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:825833,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://academy.shiftbase.info/i/180495414?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NOLC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!NOLC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aca2e5d-186d-4198-8763-3ea48be17bb8_1536x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In some places, delegation runs far ahead of design. Agents are given sweeping autonomy without corresponding clarity on boundaries, escalation, or quality thresholds. In others, distrust freezes collaboration entirely, reducing AI to a glorified drafting aid despite its wider potential. Both patterns look like adoption. Neither look like teaming readiness.</p><p>Teaming readiness is something different from tool readiness.</p><p>Tool readiness asks:</p><ul><li><p>Is the technology stable?</p></li><li><p>Is it secure?</p></li><li><p>Is it integrated?</p></li></ul><p>Teaming readiness asks a different set of questions:</p><ul><li><p>Do humans and agents share a workable definition of success?</p></li><li><p>Is it clear when an agent may act, and when a human must decide?</p></li><li><p>Do people trust the system for the right reasons?</p></li><li><p>Is learning flowing both from human to machine and from machine back to human?</p></li></ul><p>These are not questions for IT alone. They are questions of leadership design.</p><p>Moving from tool use to true teaming requires the deliberate shaping of roles, responsibilities, feedback loops, and the language through which work is coordinated. In other words, through the design of context as a shared operating grammar.</p><p>Until that grammar exists, organisations will continue to experience a familiar paradox: impressive local gains from AI deployment, alongside growing systemic fragility in how work actually holds together.</p><p>Leadership in mixed-intelligence environments therefore shifts in a subtle but fundamental way. The task is no longer simply to deploy capability. It is to make collaboration itself legible, stable and learnable at the boundary between human and machine.</p><p>That is what we mean by Human&#8211;AI Teaming Readiness.</p><h2>What &#8220;Context&#8221; Means in a Human&#8211;AI Team</h2><p>In human&#8211;AI collaboration, <em>context</em> is often treated as a technical concern: prompts, data access, memory, retrieval. These are important foundations. But they are not what makes collaboration work.</p><p>In a teaming environment, context is also <em>coordination.</em></p><p>It is the shared frame that allows different kinds of intelligence to act in relationship to one another without constant supervision or repair. It is what tells a human when to trust a system&#8217;s output, when to challenge it, and when to override it. It is what tells an agent not just what it can do, but how its actions sit within a wider field of purpose, risk and responsibility.</p><p>In world-building terms, this is the moment where physics, culture and geography stop being abstract layers and become the operating language of daily collaboration:</p><ul><li><p>The <strong>System layer</strong> provides the physics: rules, data, contracts and constraints that make certain actions possible and others impossible.</p></li><li><p>The <strong>Culture layer</strong> provides the meaning: norms, values, stories and judgments that shape what <em>should</em> happen.</p></li><li><p>The <strong>Experience layer</strong> provides the geography: the interfaces, workflows and spaces through which both humans and agents navigate the world.</p></li></ul><p>Context is the braided fabric of all three.</p><p>In this sense, it is not a static asset. It is a living operating language, made up of several intertwined elements:</p><ul><li><p><strong>Shared intent:</strong> a common understanding of what the work is ultimately trying to achieve, beyond the task at hand.</p></li><li><p><strong>Boundaries of authority:</strong> clarity on when an agent may act autonomously, when it must recommend, and when a human must decide.</p></li><li><p><strong>Decision vocabulary:</strong> stable definitions of what &#8220;good&#8221;, &#8220;acceptable&#8221;, &#8220;escalate&#8221;, &#8220;complete&#8221;, or &#8220;exception&#8221; actually mean in practice.</p></li><li><p><strong>Quality thresholds:</strong> what level of confidence, evidence, or validation is required before action is taken.</p></li><li><p><strong>Risk posture:</strong> how much uncertainty the team is willing to tolerate in different contexts.</p></li><li><p><strong>Cultural norms of judgment:</strong> whether challenge is expected or discouraged, whether speed outweighs precision, whether learning is prioritised over optimisation.</p></li></ul><p>Taken together, these form the grammar of collaboration. Without this grammar, human&#8211;AI interaction defaults to two unstable extremes. Either humans over-trust systems, surrendering judgment too early and too broadly. Or they under-trust them, turning agents into little more than sophisticated drafting assistants. In both cases, potential is left unrealised and risk is misunderstood.</p><p>The difficulty is that much of this context is normally held tacitly in human teams. It lives in shared experience, informal norms, and unspoken expectations. When agents enter the system, that tacit layer is suddenly exposed. What was once quietly inferred now has to be made explicit if coordination is to hold.</p><p>This is why many early centaur team experiments feel awkward at first. The introduction of an agent acts like a mirror, reflecting back the vagueness that already existed within the team.</p><p>To become teaming-ready, organisations must therefore do more than supply agents with data and access. They must author the world in which those agents will operate, as physics, as culture, and as navigable experience.</p><p>And it is this shared language that leadership is now being asked to design.</p><p>Let&#8217;s look at a core leadership technique for Human&#8211;AI Teaming Readiness we call <strong>Align &#8594; Bound &#8594; Learn</strong>, which is a lightweight system for:</p><ul><li><p>aligning intent,</p></li><li><p>designing authority boundaries, and</p></li><li><p>ensuring that learning compounds rather than fragments.</p></li></ul>
      <p>
          <a href="https://academy.shiftbase.info/p/leading-collaborative-centaur-teams">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>