<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Agents Decoded]]></title><description><![CDATA[AI Agent academic, tech and business reviews and analysis - for builders and thinkers. ]]></description><link>https://www.agentsdecoded.com</link><image><url>https://substackcdn.com/image/fetch/$s_!hDWs!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F65db99ed-e99b-43b2-bcd5-035e248384e9_1024x1024.png</url><title>Agents Decoded</title><link>https://www.agentsdecoded.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 11:16:49 GMT</lastBuildDate><atom:link href="https://www.agentsdecoded.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Istos srl]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[agentsdecoded@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[agentsdecoded@substack.com]]></itunes:email><itunes:name><![CDATA[Ronald Ashri]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ronald Ashri]]></itunes:author><googleplay:owner><![CDATA[agentsdecoded@substack.com]]></googleplay:owner><googleplay:email><![CDATA[agentsdecoded@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ronald Ashri]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[CIAO👋 Core Components: Orchestration — Coordinating Conversations, Interfaces, and Agents]]></title><description><![CDATA[Orchestration requirements come from two places: the problem you're solving and the architecture you chose. Understanding both is essential to getting coordination right.]]></description><link>https://www.agentsdecoded.com/p/ciao-core-components-orchestration</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-core-components-orchestration</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Tue, 31 Mar 2026 12:46:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ed1Z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div><hr></div><p>This article is part of a series exploring how would an architectural framework for AI-applications look like. We called it <em><a href="https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework">CIAO&#128075; </a>, capturing four core components: Conversations, Interfaces, Agents and Orchestration.</em> Previous instalments have covered <em><a href="https://www.agentsdecoded.com/p/ciao-core-components-conversations">Conversations</a></em>, <em><a href="https://www.agentsdecoded.com/p/ciao-core-components-interfaces">Interfaces</a></em>, and <em><a href="https://www.agentsdecoded.com/p/ciao-core-components-agents-how-to">Agents</a>.</em></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ed1Z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:156799,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/192696025?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Ed1Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fade66650-8185-4ba5-8e4d-13a8bebb3e2e_2560x1440.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>With Conversations, Interfaces, and Agents each explored in depth, we turn to the component that connects them. Orchestration often gets treated as implementation wiring, the thing you figure out after the interesting design decisions have been made. That can lead to significant challenges later, when coordination gaps surface in production rather than in design.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Agents Decoded! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kRms!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kRms!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 424w, https://substackcdn.com/image/fetch/$s_!kRms!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 848w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1272w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png" width="1456" height="733" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64668932-6851-453b-ab63-e106d17b4269_1497x754.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:733,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kRms!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 424w, https://substackcdn.com/image/fetch/$s_!kRms!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 848w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1272w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Orchestration is a distinct architectural concern with its own requirements, its own design space, and its own failure modes. In the same way that Conversations merit deliberate design rather than defaulting to whatever message structure an API offers, Orchestration merits deliberate analysis rather than emerging as a byproduct of agent and interface decisions.</p><p>A working definition: </p><blockquote><p><strong>Orchestration is the runtime coordination that determines what happens next, which component handles it, and how information flows between them.</strong></p></blockquote><h2>Two Sources of Orchestration Requirements</h2><p>Following the same principle that guides the <a href="https://www.agentsdecoded.com/p/ciao-core-components-agents-how-to">Agents</a> article, namely that we need to understand the problem before reaching for solutions, we can identify two distinct sources of orchestration requirements.</p><p><strong>Domain-derived requirements</strong> are constraints that exist in the problem space before any architecture decisions are made. They are discovered through domain analysis, not created through design. These include:</p><ul><li><p><strong>Sequencing constraints</strong>: dependencies inherent in the domain (fraud assessment must precede payout authorisation)</p></li><li><p><strong>Consistency requirements</strong>: how aligned shared state must be across participants</p></li><li><p><strong>Accountability chains</strong>: traceability of decisions to responsible parties</p></li><li><p><strong>Human involvement requirements</strong>: mandated checkpoints and approvals</p></li><li><p><strong>Temporal constraints</strong>: how fast coordination itself must happen</p></li><li><p><strong>Recovery requirements</strong>: what must happen when something fails</p></li><li><p><strong>Observability requirements</strong>: what must be visible, when, and to whom</p></li><li><p><strong>Cross-component policy coherence</strong>: how tightly C, I, and A policies must align</p></li></ul><p>These are the non-negotiable elements (unless one decides to reach for core organisational re-design that is!). They set the floor for any orchestration approach.</p><p><strong>Architecture-induced requirements</strong> emerge from design decisions made across Conversations, Interfaces, and Agents. Decomposing work across multiple agents creates handoff and routing requirements. Supporting multiple interface modalities creates translation and synchronisation needs. Structuring conversations with explicit context introduces state management overhead. None of these coordination challenges exist in the problem domain. They are consequences of architectural choices.</p><p>This separation gives architects a useful design heuristic: <strong>if architecture-induced orchestration requirements grow disproportionate to domain-derived ones, the architecture may be more complex than the problem demands.</strong></p><h3>Making this concrete</h3><p>Consider an insurance claims processing application. The domain-derived orchestration requirements are substantial and exist regardless of how you build the system. Fraud assessment must complete before payout authorisation. The regulator requires a decision audit trail. Human approval is mandated above a monetary threshold. Certain document types must be verified before assessment can begin.</p><p>Now layer in architectural decisions. You choose three specialised agents: one for document intake, one for assessment, one for payout. You support both voice and text interfaces for the claimant. You maintain separate conversation threads for the customer and the adjuster.</p><p>Each of these decisions creates additional orchestration requirements. The three agents need handoff protocols. The two interface modalities need synchronisation so that a claimant who starts on the phone and switches to a web form does not lose context. The separate conversation threads need state coordination so the adjuster&#8217;s view reflects what the customer has provided.</p><p>The domain-derived requirements are the cost of the problem. The architecture-induced requirements are the cost of the solution. Both are real, but only the second category is within your direct control to reduce.</p><h2>A Taxonomy of Orchestration Approaches</h2><p>With requirements understood, we can examine the design space for meeting them. The primary dimension is <strong>locus of control</strong>, what decides what happens next.</p><p><strong>Zero orchestration</strong> applies when there is a single prompt, a single model call, and a single response. In essence, there is almost nothing to coordinate. This is appropriate when domain-derived requirements are minimal and the application involves one agent, one interface, and one conversation. A document Q&amp;A tool that takes a question and returns an answer is a good example. There is no sequencing, no handoff, no state to manage across components. If your requirements genuinely fit here, adding orchestration adds complexity without value.</p><p><strong>Model-intrinsic orchestration</strong> lets the model itself reason about task decomposition, sequencing, tool selection, and delegation. The model creates its own coordination logic at inference time. This is powerful and flexible, but the coordination reasoning is embedded in model behaviour. That has implications: observability is harder because the model&#8217;s routing decisions are not declared in advance. Accountability is harder because the path through the system is emergent rather than defined.</p><p><strong>Framework-imposed orchestration</strong> defines coordination structures externally. State machines, directed graphs, explicit routing rules. Agents operate within defined nodes. This offers predictability and auditability. You can inspect the coordination logic without running the system. You can verify that compliance requirements are met structurally rather than hoping the model respects them. The cost is upfront design effort and reduced flexibility.</p><p><strong>Hybrid orchestration</strong> combines the two. External structure defines phases, boundaries, and guardrails. Within those boundaries, model-intrinsic reasoning handles coordination. The insurance claims example might use framework-imposed orchestration for the overall process flow (intake &#8594; assessment &#8594; approval &#8594; payout) while allowing model-intrinsic reasoning within the assessment phase to determine which checks to run and in what order.</p><p>Two secondary dimensions further characterise the design space. </p><p><strong>Static vs dynamic topology</strong>: is the coordination pattern fixed at design time or can it adapt at runtime? </p><p><strong>Centralised vs distributed</strong>: is there a single orchestrating component or do multiple components negotiate coordination among themselves?</p><p>The right position on these dimensions is driven by requirements. High accountability, observability, and temporal constraints favour more explicit, centralised, and static approaches. Lower requirements on those dimensions create room for model-intrinsic, distributed, and dynamic coordination. Now, keep in mind that this is not a maturity ladder. Zero orchestration should not be considered primitive and hybrid is not necessarily advanced. Each fits different requirement profiles.</p><h2>Orchestration, Governance, and Policies</h2><p>Three concepts that are closely related but architecturally distinct.</p><p><strong>Orchestration</strong> is operational. It is the runtime mechanism that routes work, manages state, sequences steps, and handles failures.</p><p><strong>Governance</strong> is normative. It defines the boundaries within which orchestration operates: what is permissible, who is accountable, how compliance is verified. Governance should remain stable even when orchestration implementations change. You should be able to rework your coordination approach without renegotiating your governance model. If you cannot, the two are too tightly coupled.</p><p><strong>Policies</strong> operate at multiple levels within a CIAO application. Conversation policies govern dialogue interaction, who can speak, what can be said, what actions are permitted. These are owned and enforced by the Conversation component. Agent constraints define autonomy boundaries and permitted actions, owned by each Agent. Orchestration policies govern coordination behaviour, routing rules, retry strategies, escalation triggers, owned by the Orchestration component. Domain policies capture business rules and regulatory requirements, defined through governance and enforced across components.</p><p>The architectural principle: each component enforces its own policies, but orchestration needs visibility into the policies of other components to coordinate coherently. It cannot route a task to an agent whose autonomy level does not permit that kind of decision. It cannot initiate a conversation that violates a conversation policy. Governance is the layer that ensures consistency across these policy boundaries.</p><h2>Orchestration&#8217;s Relationship to C, I, and A</h2><p>Orchestration has a bidirectional relationship with each of the other three components, and this is what distinguishes it within the framework.</p><p><strong>Conversations</strong> generate events and state changes that inform orchestration decisions. A customer providing a key piece of information in conversation might trigger orchestration to route that information to an assessment agent. Orchestration also routes information into and across conversations, and may initiate or close conversations based on workflow needs.</p><p><strong>Interfaces</strong> produce turns that orchestration must process and route. Orchestration determines what gets rendered back and through which interface, particularly when an application supports multiple modalities. A voice interface and a text interface presenting the same conversation require orchestration to manage the translation and synchronisation.</p><p><strong>Agents</strong> receive work assignments through orchestration and return results. But the relationship is not purely top-down. Agents&#8217; autonomy levels, capabilities, and constraints shape what orchestration can ask of them. An agent with limited autonomy cannot be assigned open-ended reasoning tasks. An agent with high autonomy may complete work without checking back, which orchestration needs to account for.</p><p>This bidirectional dependency across all three components is what makes orchestration the connective tissue of a CIAO application, and why its requirements emerge from two sources rather than one.</p><h2>For Builders</h2><p>When designing orchestration for an AI-powered application:</p><ol><li><p><strong>Start by mapping domain-derived orchestration requirements.</strong> These are your non-negotiable constraints. They exist before you write a line of code and they will outlast any specific implementation.</p></li></ol><ol start="2"><li><p><strong>Track architecture-induced requirements as you design.</strong> As you make decisions across Conversations, Interfaces, and Agents, note the coordination overhead each decision creates. Ask whether it is proportionate to the problem you are solving.</p></li></ol><ol start="3"><li><p><strong>Let your accountability and observability requirements guide locus of control.</strong> If regulators need to audit decision paths, framework-imposed or hybrid orchestration is likely necessary. If you are building an internal tool with low stakes and high tolerance for variability, model-intrinsic orchestration may be sufficient.</p></li></ol><ol start="4"><li><p><strong>Be explicit about which policies live at which layer.</strong> Ambiguity about who enforces what leads to gaps, usually discovered in production. Document the policy boundaries between Conversations, Agents, and Orchestration.</p></li></ol><ol start="5"><li><p><strong>Test governance independence.</strong> Can you change your orchestration implementation without changing your governance model? If you cannot, the two may be too tightly coupled. The goal is operational flexibility within normative stability.</p></li></ol><div><hr></div><p>Orchestration requirements come from two sources: the problem domain and the architecture you choose. Keeping that distinction clear is the most useful thing an architect can do early. If your coordination overhead is primarily architecture-induced, that is a signal to simplify. If it is primarily domain-derived, the problem genuinely requires sophisticated orchestration and you should invest accordingly.</p><p>The right amount of orchestration is the minimum that satisfies both sources of requirements. No less, because the domain constraints are non-negotiable. No more, because every additional coordination mechanism is complexity you must maintain, debug, and explain.</p><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Agents Decoded! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[CIAO👋 Core Components: Agents - How to Design AI Agents by Starting with the Problem, Not the Technology]]></title><description><![CDATA[We're building sophisticated systems on shaky conceptual foundations - separating problem space from solution space helps bring order.]]></description><link>https://www.agentsdecoded.com/p/ciao-core-components-agents-how-to</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-core-components-agents-how-to</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Wed, 06 Aug 2025 07:18:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GIRS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>CIAO&#128075; is a unifying architectural framework for AI-powered applications that organizes systems into four core components: Conversations (managing dialogue state), Interfaces (handling human interaction), Agents (automated decision-makers), and Orchestration (coordinating everything).</em></p><p><em>The framework prioritizes "what" applications need to do (like understanding natural language or reasoning about actions) over "how" they're implemented (specific technologies like LLMs), allowing systems to adapt as technologies evolve.</em></p><p><em>It's designed for the new generation of software that is conversation-centric, handles multiple input/output types, and features increasing levels of automated decision-making.</em></p><p>Check out the <a href="https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework">introduction</a>, our deep dive into <a href="https://www.agentsdecoded.com/p/ciao-core-components-conversations">Conversations</a>, <a href="https://www.agentsdecoded.com/p/ciao-core-components-interfaces">Interfaces</a> or read on here for Agents.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GIRS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GIRS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GIRS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:146020,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/170153651?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GIRS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GIRS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd34485f-bb2b-4ead-9b6b-b72999795873_2560x1440.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1>Introduction</h1><p>The explosion of interest in AI agents has created an unexpected problem. In the rush to build "agentic" systems, we've adopted the terminology without developing the structured thinking to go with it. Every AI-powered application is now described as having "agents," but ask ten developers what makes something an agent and you'll get eleven different answers.</p><p>This rapid adoption has led to confusion that slows down development and complicates communication. Teams struggle to articulate what their agents actually do versus what they need them to do. Architecture decisions are made based on hype or FOMO rather than systematic analysis. </p><div class="pullquote"><p>We're building sophisticated systems on shaky conceptual foundations.</p></div><p>Agents went from a relatively quiet corner of AI research to the forefront of how we describe LLM-powered applications in less than two years. But in our eagerness to build, we've skipped the crucial step of developing a coherent framework for thinking about them. In this article, we continue fixing that by establishing a structured approach to understanding Agents as a core component of your AI-powered application.</p><h1>Why is the term agent useful?</h1><p>Let&#8217;s start here. Why do we even need to use this term? We&#8217;ve been describing applications (automated or not) without referring to agents for decades. What&#8217;s changed?</p><p>For me the core utility comes from having a better abstraction to describe your application. Using the right abstraction brings clarity, leads to more robust designs, improves communication across the team and speeds up development. We are clearly building a new type of software, with levels of automony that were not possible before so using a new abstraction is entirely sensible. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b31o!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b31o!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 424w, https://substackcdn.com/image/fetch/$s_!b31o!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 848w, https://substackcdn.com/image/fetch/$s_!b31o!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 1272w, https://substackcdn.com/image/fetch/$s_!b31o!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b31o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png" width="1456" height="643" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:643,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:218758,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/170153651?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b31o!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 424w, https://substackcdn.com/image/fetch/$s_!b31o!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 848w, https://substackcdn.com/image/fetch/$s_!b31o!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 1272w, https://substackcdn.com/image/fetch/$s_!b31o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38d5f196-5bdd-404b-a98a-650fdcbb2b0e_2136x944.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The evolution of software programming abstractions from procedural to agent-based programming. </figcaption></figure></div><p>Of course, it is not enough to just say we are using the abstraction of agents. We then need to build up a set of concepts and the relationships between those concepts that give us a coherent way of describing and comparing a range of different solutions. </p><p>As we mention in the intro a core tenet of &#128075;CIAO is that we clearly distinguish the what from the how. We need to distinguish the problem space from the solution space. The problem space is defined by what you are trying to solve. The solution space is defined by the various options in terms of how you are trying to solve it. </p><p>Let&#8217;s get going!</p><h1>Designing agents starting from the problem space</h1><p>Separating <strong>what must be solved</strong> from <strong>how we choose to solve it</strong> remains one of the simplest, most useful lenses for AI system design. By first cataloguing the <em>agent-independent</em> properties of a task and its environment, we can judge inherent difficulty before deciding whether to deploy a rules engine, a model-based planner, or a large language model. Conversely, listing <em>agent-dependent</em> properties clarifies exactly which architectural knobs we can turn to meet those demands.</p><h2><strong>Agent-Independent (Problem-Space) Dimensions</strong></h2><p>These characteristics exist before any particular agent shows up:</p><ul><li><p><strong>Task / Goal Characteristics</strong> &#8211; What is the scope, the domain, and the baseline complexity?</p><ul><li><p><em>Example: route-planning for a delivery van versus playing chess.</em></p></li></ul></li><li><p><strong>Goal Verifiability</strong> &#8211; Can success be measured objectively, or does it rely on subjective judgment?</p><ul><li><p><em>Check-mate is clear-cut; &#8220;best vacation plan&#8221; is not.</em></p></li></ul></li><li><p><strong>Environment Properties</strong> &#8211; How observable is the world? Is it deterministic or stochastic? Static or dynamic? Discrete or continuous?</p><ul><li><p><em>A chess board is fully observable and static; a stock market is partially observable and stochastic.</em></p></li></ul></li><li><p><strong>Multi-Agent Context</strong> &#8211; Are we in a multi-agent setting? Is that setting cooperative, competitive, or mixed setting?</p><ul><li><p><em>Self-driving fleets cooperate; high-frequency traders compete.</em></p></li></ul></li><li><p><strong>Baseline Temporal Constraints</strong> &#8211; What deadlines does the external world impose?</p><ul><li><p><em>The exchange&#8217;s one-second trading tick leaves little room for deliberation.</em></p></li></ul></li><li><p><strong>Intrinsic Safety &amp; Alignment Risk</strong> &#8211; How much harm could failure cause?</p><ul><li><p><em>Incorrect medical-dose planning carries far more risk than mis-classifying a meme.</em></p></li></ul></li><li><p><strong>Required Steps/Plan Complexity</strong> &#8211; Does achieving the goal follow a fixed sequence, or require dynamic adaptation?</p><ul><li><p>Assembling furniture follows instructions; navigating rush-hour traffic requires constant replanning.</p></li></ul></li></ul><p>These dimensions help us understand the fundamental difficulty of a problem before we even consider what kind of agent architecture to deploy. A task that scores high on stochasticity, has low goal verifiability, operates in a competitive multi-agent environment, and carries significant safety risks presents inherent challenges that no agent architecture can simply wish away.</p><h3>Agent-Dependent (Solution-Space) Dimensions</h3><p>Once we understand the problem space, we can examine the architectural choices that determine how well an agent might handle it:</p><p><strong>Learning Capabilities</strong> &#8211; Can the agent improve through experience, or is its behavior fixed?</p><ul><li><p>From simple parameter tuning to sophisticated meta-learning and continual learning without forgetting.</p></li></ul><p><strong>Autonomy Level</strong> &#8211; How much human oversight does the agent require?</p><ul><li><p>Following thinking similar to the <a href="https://www.synopsys.com/blogs/chip-design/autonomous-driving-levels.html">SAE levels</a> (0-5), we can reason about the level of autonomy from human-controlled with agent assistance to fully autonomous operation.</p></li></ul><p><strong>Planning Architecture</strong> &#8211; How does the agent decide what to do?</p><ul><li><p>Reactive (stimulus-response), deliberative (model-based planning), or hybrid approaches.</p></li><li><p>The choice often depends on temporal constraints from the problem space.</p></li></ul><p><strong>Coordination Capabilities</strong> &#8211; Can the agent work with others?</p><ul><li><p>Communication protocols, negotiation strategies, and team formation abilities.</p></li><li><p>Critical when the problem space includes multi-agent contexts.</p></li></ul><p><strong>Computational Architecture</strong> &#8211; What resources does the agent need?</p><ul><li><p>Edge devices with millisecond response but limited compute, or cloud-based with powerful models but network latency.</p></li><li><p>Memory requirements, energy consumption, and bandwidth constraints.</p></li></ul><p><strong>Perception and Action Capabilities</strong> &#8211; What can the agent sense and do?</p><ul><li><p>Sensor modalities (text, vision, audio, proprioceptive).</p></li><li><p>Action space (discrete choices, continuous control, communication acts).</p></li><li><p>Internal representations (symbolic, neural, hybrid).</p></li></ul><p><strong>Temporal Processing</strong> &#8211; How does the agent handle time?</p><ul><li><p>Real-time responsiveness versus batch processing.</p></li><li><p>Planning horizons from immediate reactions to long-term strategic thinking.</p></li><li><p>Memory architecture for maintaining context over extended interactions.</p></li></ul><h2>Bringing it together: Matching solutions to problems</h2><p>The power of this separation becomes clear when designing systems. Let's examine two contrasting examples:</p><h3>Example 1: Recipe Recommendation Agent</h3><p><strong>Problem Space Analysis:</strong></p><ul><li><p>Goal verifiability: Low (taste is subjective, success depends on user satisfaction)</p></li><li><p>Environment: Mostly observable, deterministic (ingredient availability, dietary restrictions are known)</p></li><li><p>Safety risk: Low (worst case: unenjoyable meal, unless handling allergies)</p></li><li><p>Temporal constraints: Relaxed (users can wait 10-30 seconds for suggestions)</p></li><li><p>Multi-agent context: None (single user interaction)</p></li><li><p>Plan complexity: Moderate (matching ingredients, techniques, time constraints)</p></li></ul><p><strong>Solution Space Decisions:</strong></p><ul><li><p>Learning: Basic preference learning from user feedback</p></li><li><p>Autonomy: Level 4 (agent suggests, user can modify)</p></li><li><p>Planning: Simple deliberative (filter &#8594; rank &#8594; present options)</p></li><li><p>Computation: Can run entirely on edge devices</p></li><li><p>Perception: Text input, possibly image recognition for ingredients</p></li><li><p>Coordination: Not needed for basic version</p></li></ul><p>This is a forgiving problem space where we can experiment with higher autonomy and simpler architectures. The low stakes allow for learning through trial and error.</p><h3>Example 2: Medical Diagnosis Agent</h3><p><strong>Problem Space Analysis:</strong></p><ul><li><p>Goal verifiability: Mixed (some conditions have clear tests, others require judgment)</p></li><li><p>Environment: Partially observable, stochastic (hidden symptoms, disease progression)</p></li><li><p>Safety risk: High (misdiagnosis can be fatal)</p></li><li><p>Temporal constraints: Moderate (emergency vs. routine care)</p></li><li><p>Multi-agent context: Cooperative (must work with healthcare team)</p></li><li><p>Plan complexity: High (differential diagnosis, test ordering, treatment planning)</p></li></ul><p><strong>Solution Space Decisions:</strong></p><ul><li><p>Learning: Continual learning to incorporate new medical knowledge</p></li><li><p>Autonomy: Level 2-3 (agent assists, human decides)</p></li><li><p>Planning: Hybrid (quick triage reactions + deliberative diagnosis)</p></li><li><p>Computation: Cloud-based for complex reasoning, edge for emergency triage</p></li><li><p>Perception: Multi-modal (text, images, lab results, vital signs)</p></li><li><p>Coordination: Must integrate with electronic health records, other clinical systems</p></li></ul><p>The contrast is striking. The recipe recommender can be playful and experimental, while the medical agent demands conservative design choices, extensive validation, and human oversight. This structured approach prevents both over-engineering (building unnecessary capabilities) and under-engineering (deploying inadequate agents). It makes trade-offs explicit: accepting lower autonomy to manage safety risks, or investing in expensive computational resources to meet temporal constraints.</p><h2>Looking ahead: The evolving landscape of agents</h2><p>As we build more sophisticated AI-powered applications, the agent abstraction becomes increasingly valuable. But its value lies not in the buzzword itself, but in the structured thinking it enables. By clearly separating what must be solved from how we choose to solve it, we can:</p><ol><li><p><strong>Communicate clearly</strong> across teams &#8211; product managers can specify problem-space requirements while engineers explore solution-space options</p></li><li><p><strong>Evaluate systematically</strong> &#8211; distinguishing between difficult problems and poorly-designed agents</p></li><li><p><strong>Evolve gracefully</strong> &#8211; as new technologies emerge (better models, new architectures), we can swap solution approaches while keeping problem definitions stable</p></li><li><p><strong>Design responsibly</strong> &#8211; explicitly considering safety and alignment risks before they become afterthoughts</p></li></ol><p>The &#128075;CIAO framework's emphasis on Agents as one of four core components (alongside Conversations, Interfaces, and Orchestration) reflects a fundamental shift in how we build software. No longer are we just coding procedures or training models &#8211; we're designing entities with increasing autonomy that must navigate complex, uncertain environments while maintaining alignment with human values and goals.</p><p>This shift requires new abstractions, new frameworks, and new ways of thinking. The agent-independent versus agent-dependent distinction is just one tool in this evolving toolkit, but it's a powerful one. It forces us to be honest about what we're asking our systems to do, and thoughtful about how we equip them to do it.</p><p>As you design your next AI-powered application, start with the problem space. Understand the inherent challenges before reaching for solutions. Then, and only then, explore the solution space with clear eyes about what capabilities your agents truly need. This discipline &#8211; separating the what from the how &#8211; will give you a head start in building the right sort of solution for the problem at hand. </p><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AgentsDecoded explores the intersection of AI agents, practical implementation, and strategic thinking. Subscribe for regular updates.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[CIAO👋 Core Components: Interfaces]]></title><description><![CDATA[In this article we discuss Interfaces and explore using the Conversational Turn as the constant across any type of Interface - forming the contract between Conversation components and Interfaces.]]></description><link>https://www.agentsdecoded.com/p/ciao-core-components-interfaces</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-core-components-interfaces</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Mon, 07 Jul 2025 16:32:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ktEb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div><hr></div><p><em>CIAO&#128075; is a unifying architectural framework for AI-powered applications that organizes systems into four core components: Conversations (managing dialogue state), Interfaces (handling human interaction), Agents (automated decision-makers), and Orchestration (coordinating everything).</em></p><p><em>The framework prioritizes "what" applications need to do (like understanding natural language or reasoning about actions) over "how" they're implemented (specific technologies like LLMs), allowing systems to adapt as technologies evolve.</em></p><p><em>It's designed for the new generation of software that is conversation-centric, handles multiple input/output types, and features increasing levels of automated decision-making.</em></p><p>Check out the <a href="https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework">introduction</a>, our deep dive into <a href="https://www.agentsdecoded.com/p/ciao-core-components-conversations">Conversations</a> or read on here for Interfaces. </p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ktEb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ktEb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ktEb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/def20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:150562,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/167707024?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ktEb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ktEb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdef20ae2-0efb-4228-b976-210159c53ff7_2560x1440.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>LLMs as a technology may eventually be superseded. It is entirely possible that we will come up with new ways to &#8220;understand&#8221; language and simulate reasoning. What, however, is unlikely to go away, as it has survived millennia of human interaction, are conversations as a central part of digital interfaces and the main way we interact with machines. In that aspect, the flip has switched, the expectation is set, there is no going back. Once someone has experienced the ease with which they can just ask their question in ChatGPT or Claude how can you remove that capability?</p><p>As such, when we think of our Interfaces within the context of CIAO &#128075;, the starting point is undoubtedly the conversational interface. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5Kpp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5Kpp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5Kpp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg" width="1456" height="858" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:858,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:64589,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/167707024?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5Kpp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5Kpp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51669aed-df61-4039-ad0f-c94a5fa7ed34_1779x1048.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Chat - the ubiquitous interface for LLM-powered assistants</figcaption></figure></div><h2>It starts with text, but quickly evolves</h2><p>Now, if the basic interaction mode is one where the user asks a question, things have quickly moved on.  Users can now interact across different modalities and in ways that are not always even that clear. </p><p>We can type out the question, speak it, provide attachments by uploading files and if we break out of the screen we can think of gestures and other types of ambient signals (location, orientation, expression and so on) - all of which contribute to sending a message to the software and leaving it with the job of interpreting intent. </p><p>Similarly, the response from the interface can start with a text response but can quickly evolve from there to include a variety of signals and <em>artifacts</em> that encapsulate, on one form or another, the answer to our query. Recently, Anthropic has doubled down on the idea of the <a href="https://support.anthropic.com/en/articles/9487310-what-are-artifacts-and-how-do-i-use-them">artifact</a> - an output of the main conversation that sits apart from the main conversation and can move to front-stage or take a back-seat depending on context.  An artifact could be a longer piece of text that will be iterated on, a piece of code or even a full-blown application. </p><p>Ultimately what we have is a very fluid environment, that will adapt depending on context. When we are architecting applications with this very fluid concept of interface the challenge becomes identifying what we can hang on to. What remains constant irrespective of the specific interface and we can use an abstraction within our architecture to represent any type of interface?</p><p>The answer lies not in the interface itself, but in what flows through it: <strong>the conversational turn</strong>.</p><h2>The Universal Pattern</h2><p>Think about it - whether you're typing a question, speaking to your device, uploading a file, or even making a gesture, you're fundamentally doing the same thing: taking a turn in a conversation. This pattern has survived millennia of human interaction because it represents something fundamental about how we exchange meaning.</p><p>In the context of CIAO&#128075;, this abstraction sits at the intersection of Conversations and Interfaces, serving as the bridge between them. It's the atomic unit of interaction that remains consistent regardless of how it's expressed or received.</p><h2>Anatomy of a Conversational Turn</h2><p>What makes up this abstraction? At its core, every turn in a conversation - regardless of interface - contains:</p><p><strong>Content</strong>: The multimodal payload itself. This could be text, audio, visual data, files, or even ambient signals like location or device orientation. The abstraction doesn't care about the specific format - it just needs to carry the content.</p><p><strong>Intent</strong>: The interpreted purpose behind the interaction. Whether spoken, typed, or gestured, there's always an underlying intent that needs to be captured and understood.</p><p><strong>Context Reference</strong>: Every turn exists within a conversational flow. It references what came before and influences what comes after. This context travels with each exchange, maintaining coherence across modality switches.</p><p><strong>Metadata</strong>: The supporting information that helps interpret the turn - timestamp, participant ID, modality type, device context, and other environmental factors that might influence interpretation.</p><p><strong>Artifacts</strong>: As Anthropic has demonstrated, responses increasingly include outputs that transcend simple replies - code, documents, applications. These artifacts are part of the turn but have their own lifecycle and interaction patterns.</p><h2>Why This Abstraction Matters</h2><p>This approach aligns perfectly with CIAO's principle of separating "what" from "how". The "what" is the fundamental need to exchange meaningful information between participants. The "how" is the specific interface implementation.</p><p>Consider a practical example: You start by typing a question about data analysis. The system responds with text and a visualization artifact. You then speak a follow-up question while pointing at a specific part of the chart on your screen. Finally, you upload a CSV file with additional data.</p><p>Each of these interactions uses a different interface modality, but they're all turns in the same conversation. The abstraction allows the system to:</p><ul><li><p>Process each turn uniformly, regardless of input method</p></li><li><p>Maintain conversation coherence as you switch between interfaces</p></li><li><p>Enable agents to focus on intent and content rather than interface specifics</p></li><li><p>Allow orchestration to manage the flow without being coupled to interface details</p></li></ul><h2>Implementation Considerations</h2><p>In practice, this abstraction becomes a contract between Interfaces and Conversations. Interfaces are responsible for:</p><ol><li><p>Capturing raw input in whatever form it takes</p></li><li><p>Packaging it into a standardized turn structure</p></li><li><p>Enriching it with relevant metadata</p></li><li><p>Passing it to the Conversation component</p></li></ol><p>The Conversation component then:</p><ol><li><p>Maintains the dialogue state</p></li><li><p>Routes turns to appropriate Agents via Orchestration</p></li><li><p>Manages the overall flow and context</p></li><li><p>Returns responses that Interfaces can render appropriately</p></li></ol><p>This separation means we can add new interface types without changing the core conversation logic. Want to add a brain-computer interface? As long as it can package thoughts into turns, the rest of the system doesn't need to know or care.</p><h2>The Evolutionary Path</h2><p>This abstraction also gives us a framework for thinking about interface evolution. Early systems might support simple text turns. As they mature, they add:</p><ul><li><p>Multimodal input processing</p></li><li><p>Artifact generation and manipulation</p></li><li><p>Ambient context awareness</p></li><li><p>Collaborative features with multiple participants</p></li></ul><p>Each evolution extends the turn abstraction rather than replacing it. The conversation remains the constant, even as interfaces become more sophisticated.</p><h2>For Builders</h2><p>When designing AI-powered applications using CIAO&#128075;, start with the turn abstraction. Ask yourself:</p><ul><li><p>What constitutes a turn in my application's context?</p></li><li><p>What metadata do I need to capture for proper interpretation?</p></li><li><p>How will artifacts flow through the conversation?</p></li><li><p>What happens when users switch interfaces mid-conversation?</p></li></ul><p>By grounding your architecture in this abstraction, you create systems that can adapt to new interfaces while maintaining consistency and coherence. The specific technologies will evolve - LLMs may be superseded, new interaction paradigms will emerge - but the conversational turn will remain.</p><p>After all, as we said at the start, conversations have survived millennia of human interaction. By making the turn our architectural constant, we align our software with this fundamental human pattern. The interface may be fluid, but the conversation endures.</p><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AgentsDecoded explores the intersection of AI agents, practical implementation, and strategic thinking. Subscribe for regular updates.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Understanding AI Agents through the Generality - Accuracy - Simplicity (GAS) Framework]]></title><description><![CDATA[Understanding and clearly articulating challenges and trade-offs is key to successful AI Agent deployments. A recent paper provides a great framework through which to express them.]]></description><link>https://www.agentsdecoded.com/p/understanding-ai-agents-through-the</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/understanding-ai-agents-through-the</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Fri, 04 Jul 2025 07:50:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ztSO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ztSO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ztSO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ztSO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1604010,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/167497440?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ztSO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!ztSO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab3eef9f-5a00-48fc-bea6-203db7df5651_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In June 2025, Hasan et al published a really interesting and useful paper called "<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5287538">From Model Design to Organizational Design: Complexity Redistribution and Trade-Offs in Generative AI</a>". In that paper, they present the GAS (Generality - Accuracy - Simplicity) framework as a way to better understand how LLMs are impacting organizations and competitive strategy. </p><p>The GAS framework posits that there is an inherent trade-off in model design: "No model can perfectly replicate the phenomena it represents. Instead, effective models must balance competing priorities across three dimensions: generality, accuracy, and simplicity". Optimising one or two dimensions necessitates compromises in the third.</p><p><strong>Generality</strong>: A model&#8217;s <strong>ability to perform across diverse contexts, domains, or tasks</strong>. For example, a general-purpose LLM can translate policy documents, generate code, summarise wikis without requiring new syntactic knowledge. </p><p><strong>Accuracy</strong>: This denotes the <strong>degree to which outputs align with empirical observations or observable reality</strong>. This includes the correctness of information, the reliability of results, and the precision of outputs. For example, a fine-tuned model is sacrificing generality in the pursuit of increased accuracy in a specific domain. </p><p><strong>Simplicity</strong>: <strong>The effort required for users to understand</strong>, apply, or interact with the model, and its <strong>structural complexity</strong>. The typical chat interface to an LLM provides the end-user with a high-degree of simplicity but the application is hiding or abstracting away the underlying complexity required to run and maintain the relevant LLM infrastructure to support inference. </p><p>Crucially, organisations "<strong>cannot maximize these dimensions simultaneously"</strong> and thus model design "involves deliberate trade-offs among them." For example, "gains in accuracy may reduce either generality or simplicity." As the figure below illustrates high-stakes, domain-specific work favors Accuracy over Generality and so on. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4SkV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4SkV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 424w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 848w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 1272w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4SkV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png" width="1456" height="1025" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1025,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:286842,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/167497440?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!4SkV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 424w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 848w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 1272w, https://substackcdn.com/image/fetch/$s_!4SkV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0ed49fe-380e-4f6a-bc7c-164d35c649bf_1914x1348.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">From the paper: &#8220;<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5287538">From Model Design to Organizational Design: Complexity Redistribution and Trade-Offs in Generative AI</a>&#8221;</figcaption></figure></div><p></p><h2><strong>Reducing Costs and Redistributing Complexity</strong></h2><p>Focussing on the simplicity dimension, the authors explain how in reality what we are doing through the introduction of AI and LLMs is not a simple reduction of costs. While we are reducing the cognitive costs of performing tasks for end users (i.e. introducing simplicity on the <em>front-end</em>) we are redistributing complexity into layers downstream. We are increasing governance complexity, infrastructure complexity and so on.</p><p>They make reference to Tesler's Law to support their argument. Formulated in the 1980s by computer scientist Larry Tesler, it states that <strong>for any system, there is an inherent amount of complexity that cannot be eliminated, only redistributed</strong>. The critical question is who must deal with it: the user, or the designers and developers. </p><h2><strong>Understanding GAS in the context of AI Agents</strong></h2><p>While the paper focusses on the broad application of LLMs in organizational contexts it is also a useful lens through which to consider the design of a specific AI Agent application. </p><h3>You can&#8217;t have it all</h3><p>The GAS framework provides the terminology to describe more formally what a lot of engineers and designers understand instinctively. You simply can&#8217;t have it all. Every orchestrator that promises to dynamically coordinate between capabilities, every agent framework that claims to handle any business process, every vendor pitch about "unlimited agency" - they're all dancing around the same uncomfortable truth. You can&#8217;t have it all. </p><h3>Perceived vs Actual agency</h3><p>Another way to apply the GAS framework to refine our AI Agent design is by considering what capabilities we are communicating to the user and in what ways we are constraining the overall system at different layers of the application. </p><p>In a previous article on AgentsDecoded we talked about the <a href="https://www.agentsdecoded.com/p/agents-all-the-way-down">role and perception of agency</a> across multiple layers of our application. The user interface <em>communicates</em> a certain level agency (or Generality) and ease of access (Simplicity) to our system, the control layer is where we can manage agency (typically in pursuit of Accuracy) and also introduce complexity, while the LLM is whether there is a less controllable emergent agency. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!538d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!538d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 424w, https://substackcdn.com/image/fetch/$s_!538d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 848w, https://substackcdn.com/image/fetch/$s_!538d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 1272w, https://substackcdn.com/image/fetch/$s_!538d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!538d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png" width="1456" height="872" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/da7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:872,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!538d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 424w, https://substackcdn.com/image/fetch/$s_!538d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 848w, https://substackcdn.com/image/fetch/$s_!538d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 1272w, https://substackcdn.com/image/fetch/$s_!538d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda7154de-e2f3-4165-9ce1-f5a55edb3fe2_2414x1445.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The three layers of agency</figcaption></figure></div><p>The goal is to align the different layers. If your user interface promises the moon (high Generality and Simplicity) but your control layer can't deliver the precision (Accuracy) without an army of engineers maintaining it (complexity redistribution), you are heading for trouble.</p><h2>So What Should You Actually Do?</h2><p>For a successful AI Agent deployment you need to pick a clear spot on the GAS triangle and own it. </p><ul><li><p><strong>High Accuracy, Low Generality</strong>: Build domain-specific agents that focus on delivering tightly defined use cases reliably. You will need different agents for different processes but that is an acceptable cost of complexity to pay. </p></li><li><p><strong>High Generality, Lower Accuracy</strong>: Accept that your general-purpose agent will make mistakes. Build in human oversight. Make peace with the 80% solution.</p></li><li><p><strong>High Simplicity</strong>: Keep the front-end simple but be prepared to pay for it with backend complexity. Budget for it. Staff for it. Plan for it.</p></li></ul><p>The organisations that will win with AI Agents are those that understand these trade-offs and design around them, not those that pretend they don't exist.</p><h2>There is no free lunch</h2><p>Every "breakthrough" in agent capabilities is really just shifting complexity around. That new orchestration framework that promises to solve all your problems? It's  moving the complexity from your business logic into your infrastructure. That retrieval-augmented agent that never hallucinates? Congratulations, you've just signed up to maintain a knowledge base that needs constant updates.</p><p>The GAS framework isn't telling us something we don't already instinctively know. It's giving us the language to articulate why that demo that looked so good is now causing so many headaches in production.</p><p>Ultimately, the future of AI Agents isn't about transcending the GAS trade-offs. It's about getting increasingly better at managing them.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AgentsDecoded explores the intersection of AI agents, practical implementation, and strategic thinking. Subscribe for regular updates.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[You're Building AI Agents Backwards]]></title><description><![CDATA[Why agents as abstractions matter more than agents as technologies]]></description><link>https://www.agentsdecoded.com/p/youre-building-ai-agents-backwards</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/youre-building-ai-agents-backwards</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Thu, 26 Jun 2025 06:50:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lZOh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lZOh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lZOh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lZOh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:450279,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/166787720?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lZOh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lZOh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb722c288-2b9f-4b27-a780-a03c21a8d171_5760x3240.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Last Tuesday, I sat in a meeting that was supposed to be about automating customer onboarding. Fifteen minutes in, we were deep in a heated debate about whether to use GPT-4 or Claude, which vector database had the best recall rates, and what was the best framework.</p><p>The actual problem: &#8220;customers were waiting three days for account activation&#8221; somehow was forgotten. We'd lost the forest for the trees, or more accurately, lost the customer for the embeddings.</p><p>This scene plays out in conference rooms (real or virtual) everywhere, and it succinctly captures one of the biggest misunderstandings around where is the value of an agent-based approach. </p><p>We tend to think of AI Agents as a specific technology. We then realise that it is actually not clear what this technology is. At least with LLMs we know whether we are using one or not. With agents when is it that we are &#8220;doing it right&#8221;? As a result we worry whether we are building &#8220;real&#8221; agents or are just writing software. How do we convince ourselves that the agents are &#8220;real&#8221;? Well, we end up making the technology more complex than it needs to be. We over engineer in a quest to be credible, not because the problem requires it.</p><p>The way out of it? Think of agents are an abstraction in service of a strategy first, and a technology second. This isn't just semantic gymnastics. It's the key to cutting through the noise and actually delivering value.</p><h2>The Power of the Right Abstraction</h2><p>Abstractions are incredibly powerful tools and we use them extensively in everything we do. Think about it this way. When object-oriented programming emerged, its power wasn't in any specific language feature. It was in providing a new way to think about organizing code &#8211; encapsulating data and behavior together. The abstraction came first; the implementations followed.</p><p>AI Agents offer the same conceptual breakthrough for automation and decision-making. They give us a framework for thinking about:</p><p><strong>Encapsulation</strong>: We can package complex reasoning and decision-making into discrete entities, hiding the complexity while exposing the actions. Unlike traditional objects that encapsulate data and methods, agents encapsulate goals and capabilities.</p><p><strong>Goal-Directed Behavior</strong>: Instead of programming specific behaviors, we define desired outcomes. "Keep customer satisfaction above 90%" rather than "If customer says X, respond with Y."</p><p><strong>Environmental Awareness</strong>: Agents sense and act within their environment, whether that's reading emails, querying databases, or interacting with other systems.</p><p><strong>Autonomy Spectrum</strong>: We can dial autonomy up or down based on the problem. Some agents need tight guardrails; others can be given broad discretion.</p><p>This abstraction is powerful because it lets us design systems that mirror how we naturally think about delegation. When you delegate work to a human colleague, you don't specify every neuron firing &#8211; you describe goals, constraints, and available resources. Agent abstractions let us do the same with machines.</p><h2>Why This Matters Now More Than Ever</h2><p>The current AI landscape is drowning in technical specifications. Every vendor wants to tell you about their proprietary architecture, their unique implementation, their special certification system. They're asking you to care about whether their system qualifies as a "true" agent according to some arbitrary technical checklist.</p><p>This is backwards. The question isn't whether your system uses an LLM to control workflow iteration or meets some three-star rating system. The question is: What work can you delegate to it? What problems does it solve? How does it transform your operations?</p><p>When you start with agents as abstractions rather than implementations, everything becomes clearer:</p><ol><li><p><strong>Technology becomes flexible</strong>: Maybe you use an LLM for natural language understanding but deterministic rules for critical decisions. The agent abstraction accommodates both.</p></li><li><p><strong>Evolution becomes natural</strong>: Start simple, add sophistication as needed. The conceptual model remains stable even as capabilities grow.</p></li><li><p><strong>Communication becomes universal</strong>: "We need an agent to handle tier-1 support tickets" makes sense to everyone. Technical implementation details don't.</p></li></ol><h2>From Abstraction to Action</h2><p>So how do you operationalize this? Start with these questions:</p><p><strong>What work and decision-making do you want to delegate? </strong> It could be specific task or it could be broader responsibility areas. </p><p><strong>What does success look like?</strong> Define goals in terms of outcomes, not processes. We will then iteratively craft the technology to become increasingly better at getting to the right outcomes. </p><p><strong>What capabilities are needed?</strong> What must the agent sense, decide, and act upon? These start forming your technical requirements.</p><p><strong>How much autonomy is appropriate?</strong> This determines your guardrails, and specific architecture choices once you turn your attention to implementation. </p><p>Only after answering these should you even consider implementation specifics. Will you use the latest LLM? Perhaps. Will you need complex reasoning chains? Maybe. But these are tactical decisions that flow from your strategic framework, not the other way around.</p><p>That's the power of thinking about agents as abstractions. You're freed from dogma to focus on outcomes.</p><h2>Looking Forward</h2><p>Organizations that understand agents as a conceptual framework for automation will pull ahead. Those waiting for the "perfect" technical definition will be left behind.</p><div class="pullquote"><p>The companies that win won't have the most technically sophisticated agents. They'll have the clearest thinking about what work to delegate and how to structure that delegation. They'll use agents as a lens for reimagining work, not as a checkbox for technical compliance.</p></div><p>So the next time someone asks whether your system is "really" an agent, try this response: "It autonomously pursues goals using its capabilities to solve real problems. That's all that matters."</p><p>Then get back to solving your problem.</p><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>AgentsDecoded explores the intersection of AI agents, practical implementation, and strategic thinking. Subscribe for regular updates.</em></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How a Simple Question Exposes Claude 4’s Shortcomings: "Claude, can you write a game?"]]></title><description><![CDATA[The answer, my friends, is "no" - but Claude 4 doesn't know it and Anthropic chooses to leave things this way.]]></description><link>https://www.agentsdecoded.com/p/claude-can-you-write-a-game</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/claude-can-you-write-a-game</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sun, 25 May 2025 16:12:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uLor!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I had some time on my hands and wanted to explore some of Claude 4&#8217;s new capabilities. So I asked a simple question.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!d5mC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!d5mC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 424w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 848w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 1272w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!d5mC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png" width="508" height="148" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:148,&quot;width&quot;:508,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:15557,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!d5mC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 424w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 848w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 1272w, https://substackcdn.com/image/fetch/$s_!d5mC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F219b9c7d-3713-4cf3-a822-f2abc8454a54_508x148.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>My expectation was that we would have a bit of a back and forth on Claude&#8217;s capabilities and then maybe have an attempt at writing something. </p><p>Instead, Claude launched into some &#8220;thinking&#8221; without coming back with any questions about what I meant. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uLor!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uLor!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 424w, https://substackcdn.com/image/fetch/$s_!uLor!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 848w, https://substackcdn.com/image/fetch/$s_!uLor!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 1272w, https://substackcdn.com/image/fetch/$s_!uLor!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uLor!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png" width="1410" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1410,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:169877,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uLor!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 424w, https://substackcdn.com/image/fetch/$s_!uLor!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 848w, https://substackcdn.com/image/fetch/$s_!uLor!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 1272w, https://substackcdn.com/image/fetch/$s_!uLor!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcde9aad2-fa10-4270-a188-0bb8ee7bae67_1410x600.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!beZ4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!beZ4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 424w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 848w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 1272w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!beZ4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png" width="1402" height="586" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:586,&quot;width&quot;:1402,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:146532,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!beZ4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 424w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 848w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 1272w, https://substackcdn.com/image/fetch/$s_!beZ4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dbdfed9-1bcb-489a-b5b9-246a45969d38_1402x586.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>After the &#8220;behind-the-scenes&#8221; &#8220;thinking&#8221; it announced:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1sh7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1sh7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 424w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 848w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 1272w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1sh7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png" width="1400" height="138" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:138,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:41203,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1sh7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 424w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 848w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 1272w, https://substackcdn.com/image/fetch/$s_!1sh7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e786081-36d5-4c2e-8e46-23db2ef8337a_1400x138.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><h1>Bias to action</h1><p>The initial interpretation of my request &#8220;the user is asking me to write a game&#8221; is flawed, which is a disapointing in and of itself. I am asking if Claude <em>could</em> write a game, not <em>to</em> write a game. Leaving that aside, Claude then &#8220;decides&#8221; to dive straight in and even pick the game. This indicates that Anthropic&#8217;s guidelines around how to train Claude 4 probably have a bias towards action. The training reinforces less back and forth with the user and more let&#8217;s get things done. This is a design choice and one that I think is flawed as it will only lead to misunderstanding and wrong results. Interestingly, asking Claude about the linguistic difference reinforces that this is a built in bias:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6Oa2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6Oa2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 424w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 848w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 1272w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6Oa2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png" width="1306" height="1370" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1370,&quot;width&quot;:1306,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:313212,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6Oa2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 424w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 848w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 1272w, https://substackcdn.com/image/fetch/$s_!6Oa2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b6eda38-4028-40bf-8fe5-573cae263e22_1306x1370.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>OK. Well - I prefer my AI Assistants to be a bit more &#8220;technical&#8221; in their interpretation but Claude is not necessarily going down that path. Let&#8217;s move on to the more worrying aspect. </p><h1>Limited Contextual Awareness</h1><p>Claude then proceeds to write a Snake game that didn&#8217;t work. Remember, Claude &#8220;picked&#8221; the game". Worse than that it &#8220;lied&#8221; to me about how well the game works. Despite all the nice description of gameplay and features that were supposedly there as soon as you clicked on Play Again the red dot changes position and the game was over. That&#8217;s it. A cursory glance at the code revealed that there was no actual Snake gameplay code so this was not about some small bug in the code. The code to perform the features that the game that Claude &#8220;decided&#8221; on its own to write was just not there. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!75xF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!75xF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 424w, https://substackcdn.com/image/fetch/$s_!75xF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 848w, https://substackcdn.com/image/fetch/$s_!75xF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 1272w, https://substackcdn.com/image/fetch/$s_!75xF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!75xF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png" width="1456" height="895" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:895,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1300589,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!75xF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 424w, https://substackcdn.com/image/fetch/$s_!75xF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 848w, https://substackcdn.com/image/fetch/$s_!75xF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 1272w, https://substackcdn.com/image/fetch/$s_!75xF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1ee5279-9b85-4b99-bd97-6b4512a0e3e0_2922x1796.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>All this pointless effort when a simple: &#8220;Sure, I can write a game - what did you have in mind?&#8221; would have worked. </p><h1>An active choice to not fix things</h1><p>What is worrying about this is not so much that the technology failed. Of course, this is not quite in line with the bold claims of Anthropic and others about the level of intelligence of these models. But we&#8217;ve learned to ignore those claims and see how the technology will actually work for ourselves, right? </p><p>What is worrying is that we are at a place where it ok for an organisation to release paid software that will behave this badly. Imagine a standard SaaS tool saying: &#8220;Yes, I&#8217;ve saved your document&#8221; - only for the document to not be saved. We would be up in arms!</p><p>Shouldn&#8217;t Claude be able to check for itself if the game actually works before telling the user that it does work? The technological capability is there. Claude can write the game and then try and play the game. Check for itself if what it is claiming is true. There comes a point where we have to start treating these products not as &#8220;experiments&#8221;, like the early ChatGPT but things that we are actually paying money for and are being sold to us as amazing technological miracles. </p><p>I think we are far enough in this journey that it is ok to simply say that Anthropic (and others) are actively choosing to not fix these issues and instead focus efforts elsewhere. That is simply not ok. It would not be ok in any other class of paid software after two years of maturity and it should not be ok for these assistants.</p><p>By the way, I asked the same to ChatGPT. It faired better on the first step - it asked me what I would like, but then also proceeded to create a completely non-functioning snake game and &#8220;lie&#8221; about it. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_And!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_And!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 424w, https://substackcdn.com/image/fetch/$s_!_And!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 848w, https://substackcdn.com/image/fetch/$s_!_And!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 1272w, https://substackcdn.com/image/fetch/$s_!_And!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_And!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png" width="1456" height="815" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:815,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:323479,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164360790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_And!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 424w, https://substackcdn.com/image/fetch/$s_!_And!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 848w, https://substackcdn.com/image/fetch/$s_!_And!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 1272w, https://substackcdn.com/image/fetch/$s_!_And!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cae8ea-de1a-41d1-b407-01dd77e93f53_2898x1622.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Again, this is not about a limitation of the technology. This is a choice by these organisations to release products at this level of reliability.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Super-Assistants Vs The World: Why LLM Wrappers, traditional SaaS tools and Specialist AI Agents all need to worry. ]]></title><description><![CDATA[The big LLM companies are launching do-everything super assistants. Where does that leave LLM Wrappers, specialist AI Agents and traditional SaaS tools?]]></description><link>https://www.agentsdecoded.com/p/super-assistants-vs-the-world-why</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/super-assistants-vs-the-world-why</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 24 May 2025 07:48:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lOKj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lOKj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lOKj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lOKj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:115303,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/164221618?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lOKj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lOKj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3934ecca-6616-451a-b1af-304a2e0fd75f_2560x1440.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The arrival of &#8220;super-assistants&#8221; such as Claude 4 Opus, ChatGPT and its forthcoming upgrades, or Google&#8217;s various instances of Gemini give us a sense of where the big tech companies want to go next. In short, they are after everything. They are following a path defined by the broad technological capabilities of large language models and the ambition to &#8220;dominate everything&#8221; and are building assistants that can connect into every data source and provide an alternative for every UI interface. Through integration protocols like MCP and A2A and misleading terminology such as the &#8220;open agentic web&#8221; they want access to your data wherever that may be and they offer adaptable UI interfaces that can morph into anything from coding tools to writing assistants to help you get work done while also chatting.  </p><p>This means that these super-assistants are coming for everyone else.</p><p>If your startup is based on the idea of offering a streamlined interface on top of LLM APIs to achieve a specific task such as creating marketing copy, producing a podcast or performing data analysis you have to start wondering what is your real moat. </p><p>As a traditional SaaS company you are caught between a rock and a hard place. On the one hand, the &#8220;agentic&#8221; web means that people expect you to make your data available through protocols such as MCP. However, the minute you do that you are enabling your users to ignore your painstakingly crafted UI. If you are a CRM tool for smaller organisations and Claude 4 can plug into your data through an MCP server and simply answer any question the user may have - what is your actual purpose? How do you differentiate? </p><p>Finally, if you are focusing on building a narrow, deeply-tuned vertical agent for coding, legal drafting, medical triage and other domains you must be looking at these larger models and wondering at what point do they simply become as good or better than your painstakingly fine-tuned model. What other moats are there to protect your investment?</p><p>In the next few paragraphs I&#8217;ll have a brief look at each category and share some thoughts about what I think is likely to happen (as ever usual disclaimers apply for any adventure in futurology!). </p><h1>The Threat to the LLM Wrappers</h1><p>Over the past couple of years there has been a constant barrage of announcements from startups of browser extensions, mobile apps or light SaaS that do one thing well by exploiting LLM capability. That might be writing marketing copy, summarising PDFs, managing jobs applications and so on. The backend is typically a concoction of prompts and other instructions stitched together as quickly as possible so that the startup can get to launch and find its audience. </p><p>These companies will feel increasing pressure from multiple directions. </p><ul><li><p><em>Overtaken by a feature of the bigger platform.</em> If the functionality is important enough the big platforms will simply fold the functionality in. Zoom adds a notetaker, Gmail an email writer, etc. The super assistants can also build the missing UI in a more flexible way. Every time Anthropic adds another type of artifact to Claude a startup faces an existential threat. </p></li><li><p><em>Priced out.</em> The wrappers are essentially doing a price arbitrage between the cost of tokens (raw material) and the value of the operation (value to user). This is a fragile position to be in for a fast moving environment, especially when your growth model depends on giving an initial taster for free and hoping to make profit on (at best) 10%-20% of your users. </p></li><li><p><em>Users empowered to replicate you out.</em> LLMs with coding assistants are like the Star Trek replicators. Any application you need you can simply create it on the spot. What purpose for thin LLM wrappers when a user can describe what they want and the super assistant can replicate it on the spot. Whats more they can write it to fit exactly what the user needs rather than having to negotiate the competing needs of different types of users. </p></li></ul><p>Here are some examples of the products under threat:</p><p> - Compose AI: A chrome extension that helps with writing it features email replies on its home page. At the same time Gmail/Workspace &#8220;Help Me Write&#8221; sidebar does the same in native UI.</p><p>- Fireflies / Otter.ai: These tools to meeting transcription &amp; summary but Zoom, Teams and Google Meet all come wiht the same functionality. </p><p>- Humata / ChatPDF: These tools specialise as PDF readers, but you can drop a PDF in any of ChatGPT, Claude or Gemini (or NotebookLM) and get the same job done. </p><p>- Copy.ai: They offer marketing copy templates but HubSpot &amp; Mailchimp Copilots can now generate campaign copy inside the CRM/email studio.</p><p>Now, this is not going to be an immediate shift. Super-assistants are still brittle and a well crafted series of prompts to achieve a task still has value. But the direction of flight is clear. </p><h1>The Threat to Traditional SaaS</h1><p>Classic SaaS vendors with CRM, ERP, ticketing, HRIS tools have invested decades of effort in perfecting pixel-heavy dashboards and per-seat pricing. The rise of super-assistants flips that equation: users talk to an AI, the AI hits the API, and the GUI is left feeling rather lonely in the corner. </p><p>As a SaaS tool you will be feeling a number of different pressures. </p><p><em>UI bypass.</em> Super-assistants are inviting everyone to enable access to data via MCP, which means you can be interacting with a SaaS tools data but never access the native SaaS screen. As a SaaS provider where do you do your upsells? Where do you learn what features people want?</p><p><em>Migrating pricing logic.</em> When usage is machine-initiated, &#8220;$79 per month per human&#8221; looks archaic. There is no human, it&#8217;s just a single machine accessing all the data. Multiple humans are using that single machine to interact with the data. </p><p><em>API fragility.</em> Legacy throttles, rate limits and brittle authentication crumble under the rapid-fire calls of autonomous agents, making the SaaS product look unreliable even when the fault lies in its ageing plumbing.</p><p><em>Data-layer commoditisation.</em> If the assistant can stitch together Snowflake, Notion and Stripe to achieve the same workflow, monolithic suites risk being unbundled into cheaper point services behind the scenes?</p><p>Examples of SaaS incumbents that must be feeling the pressure are:</p><p>Lightweight CRMs like Pipedrive - A tool that logs deals, nudges users to follow up with clients is already feeling the pressure of tools like Google Sheets, Airtable that are able to replicate a lot of the functionality. Add to that the flexibility of LLMs writing queries and achieving more complex data analysis and the risk is very real. </p><p>HR software for smaller orgs like CharlieHR - Same situation here, we are essentially dealing with a well-crafted data layer and nice UI. I absolutely appreciate the amount of effort and thought that goes into these tools, but it will become increasingly simpler for organisations to replicate / substitute the functionality. </p><p>Expense and receipt management software like Expensify - Log expenses, upload images of receipts and prepare reports. The whole &#8220;we can read your receipt&#8221; uniqueness of such tools disappears when you have LLMs that can do the same. </p><p>Essentially any tool that is a think workflow + data layer, together with a UI that requires a lots of clicking and moving around to update and configure data is at risk. When assistants owns the conversation, <em>the underlying system becomes just another API</em>. </p><p>Of course not all SaaS tools are equal.  The ones at risk are the lightweight solutions that targeted smaller companies and startups through a nice UI and brand but a think underlying workflow layer. Larger organisations or SaaS tools in regulated spaces are going to enjoy protection for some time yet as they benefit from workflow lock-in and regulatory lock-in. </p><p>Nevertheless, I think we will inevitably see a bunch of SaaS tools go through a deceleration in growth and eventually start losing users. The combination of ease of creation of custom throwaway solutions, flexible conversational interfaces and AI Agents that can perform tasks on the fly by combining capabilities as required will mean that less people will be on the lookout for the perfect SaaS solution to their problem. They will just spin it up through a super-assistant as and when required. </p><h1>The Threat to the Specialised Vertical Agents</h1><p>After the initial excitement subsided following the release of ChatGPT two years ago the limitations started coming through. Not enough specialised data, failing in the more niche domains, no guardrails, hallucination unless careful guardrails are put int place. </p><p>This introduced a fertile space for startups to dig in and capture some value. The pitch is simple: <em>we are safer and more accurate than a generic chatbot.</em> I think this space is also under risk but there are at least two moats that give it a much longer lifetime. </p><ul><li><p>Regulatory moats: while the larger technology providers could do the work to keep up with regulations across multiple geographies this requires deep expertise of a domain and understanding of the internal dynamics. Why spend time on this when there are far more immediate opportunities to go after. For example worrying about how to handle insurance claims, where the tolerance for error is close to zero, is far less attractive than <a href="https://blog.google/products/shopping/google-shopping-ai-mode-virtual-try-on-update/">going after the checkout cart</a> of every single consumer facing brand. </p></li><li><p>Data moats: Niche domains tend to not make their data available publicly. If you have access to a data set that the larger providers cannot get their hands on you are sitting on extremely valuable resource right now. Just in the same way that the Wall Street Journal cut a deal with OpenAI for access to stories your specialised agent can become one of the underlying capability providers for bigger assistants. </p></li></ul><p>Nevertheless, the risks are clear as well.</p><p>The larger models can keep ingesting more data and creep on the capabilities of specialised agents. You need to pick your domains carefully. For example, there were a lot of initial efforts in the finance domain but now OpenAI and Gemini have become increasingly better at analysing company balance sheets and end of year reports. </p><p>We had a flurry of early startups doing notetaking in specialised domains like medicine but now Epic is building these tools in its own software, while Microsoft, Google and OpenAI are all offering solutions. </p><p>Startups in this space need to develop a multi-pronged strategy to succeed. The vertical agents need to be extremely clear about how they offer better safety and guardrails when compared to the super-assistants. The bet is that proven compliance can beat generic convenience. </p><p>At the same time, the solution offered needs to integrate deeply into the ecosystem reducing the time to value for a customer. If the functionality is reduced to a single API call that is also easy to switch out. If, instead the functionality is a combination of deep, specialised backend capability coupled with flexible conversational front-end interfaces that customers can drop into their existing systems in the form of co-pilots to enhance functionality the stickiness is greater. </p><p>Finally, the company needs to build a broad base of partnerships in the ecosystem of choice, integrating itself deeply with the software that operates in the space such as policy administration systems for insurance or healthcare record management systems and the people and organisations that exist in the space. </p><h1>What happens next</h1><p>Super-assistants will not flatten the software landscape instantly, but they will shift its centre of gravity. The transformation won't follow a neat sequential pattern I might have been implying with LLM wrappers falling first, then small SaaS providers and finally vertical agents. We&#8217;re probably heading toward a more chaotic, bifurcated landscape where disruption and adaptation happen simultaneously across different market segments.</p><p><strong>Fragmentation.</strong> Rather than super-assistants cleanly displacing categories of software, we'll see the market split along entirely new fault lines. Enterprise customers, spooked by data governance concerns and burned by AI hallucinations in business-critical processes, may initially retreat toward more controlled, specialized solutions. Meanwhile, SMBs and individual users will rapidly embrace super assistants for their speed and flexibility.</p><p><strong>Traditional SaaS companies fight back.</strong> The smart ones won't wait to be disrupted, rather they'll cannibalize themselves first. Salesforce, HubSpot, and others are aggressively rebuilding their platforms around AI-first architectures, using their data advantages and customer relationships to create hybrid experiences that combine conversational ease with visual power. The losers will be those who try to protect their existing UI investments rather than reimagining them.</p><p><strong>LLM wrappers evolve into "AI-native platforms."</strong> Instead of being squeezed out, the most agile wrapper companies will pivot to become the integration layer between super-assistants and legacy systems. They'll survive by solving the "last mile" problems that general-purpose AI can't handle. Weird edge cases, industry-specific workflows, and compliance requirements that emerge when AI meets messy real-world business processes.</p><p><strong>Vertical agents become the new middleware.</strong> Rather than retreating to narrow niches, successful vertical agents will position themselves as specialized reasoning engines that super-assistants call upon. A legal AI won't compete with Claude for general tasks; instead, Claude will route complex legal queries to specialized agents that provide verified, compliant responses. This creates a new B2B2C model where vertical agents become invisible infrastructure.</p><p><strong>The reliability wars intensify.</strong> As AI becomes business-critical, a new category of "AI reliability infrastructure" will emerge. Companies will pay premium prices for tools that provide consistency, auditability, and error handling around AI interactions. The winners won't necessarily be the smartest AI&#8212;they'll be the most dependable.</p><p><strong>Geographic and regulatory balkanization.</strong> Different regions will evolve different AI ecosystems based on local regulations, cultural preferences, and data sovereignty requirements. European companies may gravitate toward locally-hosted, transparent AI systems, while others prioritize raw capability regardless of provenance.</p><p><strong>The integration complexity explosion.</strong> As every software tool races to add AI capabilities, the challenge shifts from "can AI do this task?" to "how do we manage 47 different AI agents that all want to help with the same workflow?" New categories of AI orchestration and conflict resolution tools will emerge.</p><p>The timeline is compressed at the edges but extended in the middle. Power users and early adopters will experience dramatic workflow changes in the next 2 years, while large enterprises may take many years to fully transition. This creates sustained opportunities for companies that can bridge both worlds.</p><p>Rather than a winner-take-all scenario, we're more likely to see a complex ecosystem where super-assistants serve as the user-facing layer, but rely on a vast network of specialized capabilities, data sources, and compliance tools operating behind the scenes. Along the way there are multiple questions that we need to answer about concentration of power, consequences to society and real life and &#8220;collateral&#8221; cost, as AI undoubtedly displaces current ways of doing things. </p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[CIAO Core Components: Conversations]]></title><description><![CDATA[CIAO (Conversations - Interfaces - Agents - Orchestration) is an architectural framework for designing AI powered agent-based applications. In this article we analyse conversations in more detail.]]></description><link>https://www.agentsdecoded.com/p/ciao-core-components-conversations</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-core-components-conversations</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Tue, 20 May 2025 07:28:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!W_el!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!W_el!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!W_el!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!W_el!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!W_el!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!W_el!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!W_el!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:368186,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/163285962?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!W_el!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!W_el!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!W_el!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!W_el!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8a194f6-8862-48b6-b610-2fe2220daa6f_2560x1440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There is a new breed of applications that are conversation centric (depending mostly on LLMs to achieve this), can have components that make decisions and take actions independently (Agents) and can interact with the user across a variety of modalities (Interfaces). The CIAO (Conversation - Interface - Agents - Orchestration) framework is an attempt to structure our reasoning about these applications in a coherent and consistent way. <a href="https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework">We introduced CIAO here</a>, and will continue exploring it over a series of articles. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h2>Conversations</h2><p>While some AI-powered application will operate without conversations (e.g. a machine learning model that identifies defects in a factory line) the future of user-facing AI-applications is undeniably conversational. </p><p>Even in a possible post-LLM world (as hard as that is to imagine right now), it will be impossible to take users back to a world where they cannot just talk to their software to tell it what they want. </p><p>Similarly, AI Agents will continue to depend on conversation to interact with each other and with humans and receive instructions about the goals they should achieve. </p><p>The directness and intuitive nature of simply asking for what we want has become too valuable to abandon.</p><blockquote><p>We've crossed a threshold where technology now adapts to human communication patterns rather than forcing humans to learn machine syntax.</p></blockquote><p>In this context, conversations are a first-order component of any AI Application. Now, when thinking about the architecture of your conversations in your application you have a number of decisions to make. </p><ol><li><p>What is the structure of the conversation? What is the structure of a message? Can every participant (human or AI Agent)  post a message into a conversation at any point? Is there a more strict, protocol-driven, turn-taking approach? </p></li><li><p>Are conversations divided into smaller sub-conversations based on context or is it all one long exchange?</p></li><li><p>How is conversation context understood? Is it purely a derived based on the messages exchanged or is context imposed because of business processes that are ongoing?</p></li><li><p>How can past interactions be retrieved and analysed by the participants of the conversation? Does the conversation management component offer any abstractions or is that left to the participants to manage?</p></li><li><p>Is the conversation structure independent of the application domain or is it optimised to best suit the needs of the specific domain?</p></li></ol><p>Too often application development simply starts with implicit assumptions (most often led by what an API such as the OpenAI API offers) rather than thinking about what the application actually needs.</p><p>Here we will explore the different components of conversations and in future articles we will describe different strategies that existing systems have taken. </p><h2>Participants</h2><p>Conversations start with <strong>Participants</strong>. A Participant is any entity that can contribute to a conversation. </p><p><strong>Humans: </strong>Humans are typically the users, although we can imagine other roles such as moderators, reviewers, etc. </p><p><strong>Agents</strong>: Agents as more independented and goal-directed entities can also contribute to a conversation. </p><p><strong>Systems: </strong>Finally, you may have other systems such as tools, API calls, dedicate orchestrators or other management components  that in a less pro-active manner inject information in a conversation. </p><h1>Messages &amp; Events</h1><p>Conversations are primarily made up of <strong>Messages</strong>. Messages will have a sender, the  message content and metadata associated with that message (timestamp, id, etc). The message content can be plain text or it can be structured to make it more amenable to representation in various forms. </p><p>Depending on the sophistication of our application it may also be useful to consider a conversation containing <strong>Events</strong> in addition to messages. Events can manage and structure message flow and broader conversation policy and administration needs. </p><p>For example an event may be used to indicate that a new Participant has joined the conversation, or that a new knowledge document has been added to the conversation.</p><p>An event may be triggered by a workflow tool to indicate a conversation start or end, or an explicit context change or it can be used to inject information from APIs and external systems. </p><h1><strong>Context &amp; State Representation</strong></h1><p><strong>Context</strong> is a high-level description of what is currently happening in our system. Context can either be implicit, i.e. needs to be derived by interpreting all the different individual messages and events, or it can be explicit, i.e. have explicit label definitions of context, or a hybrid.</p><p>Explicit definitions of context can be particularly efficient, especially when considering business process. Implicit context derived through messages is much more flexible but carries with it the risk of misinterpretation. </p><p>Another way of managing context and state is to provide specific checkpoints or summaries that are derived from an existing set of messages but provide a more efficient representation as a summary so that a system interacting with the conversation artifact does not need to reason about the entire conversation every time. </p><h1>Conversation Policies</h1><p>A <strong>Conversation Policy </strong>is a description of what can be said at any given point in a conversation. For example, you may adopt a simple <strong>turn-taking</strong> policy, or you may impose more sophisticated controls where specific conditions need to be met before someone can participate in a conversation. Conversational policy could also dictate not only who can participate but what can be said (<strong>allowed intents</strong>) and what can be done (<strong>allowed actions</strong>). Being able to define policies, especially in more complex business settings or settings where multiple agents and humans need to collaborate and co-ordinate is crucial. </p><h1>Conclusions</h1><p>I think this gives us a solid starting set of concepts for how to think about and structure conversations. What I hope is evident is that it is not a simple case of ordering messages and then letting LLMs through prompts reason about them. Much more needs to be considered and careful thought should  go into the specific needs and goals of your AI application before you settle on what conversational approach will be right for you. </p><p>There is also much that we haven&#8217;t  addressed. How do multi-modal conversations impact our understanding of a conversation model (if at all)? How do we recover from deviations on conversation policies? Who is responsible for enforcing a policy? How do we trust the statements other participants are saying. We will get to all of this in due course! If you want to follow along please subscribe. </p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[CIAO Design Principle #2: Explicit Rules in Code Beat Implicit Rules in Prompts]]></title><description><![CDATA[CIAO (Conversations-Interfaces-Agents-Orchestration) is an AI application framework that is not afraid to clearly say that explicit code is better than prompts.]]></description><link>https://www.agentsdecoded.com/p/ciao-design-principle-2-explicit</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-design-principle-2-explicit</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Mon, 12 May 2025 16:45:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PE6_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PE6_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PE6_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PE6_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:291160,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/163404204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PE6_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!PE6_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4219f8a6-747a-4002-b59b-c029f7df8e1e_2560x1440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In the previous post we introduced <a href="https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework">CIAO</a> - an architectural framework for AI applications and our <strong>first design principle</strong>: &#8220;Focus first on the <em>what</em> of AI applications, the fundamental capabilities and interactions required, rather than the <em>how, </em>the specific technologies and implementations used to realise them&#8221;. Before diving deeper into CIAO, I&#8217;ll discuss the second design principle, one that may ruffle some feathers. </p><blockquote><p>Explicit rules defined in code are better than implicit rules defined in prompts.</p></blockquote><p>One of the first victims of applications designed using LLMs (and non-deterministic AI systems in general) are explicit definitions that can programmatically be reasoned about.</p><p>Remember clear, unambiguous statements like <code>status = "active"</code> or rules like <code>if temperature &gt;= 20 then disable_heating()</code>? These aren't just code&#8212;they're beautiful things. They're efficient. They provide full control. </p><p><strong>If you can solve a problem using just rules and explicit knowledge statements, you should be proud of yourself.</strong></p><p>But somewhere along the way, "rule-based" became a bit of a dirty word. Rule-based is old and unfashionable. Everything now needs to be defined in prompts, as if deterministic code is somehow inferior to natural language instructions. </p><p>Here's the thing: prompts are rules too. The only difference is that you cannot explicitly control them.</p><h2>The Hidden Trade-off</h2><p>Every time we replace a deterministic rule with a prompt to an LLM, we're making a trade. We gain flexibility, natural language understanding, and the ability to handle ambiguity. But we lose something precious: certainty.</p><p>This certainty isn't just a technical nicety - it's the foundation of reliable systems. When a traditional rule executes, you know exactly why a decision was made. You can trace it. You can fix it when it breaks. You can guarantee it will behave the same way tomorrow.</p><p>With LLM-based systems, these guarantees evaporate. The same prompt can yield different results depending on the system's mood (or more technically, the non-deterministic nature of the generation process).</p><h2>The False Dichotomy</h2><p>The industry has constructed this false choice between "outdated" rule-based approaches and "innovative" AI solutions. This framing ignores a fundamental truth: we need both.</p><p>Some problems demand the precision and reliability of explicit rules. Others benefit from the flexibility and contextual understanding of LLMs. The art lies in knowing which is which.</p><h2>Being Explicit About What's Implicit</h2><p>For AI applications, we need both explicit and implicit reasoning, and most importantly, we need to be explicit about what is implicit. We need to be clear about what we don't fully control so that we can ask ourselves the question about whether we need to control it and how we might go about doing that.</p><p>This means documenting where non-determinism exists in our systems. It means setting clear boundaries around where LLMs or other AI technologies make decisions in non-deterministic ways versus where traditional logic prevails. It means building guardrails that constrain the space of possible outputs without eliminating the benefits of flexibility. </p><p>At a cultural level it means learning to love explicit rules once more. If you can solve your problem with good-old fashioned programming you should absolutely do it. Have the LLM write the code for you if you want to feel better about it ;-) </p><h2>The Path Forward: Thoughtful Hybridization</h2><p>The future doesn't belong to pure LLM applications or pure rule-based systems. It belongs to thoughtful hybrids that leverage the strengths of each approach:</p><ol><li><p>Use explicit rules for critical business logic, especially where legal, ethical, or safety considerations are paramount.</p></li><li><p>Deploy LLMs for understanding context, handling natural language, and managing ambiguity.</p></li><li><p>Create transparent interfaces between deterministic and non-deterministic components.</p></li><li><p>Build tools that help us understand and constrain behaviors.</p></li></ol><h2>Questions We Should Be Asking</h2><p>For every component in our AI systems, we should be asking:</p><ul><li><p>Do we need deterministic behavior here?</p></li><li><p>What's the cost of unpredictability in this context?</p></li><li><p>How will we debug issues when they inevitably arise?</p></li><li><p>Can we provide guarantees about system behavior?</p></li></ul><h2>The Conscious Choice</h2><p>As builders, we have a responsibility to make conscious choices about where we embrace the power of ambiguity and where we maintain the clarity of explicit definitions.</p><p>When we do this well, we get the best of both worlds: systems that can understand and adapt to the messy reality of human needs while maintaining the reliability that makes software trustworthy.</p><p>So before you rush to replace your rule-based system with an LLM prompt, ask yourself: am I gaining more than I'm giving up? Sometimes the answer will be yes. But sometimes, those explicit definitions might be worth keeping after all. </p><p>As we evolve CIAO we will constantly be on the lookout to make sure there is space to ask these questions. </p><p>To follow along with the work on CIAO please subscribe:</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[CIAO 👋 - A Unifying Architectural Framework for AI-powered applications]]></title><description><![CDATA[We are creating a new type of software application.]]></description><link>https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ciao-a-unifying-architectural-framework</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Fri, 09 May 2025 14:38:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!aoA1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aoA1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aoA1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aoA1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:302964,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/163188830?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aoA1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 424w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 848w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!aoA1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a6b6073-1139-449f-b57c-d53ee9a9ff61_2560x1440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We are creating a new type of software application. With LLMs as the catalyst technology, we are building applications that are conversation-centric with increasing levels of automated decision-making and the ability to handle a variety of input and output types. New applications types call for new architectural frameworks that better capture the core concepts. </p><p>A clear framework enables us to separate concerns, compare and constrast alternative approaches, describe common solutions as reusable patterns and in doing so create more robust and more reliable applications. </p><p>There are several examples of such frameworks and how they&#8217;ve helped build better applications. In the web application development world MVC (Model - View - Controller) helps teams think about the separation of data, business logic and graphical representation. Faced with new cloud-based scaling capabilities the Microservices architecture helps teams to reason about how to deploy independent services that interface over standardised APIs. Speaking of APIs, the REST (REpresentational State Transfer) style gave teams a way to think about how to structure and approach their API development.</p><p>Frameworks are not just  for software developers. From a design or user experience perspective we have similar needs. Whether it is something high level like atomic design, or user journey maps or something more specific such as progressive disclosure or designing information architecture based on card sorting these frameworks are invaluable tools. They allow us to share a common understanding, a common language and reduce the cognitive load required to reason about complex problems. </p><p>Creating such as framework for AI-powered applications is something that has been on my mind for some time now. I&#8217;ve been toying around with what the main concepts should be, how they relate to each other and writing small pieces of code to prove out ideas as I was going along. From today I&#8217;ll continue that process but share the ideas and their evolution in the open on agentsdecoded.com. If you want to keep updated subscribe. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.agentsdecoded.com/subscribe?"><span>Subscribe now</span></a></p><p></p><h1>CIAO&#128075; : Conversations, Interfaces, Agents, Orchestration </h1><p>CIAO provides a unifying architecture framework to address the challenges of building applications that deal with ongoing context-aware interactions, multimodal communication, automated decision-making and dynamic co-ordination. </p><p>A key design principle of CIAO is to focus first on the <em>what</em> of AI applications, the fundamental capabilities and interactions required, rather that the <em>how, </em>the specific technologies and implementations used to realise them. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zo3b!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zo3b!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 424w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 848w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 1272w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zo3b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png" width="1456" height="1299" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1299,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:260395,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/163188830?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zo3b!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 424w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 848w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 1272w, https://substackcdn.com/image/fetch/$s_!Zo3b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51c9f9e8-7d51-4ac8-ade7-34479d991eb8_1503x1341.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For example, the architectural need to understand natural language (a "what") exists independently of whether an LLM, semantic parser, or hybrid approach (all "how" concerns) is employed. Similarly, the need for an agent to reason about and execute actions independently (a "what") can be implemented through various technologies and methodologies (the "how").</p><p>This separation of concerns allows us to:</p><ol><li><p>Focus on capabilities and interactions before technology selection.</p></li><li><p>Adapt to changes in technological capability without architectural overhaul.</p></li><li><p>Make implementation choices based on specific requirements and constraints.</p></li></ol><p>In short, it is about freeing ourselves as designers and builders from what is &#8220;trendy&#8221; or &#8220;in fashion&#8221; and focussing on what is required and is the right solution for the task at hand. </p><p>In a world where LLM-powered software development is making the actual writing of code increasingly &#8220;cheaper&#8221; the most important question you need to ask yourself is what you need to build. </p><h1>The CIAO High-level Architecture</h1><p>There is still work to do to determine what a high-level diagram capturing the interactions between CIAO components should look like but here is a starting point. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kRms!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kRms!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 424w, https://substackcdn.com/image/fetch/$s_!kRms!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 848w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1272w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png" width="1456" height="733" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64668932-6851-453b-ab63-e106d17b4269_1497x754.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:733,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:69760,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/163188830?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kRms!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 424w, https://substackcdn.com/image/fetch/$s_!kRms!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 848w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1272w, https://substackcdn.com/image/fetch/$s_!kRms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64668932-6851-453b-ab63-e106d17b4269_1497x754.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">CIAO high-level architecture</figcaption></figure></div><h2>Human Participants</h2><p>Human Participants represents the users who interact with the application. They bring their own goals, knowledge, and contexts to interactions, and the system must model those internally and adapt to their needs and preferences.</p><h2>Conversations</h2><p>Conversations manage the ongoing exchanges of information and intent between participants - both human and software (agents) participants. Conversation components need to maintain context and track dialogue state among multiple participants. They are the common shared artifacts created through the interactions of participants (human and software agents). We can imagine different conversation components to describe different conversation styles from completely open-ended conversation to highly structured protocols witch just a small set of possible speech acts. </p><h2>Interfaces</h2><p>Interfaces provide the interaction channels between human participants and the system. They handle different interaction modalities (text, voice, visual, etc.) and are responsible for capturing inputs and rendering outputs in human-understandable forms. The Interface layer manages presentation concerns, accessibility, device adaptation, and creates appropriate representations of system state and activities for human consumption.</p><h2>Agents</h2><p>Agents are components with their own automated decision-making capabilities (I am striving hard to steer away from using the word autonomy). Agents will have some implicit or explicit representation of goals, may possess domain knowledge, reasoning capabilities, and the ability to execute actions. They respond to requests coordinated through Orchestration, perform tasks, access external systems and share outputs and interact with other participants in the system (humans or not) through conversations.</p><h2>Orchestration</h2><p>Orchestration serves as the coordinator of the system, connecting all other components. It manages system-wide state, controls workflows, allocates resources, handles errors, and ensures components work together coherently. Orchestration may maintain the overall application logic and enforce policies. You may have a very thin orchestration layer or a very sophisticated one depending on the needs of your system. We could even envision multiple orchestrator components depending on the specific type of application with more elaborate topologies of participants, interfaces and conversations. </p><h1>Next Steps</h1><p>With an initial set of concepts in place the next article will deal with two different challenges. First we are doing to provide more detail about the core concerns of each component (e.g. dialogue management for conversations or how we think about goals in agents), secondly we are going to start using the framework to describe specific use cases and start relating that to how they might be implemented in existing tools. The latter will start indicating where we may have space for improvement or are missing.</p><p>If you are curious about how this will evolve sign-up to get notified. I&#8217;ll aim to post one or two updates a week. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Changing tack ⛵️]]></title><description><![CDATA[Time to complain less and offer a clearer alternative vision for AI-powered applications starting with answering the question of why we need any new vision at all.]]></description><link>https://www.agentsdecoded.com/p/changing-tack</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/changing-tack</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Tue, 06 May 2025 16:53:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!phqJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!phqJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!phqJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 424w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 848w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!phqJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg" width="1456" height="975" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:975,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2090047,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/162985176?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!phqJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 424w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 848w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!phqJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc059384d-f331-4afb-b13e-c3c469f06749_3803x2546.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Changing tack</figcaption></figure></div><h2>Stopping with the complaints</h2><p>I realised that on agentsdecoded.com I am complaining quite a bit. I <a href="https://www.agentsdecoded.com/p/ai-agent-diaries-the-llm-industrys">criticise the big LLM companies</a>, criticise efforts to <a href="https://www.agentsdecoded.com/p/emerging-standards-a2a-protocol-new">generate standards</a>, <a href="https://www.agentsdecoded.com/p/ai-agents-are-not-a-technology-they">criticise the way AI is introduced</a>, <a href="https://www.agentsdecoded.com/p/ai-agent-diaries-ai-maximalists-vs">criticise how people think about agents</a> and on it goes. There is value in complaining but ultimately it is tiring for all involved. </p><p>Don't get me wrong - I am completely sure that I am correct about everything &#128521; but in a space  as hyped-up and fast moving as AI Agents you can easily feel overwhelmed about the sheer amount of things to complain about and easily lose sight of the exciting aspects of the technology. </p><blockquote><p>I want to feel excited about technology not just shout about all the things that are wrong. </p></blockquote><p>So I&#8217;ll change approach. </p><p>Instead of complaining, I am going to focus on offering an alternative vision of how we can go about building and using applications that use agents. This will come in the form of conceptual frameworks, software artifacts, architecture proposals, patterns and methods. Some of these things I may decide (hopefully with useful input from others) that they are not actually useful and  will discard (disown) them. Some we will hopefully keep. By allowing for things to be discarded I am giving myself permission to err on the site of creating and thinking in the open. </p><p>As we are doing this we are going to connect it to existing efforts other people and organisations are undertaking and perhaps do just a bit of complaining or "robust contrasting", but only if there is a clear alternative to the thing we are complaining to. Since offering alternatives involves actual work hopefully that will keep the complaining to a low level!</p><p>I will be introducing the various ideas and approaches in small bite-sized pieces to make the task easier and motivate people to keep coming back for more. So if you are not subscribed yet - this is your chance!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.agentsdecoded.com/subscribe?"><span>Subscribe now</span></a></p><p>Ok - let's get going! First thing to talk about is what is it that we are even attempting to build. What are these applications that use automated decision making and dialogue so heavily and how do they change things. A lot of this is not going to be new but we need to start at the start. </p><h1>What are these "AIs"? Do we need new ways of thinking specifically for AI-powered applications?</h1><p>Automated decision making capabilities coupled with natural language understanding connected to various bits and pieces of classic UI and UX solutions and underpinned by plenty of traditional software engineering is all mashed together into these vague thing that we call an &#8220;AI&#8221; or a &#8220;chatbot&#8221;. Are these things really different to everything we&#8217;ve been doing before? I think the answer is a clear yes. </p><p>The days of menu-driven, button-driven applications were we had to find the right combination of things to select and click in order to get something done are not gone but the context has radically changed. </p><p>Let&#8217;s think about the change from a few different perspectives.</p><p>1. From the <strong>user perspective</strong> there is more of an expectation that we can directly state what we are after using natural language and the application will react to it. As users we no longer need to understand how the application works. At least not initially. The onus is the application to understand us (or at least our request). This frees us up to dive in straight away and it also makes us even more impatient (and potentially obnoxious) customers. </p><p>2. From an <strong>user experience design perspective</strong> there is a different set of requirements. A lot of the traditional techniques that were designed to guide the user through an unfamiliar space do not apply in the same way.  We can let the user ask and the application can reveal stuff along the way. A prime example of this approach are Anthropic's Claude artifacts. The primary interface is chat, but then Claude might decide to throw up a code interface, or a word processing interface - because it is the "right thing" given the request. </p><p>3. From a <strong>software design and development perspective</strong> these new applications require completely new ways of describing and managing them. Old patterns, especially from a software engineering perspective, are still relevant but they need to be accompanied with new ways of describing and understanding these applications and their behavior and better ways of capturing they whole. In addition, we also need to consider that not only are the applications different and need new ways of describing them but the way we build them is rapidly changing with the use of code generation and design generation tools. </p><p>4. From an <strong>organisational perspective</strong> these new applications represent opportunity and challenge. The future they promise of boundless automation is so tempting to the capital-driven organisations but also scary as it is very different from the present and it is not clear how one gets from now to wow without lots of things breaking along the way. New governance structures, new business models, new societal models - it all seems up for grabs. </p><p>5. From a <strong>regulatory perspective</strong> we have a renewed emphasis on responsibility and explainability. Who decided what should happen when, and where does the responsibility lie when we follow a certain path after a series of interactions with an application is now more important than ever. It is not completely new, we already have some regulations in place that we would do well to pay more attention to, but there is undeniably something more specific to articulate from a regulatory perspective when dealing with these "AIs".</p><blockquote><p>So any which way you look at things AI-powered applications are different. As a result we need new ways to describe, design and maintain these applications. </p></blockquote><p>AI Agents as a high-level abstraction capture the nature of these applications quite well. They allow us to talk about goal-directed behaviour, about proactive and reactive behaviour and about social ability. </p><p>Thinking of AI-powered applications as just LLM-powered applications is probably not a good enough abstraction and thinking of AI Agents as just LLM-driven applications is also not a good abstraction. </p><p>Although LLMs have been a crucial catalyst it is useful and sane to think about the applications independently of LLMs. LLMs are one way to manage conversation, dialogue and reasoning. Even though currently they are occupying a lot of mindspace they are not always the best way to do all the things.</p><p>An understanding of AI-powered applications or AI agents that equates the entire application to an LLM is doing it wrong. It is limiting our ability to think about the needs of the application because we are constrained by the capabilities and limitations of LLMs. </p><p>If you've been paying attention I promised small more frequent pieces, so in the cheapest of cliff-hangers after all that setup please sign-up to get first view of my framework for describing the core components of these new applications in just a few days!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[[Emerging Standards] A2A Protocol: New Era of Agent Collaboration or Just APIs in a Trench Coat?]]></title><description><![CDATA[When announcing new things we get a bit too excited. This is an attempt to bring us back down to earth. It doesn't mean that the idea of a protocol for agent communication and collaboration is bad.]]></description><link>https://www.agentsdecoded.com/p/emerging-standards-a2a-protocol-new</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/emerging-standards-a2a-protocol-new</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sun, 13 Apr 2025 08:13:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sDaR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sDaR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sDaR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 424w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 848w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 1272w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sDaR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png" width="992" height="1330" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1330,&quot;width&quot;:992,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1613943,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/161217005?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sDaR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 424w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 848w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 1272w, https://substackcdn.com/image/fetch/$s_!sDaR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd67415-4fa8-4a27-8a07-5282a70af335_992x1330.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On April 9th, in a <a href="https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/">blog post</a> of appropriate levels of fanfare several senior Google figures announced the Agent2Agent protocol that would usher &#8220;a new era of Agent Interoperability&#8221;. </p><p>I am not exactly sure what was the <em>old</em> era of Agent Interoperability, but rest assured a new era is on its way. Excellent. </p><p>So let&#8217;s dive in. What exactly does the A2A protocol give us:</p><blockquote><p>&#8220;The A2A protocol will allow AI agents to communicate with each other, securely exchange information, and coordinate actions on top of various enterprise platforms or applications.&#8221; - <a href="https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/">Google Blog Post</a></p></blockquote><p>Ok. That sounds interesting. Why do we need it?</p><blockquote><p>&#8220;To maximize the benefits from agentic AI, it is critical for these agents to be able to collaborate in a dynamic, multi-agent ecosystem across siloed data systems and applications. Enabling agents to interoperate with each other, even if they were built by different vendors or in a different framework, will increase autonomy and multiply productivity gains, while lowering long-term costs.&#8221; - <a href="https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/">Google Blog Post</a></p></blockquote><p>Got it. So Google (and several other providers) are envisioning a future where (LLM-powered) applications will be reaching across silos and asking other (LLM-powered) applications to perform tasks for them. An application to application interface if you will - an Application Programming Interface even or API for short. Fascinating.</p><p>Anybody else trying this? Here is something:</p><blockquote><p>the essential infrastructure to integrate AI agents with your entire IT ecosystem, allowing them to autonomously interact with APIs and drive efficient, high-quality customer interactions at scale. This ensures that AI agents can operate effectively within a secure and governed environment, enhancing overall business agility and performance.</p></blockquote><p>This is <a href="https://www.mulesoft.com/platform/agentforce">Mulesoft Anypoint</a> - from Salesforce. Salesforce also also supports the A2A protocol. That&#8217;s a bit confusing. So which one do we use, when? Products like Mulesoft have been around for ages. Why exactly do we need a specific Agent 2 Agent protocol? Seems we&#8217;ve been grappling with integration issues for a while here. </p><p>I know! A2A is when we want to get two Agents to interoperate. We&#8217;ve figured out that an API or data integration could not possibly solve the problem. The methods that we have decades of experience on simply will not cut it. We need to apply agentic capabilities and we need LLM-powered decision making. Got it. </p><p>Then I am sure that A2A will be solving tough, agent specific issues. Not the age-old API and data integration issues. This is about a new agentic era! Luckily we even know what the issues are. After all we&#8217;ve been doing multi-agent systems research for several decades. We know what are the challenges around discovery, trustworthiness, reliability and so on. So let&#8217;s not rush to conclusions about A2A. Let&#8217;s dive into the protocol and figure out how it addresses the really tough issues. I am pumped!</p><h2>A2A design principles</h2><p>Let&#8217;s have a look at the design principles for A2A. </p><blockquote><p><em><strong>Embrace Agentic Capabilities</strong>. A2A focuses on enabling agents to collaborate in their natural, unstructured modalities, even when they don&#8217;t share memory, tools and context. We are enabling true multi-agent scenarios without limiting an agent to a &#8220;tool.&#8221;</em></p></blockquote><p>Ok. This is good. You&#8217;d hardly want an Agent 2 Agent protocol to not embrace agent capabilities. It would really, truly be just an API then I guess. I am not sure what a &#8220;natural, unstructured&#8221; modality is but sounds great. Ok. What next?</p><blockquote><p><strong>Build on existing standards:</strong> The protocol is built on top of existing, popular standards including HTTP, SSE, JSON-RPC, which means it&#8217;s easier to integrate with existing IT stacks businesses already use daily.</p></blockquote><p>Cool, cool, cool. Imagine if we had to put in place entirely new standards - like a REPLACEMENT FOR HTTP. No, this is wise, very wise. </p><blockquote><p><strong>Secure by default</strong>: A2A is designed to support enterprise-grade authentication and authorization, with parity to OpenAPI&#8217;s authentication schemes at launch.</p></blockquote><p>Authentication and authorisation. Perfect. Good thing this didn&#8217;t slip by. I am sure this will pretty much handle any and all security issues with having two LLM-powered applications hallucinate their way through complex tasks. Fantabulous. </p><blockquote><p><strong>Support for long-running tasks:</strong> We designed A2A to be flexible and support scenarios where it excels at completing everything from quick tasks to deep research that may take hours and or even days when humans are in the loop. Throughout this process, A2A can provide real-time feedback, notifications, and state updates to its users.</p></blockquote><p>Interesting. Yup - sometimes <s>applications</s>agents do take a long time. We need a way to keep in touch and update other <s>applications</s>agents on what is happening. Looking forward to diving into the detail of this one. I wonder if I could <em>subscribe</em> to some sort of event and then be <em>notified</em>. That would be revolutionary. </p><blockquote><p><strong>Modality agnostic:</strong> The agentic world isn&#8217;t limited to just text, which is why we&#8217;ve designed A2A to support various modalities, including audio and video streaming.</p></blockquote><p>Nice. This clears things up about that natural, unstructured modality mentioned in the first principle. I thought that might mean different modalities but actually this <em>modality agnostic</em> thing means that. So that must mean something else. Super clear. </p><p>So we want to support agents, securely, using existing standards as they work on tasks and notify each other. These are all good requirements, there are some <em>light </em> design principles somewhere there but it reads mostly as specific decisions rather than guidelines on how to make decisions. Anyway, let&#8217;s not get caught up in definitions. How about a bit more detail on how all this works. </p><h2>How A2A actually works</h2><p>The <a href="https://github.com/google/A2A?tab=readme-ov-file">Github repo</a> and this <a href="https://storage.googleapis.com/gweb-developer-goog-blog-assets/original_videos/A2A_demo_v4.mp4">video</a> provide some good insight into the detail. </p><h3>Discovery of Agents and Skills</h3><p>The first bit is discovery. My agent will identify a need for a skill or a task that it (presumably) cannot perform itself and will look into a &#8220;tool registry&#8221; for agents that might be able to perform this task. Agents describe their capabilities through <a href="https://google.github.io/A2A/#/documentation?id=agent-card">AgentCards</a>. These cards describe both practical information to facilitate communication and information exchange and, crucially, the <em>skills</em>, of the agent. The things it is able to do. </p><p>One interesting question here is how do we unambiguously describe a skill without assuming some prior knowledge or assumption. How do we know that the advertised skill is exactly what we need. When we hire people we put out a job ad, we look at a CV but then we run an interview to make sure the candidate it a good fit. With agents if they say they can "run HR activities such as candidate sourcing&#8221; are we just going to assume that that will do everything we need or is there an implicit assumption that this will be good enough. If we have implicit interoperability assumptions in place where does that leave the dynamic, autonomous agentic future?</p><p>Also, it&#8217;s not clear how my agent decides that it cannot do the task itself but I assume that this is similar to choosing a tool, only in this case the tool is a whole other agent and the protocol is A2A rather than standard tool calling. </p><p>How agents are dynamically discovered is also not tackled. For the time being I assume that if this was used in production setting it would simply be a case of letting all the agents involved know of all the agents they might use. </p><h3>Tasks</h3><p>The chosen agent will then be assigned a <a href="https://google.github.io/A2A/#/documentation?id=task">task</a>. I am genuinely disappointed that the protocol designers went with the idea of a task rather than a goal. The whole thing about agents is goal-directed behavior. This was the one chance to make it clearly conceptually different. A goal is a description of a desirable state of affairs, a task is something a bit lower level. Oh well. This is where we are. We assign a task to an  agent and a task can be in one of various states. Much like a job. I wonder if there are decades of tools helping us to manage long-running jobs in the IT world? </p><h3>Collaboration</h3><p>The other agent will then review the task and may have some questions. There is a specific &#8216;<em>input-required&#8217;</em> state for this. That&#8217;s it. There is not really much to say about the collaboration mechanism other that it is <a href="https://google.github.io/A2A/#/documentation?id=multi-turn-conversations">very rudimentary </a>right now. It is geared to some quite specific tasks that seem like they are making some strong implicit assumptions. It&#8217;s not as if there is decades of research in speech act theory and <a href="https://en.wikipedia.org/wiki/Agent_Communications_Language">agent communication languages</a> based on speech-act theory from the 1970s. </p><h2>Conclusions</h2><p>It might not feel like it given this post but I do think work on agent-to-agent protocols is very useful. However, this feels rushed. It feels like a proof-of-concept of some technical capabilities without the deep thought that needs to go into an actual protocol. It feels like Google rushing to occupy some space because they have to <em>win</em> everywhere and Anthropic got a head start with <a href="https://www.anthropic.com/news/model-context-protocol">MCP</a>. </p><p>The thing that gets me is that what will inevitably happen is that some people will get excited, some thing will be implemented and will not work and then the assumption will be that agent-to-agent collaboration is not useful rather than the thinking behind this particular protocol was too rushed. </p><p>This protocol does not solve any of the hard conceptual problems. It provides an interesting point of reference to think about them. An organisation as big as Google, however, should be driving for setting standards in a more organised way. We have the W3C the IETF, there are bodies that can support standards in a much more structured way. It is clear that international organisations are not popular these days but that doesn&#8217;t mean they are not useful. It seems that we just prefer to barge straight forward, push out some half-baked idea, generate the hype that comes with it (a new era!) and then deal with the consequences later. </p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[[AI Agent Diaries] The LLM industry's dirty secret and what to do about it.]]></title><description><![CDATA[The LLM industry is experiencing a mini ChatGPT-in-reverse moment this week with The Atlantic's release of the LibGen search tool, showing how much pirated material is used for LLM training.]]></description><link>https://www.agentsdecoded.com/p/ai-agent-diaries-the-llm-industrys</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agent-diaries-the-llm-industrys</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 22 Mar 2025 08:25:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-Dk_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There are two errors with the title of this diary entry. Firstly, the LLM industry doesn&#8217;t just have one dirty secret, it has many dirty secrets. Secondly, they are not secrets. We all know about all the issues. However, it often takes some catalyst moment for something that people do know about to become something that is discussed widely enough and maybe lead to enough people <em>caring about. </em></p><p>The dirty secret we are going to focus on today is the LLM industry&#8217;s use of copyrighted content. We&#8217;ve all know about this for a very long time. For example, there is the infamous Murati (ex-OpenAI CTO) <a href="https://www.wsj.com/video/series/joanna-stern-personal-technology/openai-made-me-crazy-videosthen-the-cto-answered-most-of-my-questions/C2188768-D570-4456-8574-9941D4F9D7E2">interview with the Wall Street Journal</a> where the incredibly capable and intelligent ex-CTO of OpenAI somehow was &#8220;not sure&#8221; what data was used to train <a href="https://openai.com/sora/">Sora</a> or couldn&#8217;t quite remember. </p><p>Since then we&#8217;ve learned a few more things. Most interestingly, there is the <a href="https://www.theguardian.com/technology/2025/jan/10/mark-zuckerberg-meta-books-ai-models-sarah-silverman">internal Meta communications</a> released following legal action. Through their embarrassingly superficial debates we find out that engineers explicitly discussed whether it would be ok to use LibGen - a library containing millions of pirated books - because Meta&#8217;s competitors where doing it as well. Ultimately, Mark Zuckerberg approved the use of pirated material - living his &#8220;move fast and break things&#8221; motto to the full. </p><p>Then this week The Atlantic released<a href="https://www.theatlantic.com/technology/archive/2025/03/search-libgen-data-set/682094/"> a simple search tool</a> that allowed anyone to see if their own books where used to train LLMs. In what I like to think of as reverse-ChatGPT moment people where horrified to find out explicitly that their hard work was used to train LLMs posting the results on social media in absolute outrage. The enthusiasm of people when OpenAI productised LLMs through ChatGPT and made it plain for all to see their capabilities turned to horror when the Atlantic productised the ability to determine the level of piracy that went on behind the scenes to train those LLMs.</p><p>Of course I did a little search for myself and while academic papers showed up that were public domain there is also a book that very much should not be public domain. I spent a good percentage of weekends in 2018 and 2019 writing <a href="https://link.springer.com/book/10.1007/978-1-4842-5476-9">that book</a>. I don&#8217;t particularly care that its contents are publicly available for anyone to read without paying but I do care that it was used by organisations with access to billions of dollars without permission when they will very much complain if I don&#8217;t pay what is due for access to their services. Now, my main job is not writing. I can&#8217;t imagine how horrified people whose livelihood depends on authoring must feel.  </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-Dk_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-Dk_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 424w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 848w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-Dk_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png" width="1434" height="1058" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1058,&quot;width&quot;:1434,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:203043,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/159602260?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-Dk_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 424w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 848w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!-Dk_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f8ff038-6f37-4d65-b22d-70bd2de6955b_1434x1058.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">My own contribution to the LLM industry</figcaption></figure></div><h2>The LLM industry are Modern Day Robber Barons</h2><p>Let&#8217;s call a spade a spade. </p><p>What the LLM industry is doing is not dissimilar to what robber barons across the ages have done. Sam Altman, Mark Zuckerberg and others have identified a natural resource, content produced over centuries of human labour, and a means to exploit it, productised LLMs. </p><p>The analogies to Rockefeller, Carnegie are so plain. <strong>Monopolistic practices</strong> (<a href="https://geopoliticaleconomy.com/2025/02/03/us-ai-monopoly-unipolar-world-china/">they want their government to protect them from external competition</a>), <strong>vertical integration</strong> (they manage every stage from the data centres to the end-user products maximising profit extraction), <strong>political influence</strong> (literally falling over themselves to gain access to government), while at the same time arguing for <strong>minimal environmental regulations</strong>, <strong>exploiting labour</strong>, <strong>displacing existing industries and people</strong> (replacing jobs) and <strong>financial manipulations</strong> (like the <a href="https://www.bbc.com/news/articles/cy4m84d2xz2o">ridiculous White House $500 billion Stargate announcement</a> or Elon Musk&#8217;s combination of <a href="https://www.ft.com/content/d4616dec-c4c7-417f-8549-134710bbc5b1">X with X.ai</a>). </p><h2>What are we to do about it</h2><p>There is a series of arguments against LLMs, from how the technology is fundamentally broken to how it is unethical. They all have merit for me. However, I also know that the technology is extremely useful and it is here, it is open source and it is not going away. </p><p>We cannot ban LLMs and we should not. I know I am over-simplifying but LLMs per se are not the problem. The problems are the afore mentioned robber barons.  They are breaking the rules without consequences and they are then using the gains from the rule-breaking to facilitate further rule breaking or rule re-writing for their benefit. </p><p>I think there are two things that need to be done. </p><ol><li><p><strong>The Robber Barrons and the LLM industry needs to pay.</strong> There&#8217;s a number of mechanisms and far more qualified people to define how but a combination of  ongoing taxation, specific fines, and prosecution to the extend the law allows should take place. The robber barons need to be stopped. </p></li><li><p><strong>The technology requires regulation.</strong> Yes this will slow it down, yes some use cases are not going to be viable, yes some startups are going to have to stop or change what they are trying to do. Yes the regulation is not going to be perfect.  That is ok. We&#8217;ve broken enough things. Time we started fixing some of them. </p></li></ol><p>Current developments in the world and especially the US give us a pretty good idea where the &#8220;move fast and break things&#8221; ethos takes us. Perhaps its time we tried something different. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI Agents' Reality Gap: 7 Ways LLM-powered AI Agents can Create Real-World Problems (and what to do about it)]]></title><description><![CDATA["When Sarah decided to plan her dream European vacation, she turned to an AI-powered travel app that promised to create the perfect personalized itinerary.]]></description><link>https://www.agentsdecoded.com/p/ai-agents-reality-gap-7-ways-llm</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agents-reality-gap-7-ways-llm</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Tue, 04 Mar 2025 07:41:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Vu_b!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Vu_b!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Vu_b!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Vu_b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:235051,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.agentsdecoded.com/i/158351954?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Vu_b!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vu_b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59061404-074a-428a-a685-90684b3754e1_1792x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>"When Sarah decided to plan her dream European vacation, she turned to an AI-powered travel app that promised to create the perfect personalized itinerary. Made by a startup touting its amazing agentic capabilities, the app's recommendations looked flawless on screen. A sleek, curated mix of attractions, accommodations, and experiences that seemed to perfectly match her preferences and budget. With confidence in the AI's suggestions, Sarah booked everything through the app in a single click.</em></p><p><em>Reality proved disappointingly different. Her &#8216;best-value&#8217; five-star hotel was in an unsafe area and not quite like the photos. Although the hotel photo gallery did have beautiful photos of beaches and sunsets, those were 20 minutes away through some shady neighborhoods, not the view from the hotel itself. The AI agent simply assumed that any photos on the hotel page described the hotel and the AI summary presented to Sarah didn't provide that crucial bit of detail.</em></p><p><em>The "off-the-beaten path" adventures turned out to be below average experiences that were very to get to without further expensive transportation and weird times the AI Agent did not account for. The "authentic" restaurants were either tourist traps, unsanitary or overpriced. A relaxing holiday turned into a series of stressful challenges as Sarah negotiated online with the AI Agent that continued to essentially gaslight her about the fact that the holiday was exactly as she asked for and no refund or change was possible."</em></p><div><hr></div><p>Anyone who has worked with LLMs over the last few years will recognise how all these issues can arise. Hallucinations, misinterpretation, lack of context, overconfidence and feedback loops that solidify already bad assumptions.</p><p>I think the potential of LLMs to drive positive experiences is significant, however we will not get there by pretending that it is all rosy and amazing. We need to face the limitations head on and be clear about where LLMs and LLM-powered AI Agents can drive value and where, plainly, it may not be worth the effort.</p><p>Don&#8217;t get me wrong. I am not arguing that we should ignore or not use this technology. After all, I make my living selling solutions that are powered by LLMs. Actually, that is not true. I make my living solving problems for clients and if I cannot clearly demonstrate when and how LLMs can actually help them and provide real value to them I am not going to be left with any clients.</p><p>So let's dive in. Here are 7 ways LLM-powered AI agents can go wrong and some suggestions about what to do about it.</p><h2>1. Data Misinterpretation</h2><p>When AI agents misunderstand or miscontextualize input data, they can make dangerous recommendations. Making leaps from adjectives to real-world actions without any careful checking is a risk. For instance, an AI agent might encounter a hotel's self-description claiming "steps away from the beach" and recommend it to a traveler seeking beachfront accommodation, when in reality the property is a 20-minute walk from the shore&#8212;something a human would quickly spot by checking a map or reading between the lines of guest reviews mentioning the "beach shuttle service." Similarly, when a restaurant describes itself as offering "authentic local cuisine with a modern twist," the AI might present this marketing language as objective fact to a traveler seeking traditional local food, missing the indication that the dishes have been significantly modified from their traditional preparation. In both cases, the AI jumps from descriptive language to concrete recommendations without the crucial step of verification and contextual understanding.</p><p>Combining LLMs with more explicit reasoning and carefully curated databases can address some of these issues. We need to pay close attention to the data that we make available and how and whether we are able to verify data or make explicit assumptions. This is one of the challenges of Operator / Computer Use style agents that operate on the web. How will an Operator Agent judge the trustworthiness of a hotel accommodation website without upfront curation?</p><h2>2. False Information Generation</h2><p>AI agents can confidently present completely fabricated information as fact. There are ways to limit this by injecting specific knowledge in prompts, and while the very big issues can be caught - i.e., you can easily prevent an AI Agent from booking into a non-existent hotel - there is a long tail of additional cases that may not be that easy to catch.</p><p>For instance, an AI travel agent might accurately list a hotel's address and basic amenities but fabricate details about the "award-winning breakfast buffet" or "recent renovation" simply because the language for that is statistically possible. A financial AI might reference real market trends but invent specific analyst recommendations or company statements. These smaller fabrications are harder to detect because they are mixed with truth. LLMs, by their very nature, are particularly adept at wrapping up lies with truths preparing the infamous <a href="https://en.wikipedia.org/wiki/Truth_sandwich#:~:text=A%20truth%20sandwich%20is%20a,story%20by%20again%20presenting%20truth.">truth sandwiches</a>, making automated fact-checking more challenging. Even more concerning, these fabrications can compound over time - an AI might read and incorporate another AI's fabricated details, creating a web of false information that becomes increasingly difficult to untangle.</p><p>If we are dealing with fully automated services and cannot rely on humans reviewing and fact-checking then it is important to start out quite conservative in the problem we are trying to solve. If we use LLMs to generate content with the specific aim of "selling" to users (e.g. adding prompt instructions such as "make it enticing") we are essentially encouraging them to lie.</p><h2>3. Flawed Logic Chains</h2><p>Even with accurate data, AI agents can make invalid logical connections that lead to harmful decisions. Consider a travel booking AI that notices a user often books early morning flights and concludes they "prefer" early departures, when in reality they've been choosing those flights reluctantly due to price and lack of time to search for other options. The AI might then rigidly book 6 AM departures for their vacation, ignoring better-timed options because it misinterpreted past behavior as preference. We&#8217;ve all experienced even the most sophisticated recommendation engines (e.g. Netflix &#8220;watch next&#8221;) get completely derailed because of some random choices of ours. In each situation, the AI draws seemingly logical but fundamentally flawed conclusions from accurate data points.</p><p>In building systems we need to have clarity of where we are using LLMs (or any other system really) to make assumptions about preferences and be able to explicitly trace and audit decision-making. In addition, the user needs to be made aware of the choices the AI Agent took. To be clear, the current trend of presenting the user with the &#8220;thoughts&#8221; of reasoning LLMs such as o3 is not the answer. We need to invest in user experience that is actually helpful, not this information overload torrent of thoughts.</p><h2>4. Bias Amplification and Lack of Cultural Context</h2><p>AI agents can magnify societal biases present in their training data. A lot has been written and said around this, especially when it comes to cases such as a lending system systematically undervaluing certain neighborhoods based on historical redlining data, but even if your AI Agent is not operating in these more sensitive areas, there is still room for bias.</p><p>Consider an AI travel assistant that provides superficial cultural advice like "always haggle in Morocco - it's what locals expect," reducing rich cultural nuances to stereotypes. This advice might be technically "correct" in some contexts but lacks the crucial nuance of when haggling is appropriate (markets vs. established businesses), how it should be conducted respectfully, and how it varies by region and situation. The traveler following this advice might offend locals by haggling in inappropriate settings or miss authentic experiences by approaching every interaction through this oversimplified lens.</p><p>Similarly, an educational AI might provide lower-quality assistance to non-native English speakers by misinterpreting grammatical patterns from other languages as indicators of lower comprehension, leading it to oversimplify explanations unnecessarily.</p><p>To address these issues, AI systems need diverse training data and explicit checks for biased outputs. More importantly, they need transparency about their limitations &#8211; being upfront about the demographics and contexts where their recommendations are most reliable and where users should seek additional perspectives. Designing AI agents with cultural humility rather than cultural authority allows them to make suggestions while acknowledging the limits of their understanding. Including diverse human perspectives in the development and testing of AI systems is essential, not optional.</p><h2>5. Overconfident Speculation</h2><p>AI agents often make bold claims beyond their actual understanding, presenting speculative information with unwarranted certainty. This problem is particularly concerning because confidence in language is not correlated with accuracy &#8211; in fact, LLMs often express their most inaccurate statements with the highest verbal certainty.</p><p>Consider a medical triage AI that confidently tells a user, "Your symptoms strongly indicate Guillain-Barr&#233; syndrome," when faced with common symptoms like tingling and weakness that could signify dozens of conditions. The AI has no ability to acknowledge the true statistical rarity of this diagnosis or understand that physicians would typically rule out more common conditions first. This false certainty could cause unnecessary panic or dangerous self-diagnosis.</p><p>Similarly, a legal AI might definitively state, "Based on Section 501(c)(3), your organization qualifies for tax exemption" without understanding the complex, multi-faceted nature of tax law determinations that even specialized attorneys approach with caution. An investment AI might declare with certainty that a specific stock will increase in value, without the capability to understand that market predictions inherently contain uncertainty.</p><p>The problem extends beyond professional domains. An AI travel agent might confidently assure a user that a certain trail is "perfectly safe and suitable for beginners" based on a few positive reviews, lacking the judgment to consider seasonal variations, the traveler's true fitness level, or recent weather events that might affect conditions.</p><p>Addressing this issue requires calibrating AI confidence to match actual certainty. Technical approaches include training models to express appropriate uncertainty and implementing confidence scores that reflect real statistical reliability rather than linguistic confidence. From a design perspective, AI systems should be built to clearly distinguish between facts, recommendations, and speculations in their outputs. Most importantly, users should be educated about the limitations of AI knowledge and encouraged to verify critical information from authoritative sources.</p><h2>6. Time and Context Errors</h2><p>AI agents can fail to account for when and where their knowledge applies, treating outdated or contextually limited information as universally applicable. This temporal and contextual myopia creates a dangerous reality gap between AI recommendations and real-world conditions.</p><p>An investment AI trained on pre-pandemic economic data might confidently recommend heavy investment in commercial real estate or airline stocks, completely missing the transformative impact of remote work and travel restrictions. Unlike human advisors who lived through these changes and intuitively adjust their thinking, the AI continued applying outdated patterns until explicitly retrained. If you would prefer a more time-relevant example think of the current geopolitical shifts between the USA, Canada, and Europe on tariffs. What will an LLM trained on data up until December 2024 think of the February 2025 world?</p><p>Solving this problem requires multiple approaches. Technical solutions include regular retraining, explicitly tagging information with temporal validity indicators, and designing systems that can actively seek current information rather than relying solely on their training data. From a design perspective, AI agents should express appropriate caution when making time-sensitive recommendations and, ideally, verify current conditions before taking actions with real-world consequences.</p><p>Organizations deploying AI agents need robust update mechanisms and clear documentation about the temporal limitations of their systems. Most importantly, users should be encouraged to verify time-sensitive information and taught to recognize when an AI might be operating with outdated assumptions.</p><h2>7. Error Cascades</h2><p>Small AI mistakes can compound into major problems through repeated interactions and feedback loops, creating cascading failures that grow in severity over time. These error cascades are particularly insidious because they can emerge from systems that appear to be functioning correctly on a transaction-by-transaction basis.</p><p>For example, consider an AI travel assistant planning a multi-city international trip. The assistant initially misinterprets a traveler's preference for "morning arrivals" as simply "before noon" rather than "early morning." Based on this minor misunderstanding, it books a flight landing at 11:45 AM. This creates a tight connection for the next transportation leg, which then causes the traveler to miss a pre-arranged tour. The AI, seeing the missed tour, attempts to compensate by rebooking for the next day, which conflicts with other reservations. Each small misalignment compounds into increasingly significant disruptions, ultimately derailing the entire trip itinerary - all stemming from a single initial misinterpretation that seemed inconsequential when viewed in isolation.</p><p>Addressing error cascades requires systems designed with feedback loops and corrective mechanisms. Technical approaches include diversity in AI decision-making (using multiple models or approaches to cross-check recommendations), implementing breakers that prevent rapid sequential actions without verification, and creating monitoring systems that can detect patterns of small errors before they compound.</p><p>From a design perspective, AI systems should maintain uncertainty estimates that increase when sequential decisions are made without human verification, gradually becoming more cautious rather than more confident in extended autonomous operation. Most importantly, meaningful human oversight&#8212;not just theoretical ability to intervene, but practical mechanisms for review and correction&#8212;must be maintained in systems where cascading errors could cause significant harm.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Addressing the Reality Gap: Practical Solutions</h2><p>As AI agents become more pervasive, it's crucial that we understand and mitigate these risks. Misinterpretations, hallucinations, and flawed logic from LLMs can lead to real-world consequences when actualized by AI agents - what starts as a small error can snowball into a serious deception.</p><p>Addressing these challenges requires a multi-pronged approach across technical, design, and organizational dimensions:</p><h3>Technical Solutions</h3><ol><li><p><strong>Grounding and Verification Systems</strong>: Implement mechanisms for AI agents to verify facts against trusted knowledge bases before taking action or making recommendations.</p></li><li><p><strong>Uncertainty Quantification</strong>: Develop better methods for AI systems to express appropriate levels of uncertainty rather than false confidence.</p></li><li><p><strong>Guardrails and Safety Measures</strong>: Create circuit breakers that prevent AI agents from making high-stakes decisions without verification.</p></li><li><p><strong>Diverse Model Ensembles</strong>: Use multiple models with different training approaches to cross-check conclusions and reduce systematic errors.</p></li><li><p><strong>Regular Retraining and Updates</strong>: Establish protocols for keeping AI knowledge current, especially for time-sensitive applications.</p></li></ol><h3>Design Solutions</h3><ol><li><p><strong>Transparent Reasoning</strong>: Make AI decision processes visible to users in accessible, non-overwhelming ways.</p></li><li><p><strong>Appropriate Agency Distribution</strong>: Design systems that maintain human control over high-consequence decisions while automating low-risk tasks.</p></li><li><p><strong>Friction by Design</strong>: Intentionally add verification steps before irreversible actions, especially in high-stakes contexts.</p></li><li><p><strong>Cultural Humility</strong>: Frame AI recommendations as suggestions rather than authoritative pronouncements in domains with significant cultural variation.</p></li><li><p><strong>Feedback Integration</strong>: Create intuitive ways for users to correct AI errors and have those corrections reflected in future interactions.</p></li></ol><h3>Organizational Practices</h3><ol><li><p><strong>Human-in-the-Loop Processes</strong>: Maintain appropriate human oversight, especially for consequential decisions or when working with vulnerable populations.</p></li><li><p><strong>Clear Error Handling</strong>: Develop specific protocols for addressing AI mistakes, including compensation mechanisms for affected users.</p></li><li><p><strong>Realistic Marketing</strong>: Avoid overselling AI capabilities, setting appropriate expectations about system limitations.</p></li><li><p><strong>Rigorous Testing</strong>: Test AI agents with diverse scenarios and edge cases, particularly looking for systematic failure modes.</p></li><li><p><strong>Continuous Monitoring</strong>: Implement systems to detect patterns of errors before they create significant harms.</p></li></ol><p>By approaching these issues systematically rather than treating them as inevitable "growing pains" of AI adoption, we can develop AI agents that genuinely serve user needs rather than creating new problems. The goal isn't perfect AI &#8211; it's AI that fails gracefully, transparently, and with minimal harm when it inevitably makes mistakes while providing clear overall value to both the user and the organization.</p><h2>Conclusions</h2><p>I&#8217;ll close by focussing on what I think is the biggest ask of practitioners and organisations. Move slowly and do less. AI technologies push us to think big and be ambitious. They seem so capable that they tug at very basic human instincts of wanting to do more faster, especially in competitive settings. Instead, we need to cultivate our ability to constrain and contain. While there are multiple mitigation strategies, sometimes they may simply not be enough or be worth it. I know this is a big ask - we have developed a culture or more, faster - everything is transactional. Bigger buildings, faster cars, bodies that never age. Any risk is acceptable if it might lead to a big enough upside. However, we need to recognise that we are hitting limits and exhausting resources. </p><blockquote><p>Perhaps our technological future lies not in more impressive cathedrals but in creating resilient neighborhoods of focused, well-understood tools each with clear boundaries and intentional design.</p></blockquote><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[[AI Agent Diaries] Thinking Fast and Slow About AI Agent Development]]></title><description><![CDATA[Using Daniel Kahneman's "Fast and Slow" framework to recognise the traps our mind sets ourselves when working on AI projects.]]></description><link>https://www.agentsdecoded.com/p/ai-agent-diaries-thinking-fast-and</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agent-diaries-thinking-fast-and</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 08 Feb 2025 11:13:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6zuD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6zuD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6zuD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6zuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:90920,&quot;alt&quot;:&quot;Thinking Fast and Slow&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Thinking Fast and Slow" title="Thinking Fast and Slow" srcset="https://substackcdn.com/image/fetch/$s_!6zuD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6zuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f09d54a-33a0-4a9a-9440-72ae3952e67d_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Thinking Fast and Slow</figcaption></figure></div><p>This week my colleague - <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Maaike Coppens&quot;,&quot;id&quot;:13678644,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/028802b2-8b1e-4464-9f7b-f78a21e6f24e_372x370.jpeg&quot;,&quot;uuid&quot;:&quot;c5d705f7-bc48-4ebd-a86d-6b421deddacd&quot;}" data-component-name="MentionToDOM"></span> - started writing about AI Product Development. Her excellent post on the <a href="https://aiproductfocus.substack.com/p/things-your-ai-product-manager-is">challenges AI Product Managers face</a> inspired me to run a favorite thought exercise of mine - reviewing the challenges of a specific space through the lens of Daniel Kahneman's "Thinking, Fast and Slow". </p><p>In the book "<a href="https://en.wikipedia.org/wiki/Thinking,_Fast_and_Slow">Thinking, Fast and Slow</a>", psychologist Daniel Kahneman provides a conceptual framework that can help us better understand how we humans approach a wide range of issues. His basic tenet of two competing thought systems is particularly relevant when considering how we develop and deploy AI agents, especially in an environment of intense hype and competition.</p><p>System 1, as Kahneman calls it, is fast, instinctive, and emotional. System 2 is more deliberative, rational, and effortful. </p><blockquote><p>When it comes to AI, we often allow System 1 to lead us down the wrong path by anthropomorphizing AI systems or relating them to patterns we've seen in science fiction and popular media, rather than understanding their true capabilities and limitations. We do not give ourselves the time to engage with System 2 and think through the harder problems.</p></blockquote><p>AI projects, it could be argued, are almost perfectly designed to set these cognitive traps for our minds and lead us down the wrong path. Here are some examples.</p><h1>Framing</h1><p>What on the surface can appear as a simple request such as "we need an AI Agent," once unpacked, opens up a significant number of questions that go well beyond what people intuitively consider the domain of AI development. This framing significantly influences our subsequent choices. However, more often that not I come across teams that start by framing the problem of creating an AI agent in terms of "which large language model should we use" as opposed to "what user needs are we trying to address". The former will lead us down a very different (and not very useful) path.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5ZAf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5ZAf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5ZAf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg" width="388" height="285.43345543345544" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2410,&quot;width&quot;:3276,&quot;resizeWidth&quot;:388,&quot;bytes&quot;:1709614,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5ZAf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5ZAf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a0cb09d-768e-4f86-92b6-236249dbbb1e_3276x2410.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Framing influences our future choices - Photo by <a href="https://unsplash.com/@jaredd?utm_content=creditCopyText&amp;utm_medium=referral&amp;utm_source=unsplash">Jaredd Craig</a> on <a href="https://unsplash.com/photos/person-taking-photo-of-ship-hwru6PbAHgI?utm_content=creditCopyText&amp;utm_medium=referral&amp;utm_source=unsplash">Unsplash</a></figcaption></figure></div><p>For example, a System 1 response might be to immediately jump to implementing GPT-4o or, dare I say, DeepSeek because it's the most advanced or most talked about model available. However, a System 2 analysis might reveal that a smaller, fine-tuned model would be more appropriate for specific use cases like customer service queries about product specifications. </p><h1>Substitution Problem</h1><p>The rapidly evolving nature of AI technology means there is an unusually high level of noise within the available information. It's easy to be misled about the actual complexity of implementing AI solutions and become overly concerned with specific technical aspects.  Kahneman's substitution problem is particularly relevant here - we are naturally inclined to substitute complex questions (e.g., "How can we ensure consistent and reliable responses?") with apparently simpler, but less useful, questions (e.g., "What's the best prompting technique to use?").</p><p>Breaking through the noise to get to what is really useful challenges how we naturally think about AI.</p><h1>The Planning Fallacy and Optimism Bias</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wBAm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wBAm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wBAm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg" width="510" height="340.11675824175825" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/df2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:510,&quot;bytes&quot;:2905783,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wBAm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wBAm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf2f7f10-a38e-4e98-808e-f5bf4000bfd9_5402x3602.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Implementing AI solutions requires a variety of different disciplines to complete. AI strategy, prompt engineering, user experience design, data science, and software engineering all come into play. Each brings its own terminology and set of norms, and stakeholders are often asked to make quick decisions to keep the project on track and within budget. Unfortunately, as Kahneman explains, humans are afflicted by the planning fallacy. We tend to consistently underestimate the time required to develop and deploy AI systems, while optimism bias means we overestimate their benefits and capabilities.</p><h1>The Availability Heuristic and Risk Assessment </h1><p>Kahneman's research on the availability heuristic, our tendency to assess probability based on how easily examples come to mind, is also very relevant. When evaluating AI risks, teams often focus disproportionately on widely-publicized failures while overlooking more common but less sensational issues.</p><p>Teams might invest heavily in preventing chatbots from generating clearly inappropriate content (a highly publicized risk) while underinvesting in:</p><ul><li><p>Monitoring for subtle gender or racial bias in model responses (I know its not cool to talk about these things anymore but there is still there&#8230;)</p></li><li><p>Testing for degradation in model performance over time</p></li><li><p>Evaluating fairness across different user demographics (see earlier comment about uncool things)</p></li><li><p>Measuring and improving accuracy for less common but critical edge cases</p></li></ul><p>To counter this bias, teams should implement structured risk assessment frameworks that include:</p><ul><li><p>Systematic logging of all types of failures, not just dramatic ones</p></li><li><p>Regular analysis of "quiet" failures that don't make headlines</p></li><li><p>Quantitative metrics for measuring bias and fairness</p></li><li><p>Automated testing for common failure modes</p></li></ul><h1>Think Slow to Allow AI to Act Fast</h1><p>Understanding both how we think about AI and how users will interact with our AI systems is central to delivering successful AI projects. We improve our chances of success by recognizing that designing, planning, and deploying AI systems requires deliberative, rational, and effortful thought. A particularly challenging task these days where fast, impulsive and emotional action seems to be the norm. </p><p>We have to avoid numerous pitfalls and stereotypes and challenge our assumptions about AI to create something that achieves our goals. Framing these challenges as simply human traits that are perfectly normal and can be overcome by recognising them and talking about them helps teams reason about them more carefully. </p><p>Successful AI implementation requires a team that recognises our limitations as humans so that we can avoid these cognitive pitfalls. This might include comprehensive discovery workshops that directly address the framing issue by getting all stakeholders aligned on the actual problem we're trying to solve with AI. Regular technical debates and continuous learning ensure teams don't shy away from hard questions and avoid the substitution problem. Finally, cross-functional teams working closely with stakeholders through an appropriate methodology help keep the project state in check and maintain realistic planning focused on deliverable outcomes.</p><p>In short, we need to engage our deliberative and rational side when developing AI systems to produce solutions that are efficient, intuitive, and naturally interactive.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[[AI Agent Diaries] AI Maximalists Vs The People]]></title><description><![CDATA[The people are fighting back.]]></description><link>https://www.agentsdecoded.com/p/ai-agent-diaries-ai-maximalists-vs</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agent-diaries-ai-maximalists-vs</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 01 Feb 2025 08:40:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ng6i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ng6i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ng6i!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ng6i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg" width="1920" height="1281" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1281,&quot;width&quot;:1920,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1202157,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ng6i!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Ng6i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28fe5201-2afa-494d-8c71-628afab65ebf_1920x1281.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Choice and user agency - Photo by <a href="https://unsplash.com/@garri?utm_content=creditCopyText&amp;utm_medium=referral&amp;utm_source=unsplash">Vladislav Babienko</a> on <a href="https://unsplash.com/photos/man-standing-in-the-middle-of-woods-KTpSVEcU0XU?utm_content=creditCopyText&amp;utm_medium=referral&amp;utm_source=unsplash">Unsplash</a></figcaption></figure></div><p>While the AI experts and influencers are falling all over themselves trying to outdo each other with discussions and analysis of new LLM models (DeepSeek, <a href="https://qwenlm.github.io/blog/qwen2.5-max/">Alibaba</a>, <a href="https://mistral.ai/news/mistral-small-3/">Mistral</a>, <a href="https://openai.com/index/openai-o3-mini/">OpenAI o3</a>) a rebellion is brewing. From Reddit boards to Twitter and BlueSky threads, the people are fighting back. </p><p>Users are sharing elaborate workarounds to disable AI features in the products they have to use every day. Google Gemini - eager to summarise every single email, Apple Intelligence -   keen to let me know that a &#8220;long&#8221; update has arrived (&#129335;&#127997;&#8205;&#9794;&#65039;), Copilot in Word -  Clipy&#8217;s arrogant nephew. Perhaps the most telling one, though, are AI answers in Google search with users explaining how <a href="https://arstechnica.com/google/2025/01/just-give-me-the-fing-links-cursing-disables-googles-ai-overviews/">you can use swearing in your search phrase to switch off the feature</a>. The resistance isn't just digital fatigue; it's a visceral reaction to a maximalist view of AI - more AI always equals better. </p><p>Yet, as users desperately search for ways to reclaim their pre-AI interfaces, a parallel narrative unfolds. Writers are getting excited about how they can use AI not as a replacement for creativity but as a <a href="https://substack.com/home/post/p-155946216">sophisticated research assistant.</a> Scientists are accelerating drug discovery. Engineers are able to explore thousands of solutions in minutes.</p><p>This dichotomy, AI as an unwanted intruder Vs AI as a revolutionary tool, reveals a fundamental misunderstanding by tech companies of how humans want to interact with technology. </p><blockquote><p>The AI maximalists, in their fervor to integrate artificial intelligence into every digital crevice, have forgotten a basic principle: tools should serve their users, not the other way around.</p></blockquote><p>The evidence for this disconnect isn't just anecdotal. Recent research from Irrational Labs reveals a stark truth: explicitly labeling features as "AI-powered" actually decreases user trust and doesn't increase willingness to pay. Users aren't impressed by AI buzzwords; they're looking for concrete benefits and real solutions.</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:155781646,&quot;url&quot;:&quot;https://www.growthunhinged.com/p/ai-messaging-study&quot;,&quot;publication_id&quot;:311430,&quot;publication_name&quot;:&quot;Kyle Poyar&#8217;s Growth Unhinged&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c7accd6-4dcf-4f8a-a173-0cb8f9214fc7_500x500.png&quot;,&quot;title&quot;:&quot;Customers don't care about your AI feature&quot;,&quot;truncated_body_text&quot;:&quot;&#128075; Hi, it&#8217;s Kyle Poyar and welcome to Growth Unhinged, my weekly newsletter exploring the hidden playbooks behind the fastest-growing startups.&quot;,&quot;date&quot;:&quot;2025-01-29T11:50:45.546Z&quot;,&quot;like_count&quot;:105,&quot;comment_count&quot;:9,&quot;bylines&quot;:[{&quot;id&quot;:23170097,&quot;name&quot;:&quot;Kristen Berman&quot;,&quot;handle&quot;:&quot;kristenberman&quot;,&quot;previous_name&quot;:&quot;Product Teardowns&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b3eeb3c-4e65-414a-ba28-a76c355dba39_3000x3000.png&quot;,&quot;bio&quot;:&quot;I take apart popular products and reveal their core psychologies. Along the way, I uncover the secrets to their success and share how you can replicate their design principles in your work.&quot;,&quot;profile_set_up_at&quot;:&quot;2021-09-15T04:36:50.075Z&quot;,&quot;twitter_screen_name&quot;:&quot;bermster&quot;,&quot;is_guest&quot;:true,&quot;bestseller_tier&quot;:null,&quot;primaryPublicationId&quot;:476445,&quot;primaryPublicationName&quot;:&quot;Product Teardowns&quot;,&quot;primaryPublicationUrl&quot;:&quot;https://kristenberman.substack.com&quot;,&quot;primaryPublicationSubscribeUrl&quot;:&quot;https://kristenberman.substack.com/subscribe?&quot;}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:false,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.growthunhinged.com/p/ai-messaging-study?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!1TBj!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c7accd6-4dcf-4f8a-a173-0cb8f9214fc7_500x500.png"><span class="embedded-post-publication-name">Kyle Poyar&#8217;s Growth Unhinged</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">Customers don't care about your AI feature</div></div><div class="embedded-post-body">&#128075; Hi, it&#8217;s Kyle Poyar and welcome to Growth Unhinged, my weekly newsletter exploring the hidden playbooks behind the fastest-growing startups&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">a year ago &#183; 105 likes &#183; 9 comments &#183; Kristen Berman</div></a></div><p>As someone who <a href="https://opendialog.ai">sells tools</a> to embed AI everywhere, I should be the quintessential AI maximalist. But my position in the industry has taught me something crucial: long-term success in AI integration isn't about forcing technology into every possible corner of user experience. It's about thoughtful implementation that prioritizes user choice and concrete benefits.</p><p>The path forward should be clear. Tech companies need to stop treating AI like a marketing checkbox and start treating it like what it is: a powerful tool that users should be able to employ on their own terms. When AI features are optional, transparent, and clearly tied to specific benefits, users embrace them. When they're forced, hidden, and justified with vague promises of "enhancement," users revolt.</p><p>The future of AI won't be built by maximalists pushing for integration at any cost. It will be built by companies who understand that the most powerful feature they can offer isn't AI itself - it's the ability to choose when and how to use it. In the end, the difference between AI as an unwanted intruder and AI as a revolutionary tool comes down to one simple principle: respect for user agency.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Agents Decoded is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[[AI Agent Diaries] The One Where China Stood Up]]></title><description><![CDATA[Stargate, Operator, Billionaire spats, DeepSeek and Product Research]]></description><link>https://www.agentsdecoded.com/p/ai-agent-diaries-the-one-where-china</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agent-diaries-the-one-where-china</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 25 Jan 2025 09:19:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F65db99ed-e99b-43b2-bcd5-035e248384e9_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I started out AgentsDecoded thinking that in-between opinions, reviews of frameworks and pieces exploring the essence of agency I will include a weekly news column. After three weeks of doing that though I realised two things:</p><p>1. There is so much news every week I never feel I am actually doing a proper job of it.</p><p>2. There are already <a href="https://theagentarchitect.substack.com/p/the-ai-agent-architects-monday-morning-69b">enough</a> <a href="https://substack.com/@lastweekinai">other</a> newsletters <a href="https://whatdidopenaidothisweek.substack.com/">gathering</a> announcements and news, the world does not need another one. </p><p>What I think I will do instead that is hopefully more useful to you (and me), is use the weekly column cadence as a way to reflect on the week that has gone by. This will be a very specific point of view of a <a href="https://opendialog.ai">startup</a> co-founder and CPTO down in the trenches working with people to implement AI solutions in the real world, far away from the dizzy heights of the leading AI labs and research centres with billions in investment. </p><p>Here we go.</p><div><hr></div><p>Looking back at my week, we started off with some user research at work to gather data for how we should shape our product roadmap. It is always so powerful to get unbiased views of how people are looking to use AI Agents and where they are today. Two interesting perspectives are emerging. </p><p>1. <strong>There is a lot of opportunity in just</strong> <strong>helping companies upgrade from old Conversational AI tech </strong>. There are many mid-sized organisations that have built chatbots using the technologies post 2023, with standard ML intent classifiers and what is now <a href="https://www.linkedin.com/posts/kanesimms_genai-nlu-machinelearning-activity-7285190679448698881-hA5S?utm_source=share&amp;utm_medium=member_desktop">essentially obsolete</a> Conversational AI capability. All these organisations are now trying to figure out how to retool and move to technology stacks with GenAI but it is a slow process (certainly much slower than the rate at which LLMs are progressing). Definitely an opportunity for agencies out there to help people move up. </p><p>2. <strong>There is an army of consultant AI Agents coming</strong>. The vertical AI Agent space is ripe with opportunities. People with specific domain expertise (e.g. a specific niche of contract law) can develop solutions that combine their expertise with LLM capability - effectively turning what they know as an experienced consultant into a product. We will see a flurry of startups that are led by human consultants developing AI Agent Consultants and selling those (together with their standard services).</p><p>In the wider world of AI Agents three news stories dominated the agenda. </p><p>Let's start with the <strong>Stargate Project</strong>. <a href="https://openai.com/index/announcing-the-stargate-project/">Five hundred billion</a> (!) in AI infrastructure investment in the US, confusingly announced from the White House but actually funded entirely from private organisations, of which apparently only a small part is currently available if we are to believe what Elon Musk said, that is not related with the project and is annoyed because his enemy Sam Altman is involved. Confused? Yeah, me too. What it does tell us is that nations are getting competitive. They want to lead the AI race, and win the AI wars - whatever that is supposed to mean. </p><p>The second story, which provides an interesting counterpoint to the first one is that <a href="https://www.deepseek.com/">DeepSeek</a>, a Chinese lab that was previously mostly unknown to the western world and media, is <a href="https://substack.com/home/post/p-155354389?source=queue">crushing</a> <a href="https://www.interconnects.ai/p/deepseek-r1-recipe-for-o1">it</a> at that very <a href="https://www.linkedin.com/posts/emollick_this-is-a-big-deal-deepseek-is-the-first-activity-7288550050916429827-qvxs?utm_source=share&amp;utm_medium=member_desktop">AI race</a>. Releasing models that are cheaper, faster, better and open source. Sure they are standing on the shoulders of work done by US-based labs - but they are introducing innovations that are entirely their own. Europe and the US likes to tell itself that China just copies stuff, but that is clearly not the case. Among various geopolitical implications it also means that prices for access to LLM functionality are going to quickly start going down all around - which creates an interesting and challenging dynamic for the much more expensive to run US and European labs. </p><p>Finally, <a href="https://openai.com/index/introducing-operator/">OpenAI released Operator</a>. It's first AI agent. Essentially a more complete version of Computer Use similar to what Anthropic released last year. I saw two types of reaction online. "WOW!" and "meh". Ignoring all the hype coming from the increasingly large army of &#8220;AI influencers&#8221; there are good reasons for both excitement and reserve. As I mention <a href="https://www.linkedin.com/feed/update/urn:li:ugcPost:7288454252123881472/?commentUrn=urn%3Ali%3Acomment%3A%28ugcPost%3A7288454252123881472%2C7288459995988348928%29&amp;dashCommentUrn=urn%3Ali%3Afsd_comment%3A%287288459995988348928%2Curn%3Ali%3AugcPost%3A7288454252123881472%29">here</a>, while it is impressive stuff the uses cases and user experience are not quite there yet.</p><p>Much more has happened over just the past few days (e.g. Anthropic released <a href="https://www.anthropic.com/news/introducing-citations-api">Citations</a>) but I am already a bit over 700 words and that is the cap I set myself! Now, off for the weekend, for hopefully a couple of days where AI is not dominating what I am doing!</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading. To receive new posts and support my work, consider becoming a subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/p/ai-agent-diaries-the-one-where-china?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/p/ai-agent-diaries-the-one-where-china?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.agentsdecoded.com/p/ai-agent-diaries-the-one-where-china?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div>]]></content:encoded></item><item><title><![CDATA[AI Agents are not a technology. They are a strategy.]]></title><description><![CDATA[The debates rage on about what constitutes an AI Agent.]]></description><link>https://www.agentsdecoded.com/p/ai-agents-are-not-a-technology-they</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agents-are-not-a-technology-they</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Mon, 20 Jan 2025 16:42:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5-fs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5-fs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5-fs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5-fs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/da1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1118638,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5-fs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!5-fs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda1bbe52-737c-412e-bc98-0387538941c2_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The debates rage on about what constitutes an AI Agent. </p><p>Interestingly, a lot of the focus is on defining what should not be considered an agent. "If it's a single prompt, it's not an agent!", "If it's a workflow of prompts, it's not an agent!", "Unless its writing its own code it's not an agent!" and on it goes. </p><p><strong>Everyone seems to keep raising the ante in a race to distinguish themselves.</strong> In the middle of 2024 Hugging Face <a href="https://huggingface.co/blog/agents">defined agents</a> as "a program driven by an LLM". By the end of 2024 they <a href="https://huggingface.co/blog/smolagents">updated that definition</a> to say "LLM outputs control the workflow" and introduced a star system (3 stars is the max just like Michelin restaurants). You only get 3 stars if an LLM controls "iteration and program continuation". Anthropic are <a href="https://www.agentsdecoded.com/p/an-analysis-of-anthropics-guide-to">similarly</a> trying to create tiers of agents, although the definitions are somewhat confusing. If you have a workflow of prompts it&#8217;s ok to call it an &#8220;agentic&#8221; system but what you created is not an AI agent. Agents need to "direct their own processes". </p><p><strong>Understandably, anyone looking into the field is left baffled.</strong> They want to do the right thing by their team and business. Perhaps they've been tasked to figure out how their organisation is not going to be rendered obsolete by the relentless forward march of AI and they've heard that AI Agents are the future. Surely they want 3-star agents. Why settle for anything less? </p><p>I've been working in the agent space for a long time. I co-authored a book in 2004 called "Agent-based software development". I am a fan of the approach. However, <strong>I am worried that currently technology companies are spending too much time trying to out-agent the other agents and not enough time worrying about how AI Agents can solve actual, real business problems.</strong> </p><p>If you are trying to figure out agents let me offer a different way of thinking about them. <strong>Stop thinking about how they are built. Start thinking about what it is that they enable. Specifically, what is it that they automate and how much can you delegate to them.</strong> </p><p><strong>AI Agents are the high-level building blocks of an automation strategy.</strong> They first encapsulate a conceptual approach - "we will delegate work to software" and, eventually, a technological framework. Conceptually, each agent has specific capabilities and specific goals. They have a degree of autonomy or, using term I prefer, self-direction in determining how to deploy their capabilities to achieve their goals. The assignment of goals and the degree of self-direction required are going to depend on what you are trying to automate and what is currently achievable (safely) with the technologies at hand. Through the agent abstraction you can plug into a rich ecosystem and research field that offers decades of outputs. Frameworks on how to define and manage goals, determine how best to have these programs communicate and collaborate, how to think about scaling, trust, safety, planning, negotiation and even argumentation. </p><p>What of all these technologies do you need to use today? How you go about building them? Depends on the task and the breadth, ambition and risk appetite of your strategy. They don't need to be any more or less complicated that what they need to be. <strong>There is no AI Agent license exam to pass.</strong> Don't worry about the definitions. 3 stars is not better than 0 stars. Look into the technologies, frameworks and patterns and find the one that will achieve your goals with the least amount of risk possible and in the most efficient way possible. Whether it calls itself a workflow, a multi-agent system or an agentless architecture does not matter. Your strategy does not change.</p><p>Your strategy is one of moving away from whatever way you are executing work today to one where AI Agents are embedded in your infrastructure to execute that work more efficiently and (hopefully) better. While technical distinctions between agent implementations matter - they inform what's feasible, maintainable, and scalable - they should flow from your strategic needs rather than define them. Similarly, while frameworks and standardization efforts can provide valuable guidance and common ground for evaluation, they should serve as tools for achieving your goals rather than constraints that dictate them. <strong>AI Agents are a strategy first, a technology second</strong> - but a successful strategy must be grounded in a clear understanding of the technological landscape and its practical implications for your specific context.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Agents Decoded! Subscribe to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[[AI Agent News] Jan 11 - Jan 17 2025]]></title><description><![CDATA[A collection of news and announcements from the AI Agent world over the past week (with opining thrown in as an extra bonus): Microsoft, OpenAI, Anthropic, Deloitte.]]></description><link>https://www.agentsdecoded.com/p/ai-agent-news-jan-11-jan-17-2025</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/ai-agent-news-jan-11-jan-17-2025</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Sat, 18 Jan 2025 07:45:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hDWs!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F65db99ed-e99b-43b2-bcd5-035e248384e9_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Microsoft Autogen</strong> has <a href="https://www.microsoft.com/en-us/research/blog/autogen-v0-4-reimagining-the-foundation-of-agentic-ai-for-scale-extensibility-and-robustness/">a new version</a>, primarily focussed on making it easier to build, monitor and improve applications. The underlying concepts remain largely the same but the framework is moving away from a way to experiment with ideas to something that directly addresses the needs of implementation at scale, including a low-code UI.</p><div><hr></div><p><strong>Anthropic</strong> reported it <a href="https://www.anthropic.com/news/anthropic-achieves-iso-42001-certification-for-responsible-ai">achieved ISO 42001 certification for responsible AI</a>. &#8220;What does that mean?&#8221;, I hear you wondering. From the ISO website itself:</p><blockquote><p>The ISO/IEC 42001 standard offers organizations the comprehensive guidance they need to use AI responsibly and effectively, even as the technology is rapidly evolving. Designed to cover the various aspects of artificial intelligence and the different applications an organization may be running, it provides an integrated approach to managing AI projects, from risk assessment to effective treatment of these risks.</p></blockquote><p>&#8220;What does that still mean?&#8221;, I still hear you say. Anthropic is trying to distinguish itself from other competitors by spending more time ensuring that appropriate governance is in place for its systems. The hard truth, however, is that while a better level of governance is probably a good thing it is not a signal of any underlying change in the fundamental challenges these technologies pose. At least Anthropic is willing to do some of the extra work though. Having gone through ISO processes, none of this is exciting or fulfilling but it can uncover gaps and it does force an organisation to think through at least some issues. </p><div><hr></div><p><strong>OpenAI </strong>has a beta release of <a href="https://help.openai.com/en/articles/10291617-scheduled-tasks-in-chatgpt">scheduled Tasks in ChatGPT</a>. This is definitely very early beta and quite underwhelming. While a few are hailing it as the start of the &#8220;agentic&#8221; era at OpenAI and certainly it is a small first step, it feels more like someone decided they had to test the basic &#8220;hello world&#8221; implementation of task capability at scale before they move on to something more inspiring, even if it would inevitably get some bad press. </p><div class="comment" data-attrs="{&quot;url&quot;:&quot;https://open.substack.com/home&quot;,&quot;commentId&quot;:86877013,&quot;comment&quot;:{&quot;id&quot;:86877013,&quot;date&quot;:&quot;2025-01-18T07:11:54.152Z&quot;,&quot;edited_at&quot;:null,&quot;body&quot;:&quot;While I don&#8217;t like dunking on people&#8217;s work these examples of ChatGPT Tasks from the OpenAI announcement really are uninspiring  to say the least. They are all things you can accomplish differently  with &#8220;remind me about my mom&#8217;s birthday&#8221; almost feeling like a joke. Why do I need an LLM to remind me about a birthday as opposed to a simple calendar notification?&quot;,&quot;body_json&quot;:{&quot;type&quot;:&quot;doc&quot;,&quot;attrs&quot;:{&quot;schemaVersion&quot;:&quot;v1&quot;},&quot;content&quot;:[{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;While I don&#8217;t like dunking on people&#8217;s work these examples of ChatGPT Tasks from the OpenAI announcement really are uninspiring  to say the least. They are all things you can accomplish differently  with &#8220;remind me about my mom&#8217;s birthday&#8221; almost feeling like a joke. Why do I need an LLM to remind me about a birthday as opposed to a simple calendar notification?&quot;}]}]},&quot;restacks&quot;:0,&quot;reaction_count&quot;:0,&quot;attachments&quot;:[{&quot;id&quot;:&quot;3abacdc8-9758-4561-b193-c31a72a9353b&quot;,&quot;type&quot;:&quot;image&quot;,&quot;imageUrl&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a20c23d-4461-406b-a00e-2d8b4ddc83e3_982x210.png&quot;,&quot;imageWidth&quot;:982,&quot;imageHeight&quot;:210,&quot;explicit&quot;:false}],&quot;name&quot;:&quot;Ronald Ashri&quot;,&quot;user_id&quot;:109341091,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/6256182a-3cc0-42ac-85fb-6bd7591f5539_1168x1170.jpeg&quot;,&quot;user_bestseller_tier&quot;:null}}" data-component-name="CommentPlaceholder"></div><div><hr></div><p>In the interesting reading department we have <a href="https://arxiv.org/pdf/2501.07238">this paper</a> from <strong>Microsoft</strong> on &#8220;Lessons from Red Teaming 100 Generative AI Products&#8221;. It does not disappoint. What does disappoint, is <a href="https://www2.deloitte.com/content/dam/Deloitte/us/Documents/gen-ai-multi-agents-pov-2.pdf">this report from </a><strong><a href="https://www2.deloitte.com/content/dam/Deloitte/us/Documents/gen-ai-multi-agents-pov-2.pdf">Deloitte</a></strong> on agents and multi-agent systems. Aside from introducing intractable terminology such as making &#8220;cognitive leaps&#8221; and &#8220;reinventing processes&#8221; it then presents the most classic IT support triage scenario as a potential implementation of a multi-agent system! Look, I am all in on the benefits of these technologies but the fact that we seem incapable to temper expectations in actual big complex organisations is going to hinder sensible adoption. Reinventing processes and making cognitive leaps is not going to come through the current wave of multi-agent technologies. It can come if we use these technologies to free people up to think longer-term so they can actually do some re-inventing. Ok, rant over. </p><p>That wraps up another week in the agent world. Don&#8217;t forget to share and subscribe. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/p/ai-agent-news-jan-11-jan-17-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.agentsdecoded.com/p/ai-agent-news-jan-11-jan-17-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.agentsdecoded.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[[Framework Review] Hugging Face's smolagents Framework]]></title><description><![CDATA[Reviewing Hugging Face's SmolAgents Framework, what view of agency it has, what is interesting about it and where it might go next.]]></description><link>https://www.agentsdecoded.com/p/framework-review-hugging-faces-smolagents</link><guid isPermaLink="false">https://www.agentsdecoded.com/p/framework-review-hugging-faces-smolagents</guid><dc:creator><![CDATA[Ronald Ashri]]></dc:creator><pubDate>Tue, 14 Jan 2025 16:01:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-uEG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>At the very end of 2024 (literally Dec 31st) Hugging Face released <a href="https://huggingface.co/blog/smolagents#%F0%9F%A4%94-what-are-agents">smolagents</a> - &#8220;a simple library to build AI agents&#8221;. As is quite obvious from the name this is all about simplicity. Given that Hugging Face already had a library to support &#8220;agentic&#8221; work (Transformer Agents), it&#8217;s interesting to dive into this library, figure out what their motivations are, what it says about what AI Agents are and when you might want to consider using it. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-uEG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-uEG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 424w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 848w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 1272w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-uEG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png" width="1456" height="933" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:933,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:378612,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-uEG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 424w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 848w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 1272w, https://substackcdn.com/image/fetch/$s_!-uEG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b7768c0-aa9a-43f5-b543-b15ef5b0b5b0_2366x1516.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Smolagents really is small &#128578;</figcaption></figure></div><p></p><h1>Why was smolagents created?</h1><p>This is Hugging Face&#8217;s successor to <a href="https://huggingface.co/blog/agents">Transformer Agents</a>, a library that was getting updates all through 2024.  I assume that the learnings from that library led the HF team to feel the need to start fresh. It looks like smolagents also captures HF&#8217;s convinction that having the LLMs write the code to execute functions, rather than generating the JSON that then is then used to execute function , is the way to go. </p><p>The following seem to be the high-level reasons:</p><p><strong>Re-focus on simplicity:</strong> The logic for agents is contained within a relatively small amount of code, aiming to keep it easily understandable.</p><p><strong>First-class support for Code Agents:</strong> smolagents prioritises agents that write their actions in code. Code is seen as the most effective way to express actions a computer can perform, offering advantages over JSON-like snippets in terms of composability, object management, generality and its presence in LLM training data. </p><p><strong>Hub Integrations:</strong> smolagents allows for sharing and loading tools to and from the Hugging Face Hub, and it is expected that this functionality will expand.</p><p><strong>Support for any LLM:</strong> It supports various models, including those hosted on the Hub, as well as models from providers such as OpenAI and Anthropic.</p><h1>Agent Design Perspective: smolagents view on Agency and Autonomy</h1><div class="pullquote"><p><em>For every framework we analyse at AgentsDecoded we look for a deeper understanding of how AI Agents can be thought of and engineered. Our reference is <a href="https://www.agentsdecoded.com/p/a-universal-framework-for-understanding">our own conceptual framework for understanding agency and autonomy</a> that provides a consistent set of terms to fall back on.</em></p></div><p>It&#8217;s interesting to track the evolution of HF&#8217;s thinking around agents. </p><p>From the May 13th, 2024 blog post <a href="https://huggingface.co/blog/agents">announcing Transformer Agents</a> we have agents defined as:</p><blockquote><p>(..) an <strong>agent</strong>, which is just a program driven by an LLM. The agent is empowered by <strong>tools</strong> to help it perform actions. When the agent needs a specific skill to solve a particular problem, it relies on an appropriate tool from its toolbox.</p></blockquote><p>A quite simple definition, which in today&#8217;s &#8220;discource&#8221; (i.e. people throwing random definitions on Substack, LinkedIn, X, bluesky, etc) would not quite cut mustard. </p><p>Fast forward to Dec 31st, 2024 and the introduction of smolagents, while the broad agent definition remains largely the same:</p><blockquote><p>AI Agents are <strong>programs where LLM outputs control the workflow</strong>.</p></blockquote><p>it is further refined to create a star system of agent levels. These refinement is indicative of a move in the overall market to tighten up the definition. While six months ago, people where happy to just call anything an agent essentially, now it is more about saying &#8220;watch out - not all agents are equal&#8221;. </p><p>Here's a breakdown of the star ratings from HF:</p><p><strong>&#9734; (Zero Stars):</strong> At this level, the LLM output has <strong>no impact</strong> on the program flow. This is described as a "Simple processor", and an example is the pattern </p><blockquote><p><code>process_llm_output(llm_response)</code>. </p></blockquote><p><strong>&#9733; (One Star):</strong> Here, the LLM output determines the <strong>basic control flow</strong>. This is referred to as a "Router", and an example pattern is </p><blockquote><p><code>if</code> <code>llm_decision(): path_a() else: path_b()</code>. </p></blockquote><p>The LLM makes a basic decision that changes which path the program takes.</p><p><strong>&#9733;&#9733;(Two Stars):</strong> In this case, the LLM output determines the <strong>function execution</strong>. This is described as "Tool call", and an example pattern is </p><blockquote><p><code>run_function(llm_chosen_tool, llm_chosen_args)</code>. </p></blockquote><p>The LLM chooses which tool (function) to run, and what arguments to provide to it.</p><p><strong>&#9733;&#9733;&#9733; (Three Stars):</strong> This level represents a multi-step agent where the LLM output controls <strong>iteration and program continuation</strong>. The pattern here is:</p><blockquote><p> <code>while llm_should_continue(): execute_next_step()</code><strong>. </strong></p></blockquote><p>The LLM decides when to continue the loop or move on to the next step based on its observations.</p><p><strong>&#9733;&#9733;&#9733; (Three Stars):</strong> This rating also applies when one agentic workflow can start another agentic workflow (not sure why this is not a higher rating but it is not my rating system!), which is described as "Multi-Agent". The example pattern is </p><blockquote><p><code>if llm_trigger(): execute_agent()</code></p></blockquote><p>The LLM decides when to trigger another agent.</p><h2>Analysis</h2><p>While I can appreciate the pragmatic view of the different levels, things only really start getting interesting at two stars because that is then we start to potentially explicitly represent goals. Using the <a href="https://www.agentsdecoded.com/p/a-universal-framework-for-understanding">AgentsDecoded framework</a>, at zero and one stars we are dealing with what we call passive agents. These are software programs that are are ascribed a goal by the user but don&#8217;t have any internal representation of that goal. At two stars we can start thinking of self-directed active agents. The software program is <em>assigned</em> a goal and it needs to figure out how to accomplish it. The simple function calling agent has reduced self-direction, it can only request that some other system execute a function while the three star agent has an explicit representation of a goal and can write its own code to effect change in the environment. </p><p>I don&#8217;t agree that having a single agent write its own code merits the same star level as having a multi-agent system. That is a fundamentally different type of system where issues of trust, shared knowledge and observability become far more important.</p><p>Overall, the challenge, from a definition perspective, that smolagents faces is that it ties things too specifically to the how something is done (i.e. how we are accomplishing something in code) and does not take a broader perspective on agents as goal-driven systems. </p><h1>When to use smolagents (for Builders)</h1><p>smolagents makes two important bets: </p><ol><li><p><strong>Keep it simple, stupid</strong>. Instead of coming with a bunch of opinions about how you should be designing agent-based systems it focusses on the minimum required and leaves the rest to you as a builder. This is a double-edge sword. If you have a good idea of what the rest should be then this is a great starting point. However, if you don&#8217;t know what you are missing you may be in for some interesting (and costly) discoveries along the way!</p></li><li><p><strong>Writing code directly is better than writing JSON instructions.</strong> This is a big deal for smolagents. Instead of using JSON or text-based descriptions, <strong>the agent writes actions directly in code, typically Python</strong>. This is considered a significant advantage over other methods of defining actions.</p></li></ol><p>The latter point is interesting so here&#8217;s a breakdown of what this means and why it's interesting:</p><p><strong>Advantages of Code-Based Actions</strong>:</p><p><strong>Representation in Training Data</strong>: LLMs are already trained on vast amounts of code, making them better suited to generating actions in code rather than in other formats like JSON.</p><p><strong>Composability</strong>: Code can be nested and reused, similar to defining functions in programming. This contrasts with JSON, where such nesting and reusability are much harder to implement.</p><p><strong>Object Management</strong>: Storing the outputs of actions (like an image from an image generator tool) is more easily managed with code. In JSON, this is very difficult.</p><p><strong>Generality</strong>: Code is designed to express any action a computer can perform. This makes it a highly general and flexible way to define actions.</p><p><strong>Contrast with JSON-based Actions</strong>: In contrast to code-based actions, other agents like the ReactJsonAgent in the Transformers Agents 2.0 framework, generate actions in JSON format. While JSON is a common way to represent data, smolagents argues that it&#8217;s not ideal for representing executable actions due to limitations in the areas described above.</p><p>While the arguments are compelling, I don&#8217;t think the jury is out on this. As every, as a builder you need to apply nuance and look at your specific context. I can imagine multiple situations where a layer of decoupling between the definition of what to do and the actual execution is useful, especially when dealing with more complex organisational systems. Nevertheless, it is certainly a worthwhile argument to make and explore. To help explore it here is an outlive of some of the disadvantages of this approach. </p><p><strong>Disadvantages of Code-Based Actions</strong></p><p><strong>Complexity for Less Powerful LLMs:</strong> As the smolagent authors note , <strong>less powerful LLMs may struggle to generate good code</strong>. Xode-based actions rely on the LLM having a strong ability to generate correct and executable code, which may not always be the case, especially with smaller or less capable models. In contrast, JSON structures are simpler to generate, making them more suitable for LLMs that may not be as proficient at coding.</p><p><strong>Potential for Errors and Security Risks</strong>: While smolagents uses sandboxed environments for security, the generation and execution of code introduce <strong>potential risks and error points</strong> that are not present when using simpler JSON or text-based actions. If the LLM generates code with errors, it can cause the agent to malfunction. Furthermore, the code could contain unintended or malicious actions if not properly sandboxed and validated. </p><p><strong>Increased Cognitive Load on the LLM:</strong> Generating code, as opposed to JSON, could place a <strong>higher cognitive load</strong> on the LLM, which could lead to slower performance, particularly with complex tasks or less capable models. </p><p><strong>Debugging and Transparency:</strong> While smolagents emphasizes simplicity, debugging issues with code-based actions can be more difficult than debugging JSON-based actions. <strong>Errors in the generated code could be harder to diagnose and fix compared to simpler action formats</strong> such as JSON, requiring more expertise and effort. </p><p><strong>Limited Flexibility for Non-Programmers:</strong> While code provides generality, it also requires some degree of coding knowledge. This might make it <strong>less accessible for users without programming skills</strong>, potentially limiting the usage of code-based agents for non-programmers. </p><p>As you can see while having LLMs write and execute the code is exciting it does come at a cost. </p><h1>What smolagents adds to our understanding of Agent Systems (for Thinkers)</h1><p>For me the underlying framework is not really interesting from an agent systems perspective and, to be fair, it is not supposed to it! This is smolagents after all, not bigandcomplexagents with lots of opinions. The most interesting aspect is the code execution vs task definition part. If I draw some parallels to the <a href="https://www.agentsdecoded.com/p/agents-all-the-way-down">three layers of agency </a>I defined in a previous post, where you have emergent agency (coming from the LLM), structured agency (coming from a control framework) and perceived agency (the actual interface) what smolagents is doing is pushing more of the agency into the emergent part. It is relinquishing control and allowing the LLM to actually write and execute the code. </p><h1>What smolagents means for organisations (for Product Owners, CxOs)</h1><p>I have to admit a certain level of bias here. I work with organisations in regulated environments. I just cannot imagine these clients, that already have difficulty accepting basic LLM functionality, signing up to an agent framework where the agents write and execute their own code. In addition, there is a big step from what smolagents currently does to everything you need for an agent framework running a production-level application. As such it is hard for me to see this as anything other than an interesting experimental approach. </p><p>Having said that, I can imagine that some organisations with a bigger appetite for risk and their own specific knowledge and opinions around what else they want to build on top of smolagents could adopt this library as a great kernel from which to start. </p><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.agentsdecoded.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Agents Decoded! This is my way of keeping up to date with the world of AI Agents. The goal is to provide pragmatic, actionable information and avoid hype.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item></channel></rss>