<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The Compounding Founder]]></title><description><![CDATA[Build once, extract forever. Notes from a solo founder turning work into compounding assets.]]></description><link>https://www.thecompoundingfounder.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Ay6o!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F365e2f47-28e8-428d-9c67-44413c616ce1_1287x1287.png</url><title>The Compounding Founder</title><link>https://www.thecompoundingfounder.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 13 Apr 2026 02:04:04 GMT</lastBuildDate><atom:link href="https://www.thecompoundingfounder.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Eduardo]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[thecompoundingfounder@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[thecompoundingfounder@substack.com]]></itunes:email><itunes:name><![CDATA[Eduardo]]></itunes:name></itunes:owner><itunes:author><![CDATA[Eduardo]]></itunes:author><googleplay:owner><![CDATA[thecompoundingfounder@substack.com]]></googleplay:owner><googleplay:email><![CDATA[thecompoundingfounder@substack.com]]></googleplay:email><googleplay:author><![CDATA[Eduardo]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Your business isn't queryable (and your agents know it)]]></title><description><![CDATA[Every business framework you studied was designed for humans in meetings. This is the operating layer for what comes next.]]></description><link>https://www.thecompoundingfounder.com/p/your-business-isnt-queryable-and</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/your-business-isnt-queryable-and</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Sun, 12 Apr 2026 02:18:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!05c0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!05c0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!05c0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 424w, https://substackcdn.com/image/fetch/$s_!05c0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 848w, https://substackcdn.com/image/fetch/$s_!05c0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 1272w, https://substackcdn.com/image/fetch/$s_!05c0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!05c0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png" width="1456" height="585" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:585,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:898710,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.thecompoundingfounder.com/i/193913239?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!05c0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 424w, https://substackcdn.com/image/fetch/$s_!05c0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 848w, https://substackcdn.com/image/fetch/$s_!05c0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 1272w, https://substackcdn.com/image/fetch/$s_!05c0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd4e4dd-c1c6-4f3a-837e-2f8c39f97a0d_1632x656.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>During my masters program it wasn&#8217;t uncommon to spend a week evaluating a single framework. TOGAF for enterprise architecture. ITIL for service management. COSO for internal controls. RACI for decision rights. Lean for process improvement. BIZBOK for the gaps between all of them.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.thecompoundingfounder.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">The Compounding Founder is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>They gave me vocabulary. Mental models. A sense of where the edges were. Useful, the way a map of a city is useful before you&#8217;ve walked any of the streets.</p><p>Then I started running businesses with agents. And I realized the map was drawn for pedestrians.</p><p>Every one of those frameworks has the same hidden assumption baked in: humans are doing the work, humans have bounded attention, and the framework exists to coordinate humans across that bound. RACI tells you who&#8217;s responsible because, without it, two people will assume the other one did it. ITIL prescribes change advisory boards because humans make bad decisions under pressure and a slow, structured process catches errors. TOGAF produces architecture diagrams because the people who need to act on a decision aren&#8217;t in the room when it&#8217;s made.</p><p>Take away the human bandwidth constraint and most of what a business framework does becomes ceremony.</p><p>Agents don&#8217;t forget what was decided last Tuesday, as long as it&#8217;s in their context window. They can query 47 tables before you finish reading the meeting agenda, but they&#8217;ll hallucinate a number if the table doesn&#8217;t exist. They have no persistent memory across sessions. Their intelligence is jagged: superhuman at some tasks, confidently wrong about basic things.</p><p>The honest framing: </p><blockquote><p>Humans are bad at attention and memory across time. Agents are bad at persistent learning and consistency across sessions. A framework that only compensates for human bandwidth limits is the wrong framework. But agents still need structure &#8212; just structure that&#8217;s queryable rather than coordinative.</p></blockquote><p>The ceremony isn&#8217;t what makes the frameworks valuable. The data model underneath is.</p><p>A few weeks before I started drafting this post, I read Jack Dorsey and Roelof Botha&#8217;s <a href="https://sequoiacap.com/article/from-hierarchy-to-intelligence/">piece for Sequoia</a>, published in March 2026. The core argument landed hard: hierarchy isn&#8217;t an ideology, it&#8217;s an information routing protocol. It was invented to move decisions through organizations that couldn&#8217;t move information fast enough. Middle management existed to pre-compute decisions at scale. Every reorg since, from matrix structures to Holacracy to Spotify squads, rearranged the routing without replacing what was doing the routing. </p><blockquote><p>The copilot framing gets this wrong too: putting AI inside the existing structure improves throughput without touching the architecture.</p></blockquote><p>Their answer is a company world model: atomic capabilities, an intelligence layer that composes them, and a small number of human roles. No product manager decides what to build. The intelligence layer recognizes the moment and composes what&#8217;s needed.</p><p>But the piece stops at the architectural sketch. How does the world model stay accurate over time? What happens when declared state drifts from observed state? Who&#8217;s accountable when the model is wrong? There&#8217;s no answer for drift detection, no audit trail, no enforcement mechanism. The vision is clear. The operating manual doesn&#8217;t exist.</p><p>So I built one.</p><div><hr></div><h2>The four things left standing</h2><p>So I started asking a different question about every framework: strip away the meetings, the diagrams, the certifications, and the coordination rituals. What&#8217;s left?</p><p>Almost always the same four things.</p><p><strong>Tables.</strong> Every framework secretly assumes certain entities and relationships: services, incidents, and change requests (ITIL); decisions and responsible parties (RACI); controls and the risks they&#8217;re meant to catch (COSO). Those assumptions are tables. The entities are rows. The relationships are foreign keys.</p><p><strong>Constraints.</strong> Every framework has things it forbids. ITIL says a change can&#8217;t deploy without an owner. COSO says a control can&#8217;t be owned by the person it governs. RACI says every decision needs exactly one accountable party. Those prohibitions are NOT NULL columns, CHECK constraints, and pre-action validators. They&#8217;re not policies written in a manual. They&#8217;re schema. And when an agent tries to violate one, the database rejects the write before it lands. The enforcement is mechanical, not supervisory.</p><p><strong>Loops.</strong> Every framework has a measure-decide-act cycle. Lean&#8217;s Plan-Do-Check-Act. ITIL&#8217;s incident-to-resolution flow. OKR check-ins. DMAIC&#8217;s Define-Measure-Analyze-Improve-Control. Each of those cycles is a scheduled agent reading from an event queue, detecting something, and writing a response. The meeting where humans do this is the slow version of the loop. The database trigger and the scheduled job are the fast version.</p><p><strong>Queries.</strong> Every framework teaches humans to ask certain questions. &#8220;Who owns this decision?&#8221; &#8220;What&#8217;s the current status of this incident?&#8221; &#8220;Is this OKR on track?&#8221; Those questions are saved views. They&#8217;re SQL. They don&#8217;t need to live in a meeting agenda. They run at 7:15 AM and the results are waiting for you when you open your laptop.</p><p>The test I use: if a framework&#8217;s value can&#8217;t be expressed as tables, constraints, scheduled jobs, and saved queries, it&#8217;s ceremony. It&#8217;s not something an agent can use.</p><div><hr></div><h2>Capability modeling: the first translation</h2><p>The framework that survives this filter most cleanly is capability modeling, and it&#8217;s worth understanding why.</p><p>A capability is a stable description of something a business can do. &#8220;Publish weekly content&#8221; is a capability. How it gets delivered changes: the skill changes, the operator changes, the tools change. The capability stays the same because it describes the function, not the implementation. In my model, a capability is what you get when a skill meets an operator. The skill is the how. The operator is the who. The capability is the what.</p><p>This matters because a single capability can have mixed operators: human decides the surface, agent researches, agent distributes, human audits. The capability is stable. The operator assignment shifts as trust and reliability change.</p><p>In an agent-native business, capability modeling is the spine. Every skill maps to a capability it delivers. Every OKR targets a capability. Every KPI tree measures one. Every decision affects one. When something breaks, you query which capabilities are affected. When you&#8217;re adding a new agent, you check which capabilities are already covered. The map isn&#8217;t a diagram. It&#8217;s a table with a parent_id column for hierarchy and a junction table linking skills to capabilities.</p><p>The consulting exercise becomes a schema. The PowerPoint becomes a query. And the constraint layer rejects a capability record without a verified skill link. The enforcement is structural, not a review meeting.</p><div class="callout-block" data-callout="true"><p>If you use Claude Code, Cursor, Codex, etc&#8230;., you already do this for your codebase. <a href="http://CLAUDE.md">CLAUDE.md</a>/<a href="https://agents.md">AGENTS.md </a>tells agents what the repo is, what conventions to follow, where to find things. Your business needs the same declaration: what it can do, what it&#8217;s trying to achieve, what rules govern it, and how to tell when it&#8217;s drifting. Capabilities are the spine of that declaration. Goals measure the spine. Constraints protect it.</p></div><div><hr></div><h2>OKRs as queries</h2><p>The second framework worth examining closely is OKRs, because the agent-native version of it is almost unrecognizable from the original.</p><p>In an agent-native business, a key result is a SQL query that returns a number. Not a description of what the number measures. Not a link to a spreadsheet someone is updating. An actual query, stored in a column called <code>query</code>, that runs on a schedule and writes its output to <code>current_value</code>. The OKR check-in isn&#8217;t a meeting where someone reports progress. It&#8217;s a scheduled job that ran at 7:15 AM and already wrote the result. Progress is always real. It&#8217;s never self-reported.</p><p>&#8220;Reach 500 free subscribers by July 15th&#8221; becomes a row in an <code>objectives</code> table and a row in a <code>key_results</code> table where the query column is something like <code>SELECT COUNT(*) FROM subscribers WHERE tier = 'free'</code>. A scheduled job runs it every morning. The current value is always accurate. If the key result falls behind, an anomaly detector writes a gap. The gap routes to the operator. No meeting required.</p><p>You&#8217;re removing the gap between declared progress and observed progress. In a human OKR system, those two things drift apart because humans self-report. Here, the only progress that exists is the progress the query can measure. That constraint forces you to be specific about what you&#8217;re trying to accomplish.</p><p>&#8220;Grow the audience&#8221; can&#8217;t be a key result because you can&#8217;t write a query for it. &#8220;500 free subscribers by July 15th&#8221; can. The discipline the agent requires is more honest than the discipline the quarterly meeting requires.</p><div><hr></div><h2>The three states</h2><p>Three states matter in any agent-native business, and conflating them is the most common mistake. Most agent systems only track one. That gap is what breaks things.</p><p>Two weeks after we declared The Compounding Founder&#8217;s publishing cadence as weekly, we had two gaps in the observed record. The config file still said weekly. Nobody flagged it. That&#8217;s what having only declared state looks like.</p><p><strong>Reference state</strong> is what should exist according to some standard or template. If you&#8217;ve adopted a framework, the reference state is what that framework says a healthy implementation looks like. If you&#8217;re using an industry benchmark for open rates, the reference state is the benchmark.</p><p><strong>Declared state</strong> is what you claim exists. Your config file. Your business declaration. What you told the agent at startup.</p><p><strong>Observed state</strong> is what evidence shows actually happened. Metrics. Agent logs. Decisions with recorded outcomes. Briefings with timestamps.</p><p>Drift between these three states is the signal. When your declared state says you publish weekly and your observed state shows two gaps in the last six weeks, that&#8217;s drift. When your reference state says every decision needs an accountable owner and your observed state shows twelve decisions with no owner logged, that&#8217;s drift. The job of the retrospective loop isn&#8217;t to run the business. It&#8217;s to find the drift and surface it.</p><p>Most agent systems only track declared state. They have the config, the prompt, the system message, and they treat it as the truth. No observed state to compare against. No reference state to know what good looks like. When things go wrong, there&#8217;s no structural way to find out. A human checks in and notices. Or doesn&#8217;t notice for a while.</p><p>Building all three states, and treating the drift between them as first-class information, is what separates a business that agents can run from one they can only assist.</p><div><hr></div><h2>Build the spine first</h2><p>This doesn&#8217;t require building everything at once. I&#8217;ve walked you through two translations in this post: capability modeling and OKRs as queries. The full set I work with is seven (also DMAIC retrospective loops, KPI trees, decision rights, three lines of defense, and ITIL services). They aren&#8217;t a mandate. They&#8217;re a menu. Capability modeling is the only one that&#8217;s mandatory from the start, because it&#8217;s the spine everything else hangs from. The others come in as the business scales and the need becomes real.</p><p>A business in its first month needs capability modeling and maybe a simple OKR query for its one key metric. It doesn&#8217;t need a full ITIL incident management system or formal three-lines-of-defense segregation of duties. Those come later, when agents are making enough consequential decisions that oversight needs to be structural.</p><p>The dependency order matters too. OKRs and KPI trees both depend on having capability modeling in place, because both need something stable to target and measure. Decision rights depend on having a clear capability map because rights are scoped to capabilities. Three lines of defense depend on decision rights because you can&#8217;t segregate duties you haven&#8217;t defined. DMAIC depends on all of it, because the retrospective loop needs something to learn from.</p><p>Spine first. Diagnostics. Controls. Then the learning loop.</p><div><hr></div><h2>Task-capable, business-blind</h2><p>The point of all of this isn&#8217;t to rebuild your business in Postgres. A Postgres schema and an MCP server are not what I&#8217;m recommending here.</p><p>The point is to ask the question that most businesses using agents haven&#8217;t asked: what would your business look like if it were a queryable, auditable entity instead of a collection of documents and meetings and conversations that live partly in the heads of the human operators?</p><p>Right now, your agents are task-capable and business-blind. They can write the email, draft the post, run the analysis. They can&#8217;t tell you which decision from six weeks ago is causing the problem today. They can&#8217;t surface the drift between what you said you&#8217;d do and what actually happened. They can&#8217;t find the gap before it becomes a crisis. Not because they lack intelligence. Because there&#8217;s nothing to query.</p><p>Start with the spine. Write down what your business can do in stable terms. Not the tools you use. Not the agents you run. Not the current process. The capabilities. That&#8217;s one column in a spreadsheet or five lines in a markdown file. Everything in this post hangs from that list.</p><p>Dorsey and Botha described the destination: a company organized as an intelligence rather than a hierarchy. This framework is an attempt at the operating manual.</p><p>In the next post, I&#8217;ll show you what happens when you actually build this. A Postgres brain, an MCP server, and the morning an operator agent found six problems at 7:15 AM without being asked. That&#8217;s not a debugging story. That&#8217;s a control plane working.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.thecompoundingfounder.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">The Compounding Founder is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[You Did Not Delegate to Your Agent. You Abandoned It.]]></title><description><![CDATA[Delegation is not a trust problem. It is a specification problem.]]></description><link>https://www.thecompoundingfounder.com/p/you-did-not-delegate-to-your-agent</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/you-did-not-delegate-to-your-agent</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Tue, 07 Apr 2026 03:45:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2Ni6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2Ni6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2Ni6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2Ni6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg" width="1456" height="585" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:585,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:498805,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.thecompoundingfounder.com/i/193423884?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2Ni6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2Ni6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3d43f59-88a4-4a00-ae08-0073fb468c63_1632x656.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Priya runs a fourteen-person design studio in Toronto. Last Tuesday at 9:14 PM she forwarded an email to her operations manager with one line: &#8220;Can you handle this?&#8221; The email contained a client request for a revised project timeline, a budget question, and an implicit complaint about response times. Her operations manager responded the next morning. He answered the budget question, missed the timeline request, and did not register the complaint at all.</p><p>Priya did not have a trust problem. She had a specification problem. She gave a competent person a task with no structure, no context about what mattered, and no definition of what &#8220;handled&#8221; meant. He did what any reasonable person would do: he answered the easiest question and moved on.</p><p>This is exactly what most founders do with AI agents. They call it delegation. It is actually abandonment.</p><div><hr></div><h3>The Conventional Argument</h3><p>The conversation around AI delegation follows a predictable script. On one side: agents are not ready. They hallucinate. They lack judgment. You cannot hand them anything consequential because they will get it wrong and you will not catch it in time. On the other side: agents are already better than junior employees. Just set them loose and review the output.</p><p>Both sides share the same assumption. Delegation is primarily a question of capability. Can the agent do the work? Is it smart enough? Reliable enough?</p><p>This framing is not wrong. It is incomplete.</p><div><hr></div><h3>The Dismantle</h3><p>Think about the last time you delegated something important to a human. Not a task &#8212; a function. Accounting. Customer support. Content production. Social media.</p><p>You did not hand them the job and walk away. You gave them:</p><ul><li><p>a definition of what good looks like</p></li><li><p>the metrics you would use to know if it was working</p></li><li><p>the boundaries of what they could decide without asking</p></li><li><p>the context they would need to make those decisions well</p></li><li><p>a cadence for checking in</p></li></ul><p>When the function failed, you almost never fired the person first. You fixed the specification first. You realized you had not told them what you actually meant by &#8220;handle this.&#8221;</p><p>AI agents fail for the same reason. Not because they lack capability. Because the founder never built the operating surface the agent needs to do the job.</p><p>The first five posts of The Compounding Founder were written by hand between February and March 2026. They followed no documented system &#8212; the voice existed in my head, the structure was improvised per post, and the distribution happened when I remembered to open Substack The posts were good. They were also unrepeatable. If I got sick for a week, the newsletter stopped. If I changed one thing about the writing process, there was no artifact to update &#8212; just a vague sense that the last post felt different from the one before it.</p><p>In the first week of April, I built the specification layer: a voice system with seven traits and thirty banned words, a content structure template, a distribution config, an asset strategy, and a scorecard. Six artifacts where there had been zero. Then I handed the content function to an agent.</p><p>The specification took about four hours to build. The agent produced a full draft, seven X posts, two LinkedIn posts, seven Substack Notes, and a paid deliverable template in a single run. The draft scored 88% on the voice system &#8212; above the 85% threshold for human review, below the 90% target. The gap was data density: not enough hard numbers. That is a specific, improvable failure. Not &#8220;the agent is not ready.&#8221; The specification was not complete enough.</p><div><hr></div><h3>The Core Idea</h3><blockquote><p><strong>Delegation without specification is not delegation. It is abandonment with a feedback delay.</strong></p></blockquote><div><hr></div><h3>1. The Specification Stack</h3><p>Every function you delegate, to a human or an agent, requires four layers of specification. Miss one and the output degrades predictably.</p><p><strong>Layer 1: Identity.</strong> What is this function? What business does it serve? Who is the audience? This is not a mission statement. It is the context that prevents drift. Without it, the agent optimizes locally, each individual output looks fine, but the collection has no coherence.</p><p><strong>Layer 2: Standards.</strong> What does good look like? Not &#8220;high quality&#8221; &#8212; that means nothing. Specific, auditable criteria. For a newsletter, this means: voice rules, banned vocabulary, structural templates, sentence-level patterns. For customer support, this means: response time targets, escalation rules, tone guidelines, resolution definitions.</p><p><strong>Layer 3: Boundaries.</strong> What can the agent decide? What requires a human? This is where most founders fail. They either restrict everything (making the agent useless) or restrict nothing (making the agent dangerous). The answer is a decision map: these categories are autonomous, these require review, these are human-only.</p><p><strong>Layer 4: Feedback.</strong> How does the agent know if it is working? Not &#8220;the human checks sometimes.&#8221; A structured evaluation loop where output is scored against the standards from Layer 2, and the scores route back into the next cycle.</p><p>Miss Layer 1 and you get drift. Miss Layer 2 and you get mediocrity. Miss Layer 3 and you get either paralysis or chaos. Miss Layer 4 and you get all three, slowly, without noticing.</p><div><hr></div><h3>2. Why &#8220;Just Review the Output&#8221; Fails</h3><p>The most common delegation pattern I see: give the agent the work, review everything it produces, fix what is wrong. This feels responsible. It is actually the worst of both worlds.</p><p>You have not saved time because you are reviewing everything. You have not improved quality because your corrections do not feed back into the system. And you have trained yourself to distrust the agent, which means you review more carefully over time, not less. The cost of the agent stays constant. The cost of your attention increases.</p><p>Review-everything is not delegation. It is supervision cosplaying as leverage.</p><p>The alternative is to build the evaluation into the system. Define the rubric. Score the output before it reaches you. Route only the failures to human review.</p><p>This is how any review-heavy function improves. The first operator run of The Compounding Founder included a manual evaluation step: the agent scanned its own draft for banned words, checked structural patterns, and scored voice trait compliance one by one. That evaluation took the full output of the run. By the time the taxonomist agent within the agent control plane saw the failure report, a programmatic content scorer had been scoped, built, and activated, designed to do in seconds what the manual scan did in paragraphs. The review bottleneck did not disappear because the agent got smarter. It disappeared because the evaluation became an artifact, not a task.</p><div><hr></div><h3>3. The Artifact Test</h3><p>Here is a practical test for whether you have actually delegated a function or just abandoned it.</p><p>Open the folder, document, or system where the agent does its work. Count the artifacts. Not the outputs &#8212; the inputs. The things you created to tell the agent how to operate.</p><p>If the count is zero or one, you have not delegated. You have assigned a task and hoped.</p><p>A properly delegated function should have:</p><ul><li><p>an identity document (what this function is, who it serves)</p></li><li><p>a scorecard (what metrics matter, what the targets are)</p></li><li><p>a standards artifact (what good looks like, specifically)</p></li><li><p>a structure artifact (the default patterns and templates)</p></li><li><p>a boundary map (what the agent decides, what it escalates)</p></li><li><p>a feedback loop (how output is evaluated and how evaluations route back)</p></li></ul><p>Six artifacts minimum. Most founders I talk to have zero. They have a prompt and a prayer.</p><p>The number of artifacts is not bureaucracy. It is the surface area of your specification. Less surface area means more drift, more review, more frustration, and eventually more abandonment disguised as a conclusion that &#8220;agents are not ready.&#8221;</p><div><hr></div><h3>4. Delegation as Architecture</h3><p>The uncomfortable implication is that delegation is not a soft skill. It is architecture.</p><p>When you delegate accounting to a bookkeeper, you give them a chart of accounts, a set of categorization rules, access to the bank feeds, and a monthly review cadence. You do not give them your bank login and say &#8220;handle the money.&#8221; The chart of accounts is an artifact. The categorization rules are an artifact. The review cadence is a feedback loop. You built an operating surface without thinking about it because accounting has had centuries to develop the specification layer.</p><p>AI agent delegation is new enough that the specification layer does not exist yet for most functions. So founders skip it and blame the agent when the output is wrong.</p><p>Building the specification layer feels slow. It feels like overhead. It feels like the opposite of the speed advantage agents are supposed to provide. But the math is simple. Two hours building a voice system saves twenty hours of post-by-post corrections over the next quarter. One hour defining a decision boundary map prevents the three-day crisis when the agent makes the wrong call on something you never told it was sensitive.</p><p>The founders who get the most from agents are not the ones with the best prompts. They are the ones who spent the most time on the specification layer before the agent started working.</p><div><hr></div><h3>5. What This Looks Like in Practice</h3><p>I&#8217;ve been publishing The Compounding Founder for just a couple months, five posts, all written by hand, each one taking a full day from outline to distribution. In the first week of April, I stopped writing and started specifying. I built the voice system, the content structure, the distribution config, the asset strategy, and the scorecard. Six artifacts. Four hours of specification work.</p><p>Then I handed the content function to an agent.</p><p>This post is the output of that handoff. An agent read the specification layer, selected the topic based on gap analysis of the first five posts, wrote the draft, produced the share assets, scored itself against the voice system, and filed a failure report listing ten things it could not do.</p><p>That sentence should create discomfort. It creates discomfort here too. But the discomfort is the point. The question is not whether an agent can write a newsletter post. The question is whether the specification layer is precise enough that the output meets the standard and whether the evaluation loop is honest enough to catch it when it does not.</p><p>The first draft scored 88% on voice compliance. The gap was data density, not enough hard numbers that hit the chest. That is not &#8220;the agent failed.&#8221; That is &#8220;the specification needs one more layer.&#8221; The system improves by improving the specification, not by hoping for a better model.</p><div><hr></div><p><strong>If this changed how you think about delegation, reply with your current setup &#8212; I will tell you which artifacts are missing. If you are not subscribed, this is what we do here every week: one operating principle, one framework, one deliverable that makes the idea real.</strong></p><p><strong>The insight is free. The specification stack I use to delegate an entire business function to an agent, including the six artifacts, the decision boundary template, and the evaluation rubric, is below. </strong></p>
      <p>
          <a href="https://www.thecompoundingfounder.com/p/you-did-not-delegate-to-your-agent">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Your AI Agent Is Building 5-Star Experiences. That's the Problem.]]></title><description><![CDATA[The 11-star framework that turns AI from a shortcut into a standard]]></description><link>https://www.thecompoundingfounder.com/p/your-ai-agent-is-building-5-star</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/your-ai-agent-is-building-5-star</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Sat, 28 Mar 2026 12:10:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pw-q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pw-q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pw-q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pw-q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg" width="1456" height="585" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:585,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:550943,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.thecompoundingfounder.com/i/192399530?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pw-q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pw-q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f59b719-c237-4755-9424-8563ef8e96ad_1632x656.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Mar&#237;a opens Figma at 6 AM in her apartment in Medell&#237;n. She has four hours before her contract shift starts. She is building a wardrobe app, solo, no team, no funding. She types a prompt into her AI coding agent: &#8220;Create a screen where users can browse their closet.&#8221; The agent returns a grid of thumbnails. Rounded corners. A search bar at the top. A filter icon. It works. It is also identical to every other closet screen on every other app that has ever existed.</p><p>Mar&#237;a stares at it. She knows something is wrong but cannot name it. The screen does what she asked. It does not do what her users need.</p><p>She is building a 5-star experience. Functional. Forgettable. Dead on arrival.</p><div><hr></div><h2>The conventional argument</h2><p>The current AI-agent discourse goes like this: agents make you faster. You describe what you want. The agent builds it. Ship it. Move on. Repeat. The thesis is volume, more features, more screens, more iterations per hour than any solo founder could produce manually.</p><p>The problem is that speed toward mediocrity is still mediocrity. It just arrives sooner.</p><p>Every AI coding agent on the market, Cursor, Claude Code, Codex, Copilot, defaults to the same pattern: functional, generic, forgettable. You ask for a dashboard, you get a dashboard. You ask for an onboarding flow, you get an onboarding flow. The shapes are correct. The structure is sound. And no user will ever text a friend about it.</p><p>This is the 5-star trap. It looks like progress. It compiles. It deploys. And it compounds into a product that feels like every other product.</p><div><hr></div><h2>Where the framework comes from</h2><p>In 2015, Brian Chesky sat down with Reid Hoffman for what would become one of the most cited episodes of the <em>Masters of Scale</em> podcast. Hoffman asked Chesky how Airbnb thinks about experience design. Chesky described an exercise he runs with his team.</p><p>Start at one star. The experience is broken, the host does not show up, the door is locked, nobody answers the phone. Move to three stars. You get in. The place is fine. Nothing memorable. Five stars. The place is clean, the bed is comfortable, there is a welcome note. This is where most companies stop.</p><p>Then Chesky pushed further. What is a six-star experience? The host leaves a handwritten note with restaurant recommendations tailored to your taste. Seven stars? A welcome basket with your favorite snacks, how did they know? Eight stars? Elon Musk greets you at the airport. Nine? A parade. Ten? You arrive and the Beatles are there to play a concert for you.</p><p>Eleven stars is absurd. It is deliberately impossible. And that is the point.</p><p>The exercise is not about building the 11-star version. It is about <em>thinking from the 11 and working backwards to what is actually shippable</em>. Because when you start at 5 and try to push to 6, you add features. When you start at 11 and pull back to 8, you rethink the entire experience.</p><p>The gap between those two approaches is the gap between a product people use and a product people remember.</p><div><hr></div><h2>The emotional job</h2><p>Every interface has two jobs. The functional job is obvious, browse clothes, schedule a meeting, track expenses. The emotional job is invisible and more important.</p><p>The functional job of a wardrobe app is: show me my clothes. The emotional job is: make me feel like I know what I am doing when I get dressed.</p><p>The functional job of a dashboard is: display metrics. The emotional job is: make me feel like I understand my business right now, in this moment, without digging.</p><p>AI agents only see the functional job. They cannot see the emotional one. They do not know that the user arriving at your closet screen feels overwhelmed, not curious. They do not know that the person opening your dashboard at 7 AM is anxious, not analytical.</p><p>This is why default AI output lands at 5 stars. The machine solves the functional job perfectly and ignores the emotional job entirely.</p><div><hr></div><h2>What 11-star thinking does to your agent workflow</h2><p>Here is what changes when you make this framework the foundation of every interaction with your AI agents.</p><h3>You stop prompting for features. You start prompting for feelings.</h3><p>Instead of: &#8220;Build a screen where users browse their closet.&#8221;</p><p>You write: &#8220;This interface transforms &#8216;I have no idea what to wear&#8217; into &#8216;I look incredible and I barely tried&#8217; by making outfit selection feel like a stylist handed you three perfect options. The user should feel relief within 3 seconds of landing on this screen.&#8221;</p><p>The output from that prompt is categorically different. Not because the agent suddenly has taste, it does not. But because you gave it the emotional specification that determines every layout decision, every copy choice, every animation.</p><h3>You map the trajectory before you write a single prompt.</h3><p>Before touching the agent, write out the star levels for the specific experience you are building:</p><ul><li><p><strong>1 star</strong>: The screen loads. The user sees a blank grid. No guidance. They close the app.</p></li><li><p><strong>3 stars</strong>: Clothes appear in a grid. No organization. The user scrolls endlessly. They find something by accident.</p></li><li><p><strong>5 stars</strong>: Clothes are categorized. There is a search bar. Filters work. The user finds what they want in 30 seconds. Forgettable.</p></li><li><p><strong>7 stars</strong>: The app suggests three outfits based on the weather and the user&#8217;s calendar. The user smiles. Small surprise.</p></li><li><p><strong>9 stars</strong>: The app knows the user has a job interview tomorrow. It surfaces the outfit that got compliments last time it was worn. It accounts for the weather, the dress code, and what is clean.</p></li><li><p><strong>11 stars</strong>: The user opens the app and a stylist has already laid out tomorrow&#8217;s outfit on their bed. Not a screen. The physical clothes. On the actual bed.</p></li></ul><p>Obviously you cannot ship 11. But you can ship 9, an experience that <em>anticipates</em> rather than <em>reacts</em>. And you would never have designed that 9-star version by starting at 5 and adding features.</p><h3>You give your agent a design conviction, not a style preference.</h3><p>Most people prompt their agents with aesthetic requests: &#8220;Make it clean and modern.&#8221; This produces the visual equivalent of elevator music. Pleasant. Unmemorable. Interchangeable.</p><p>A design conviction is a single sentence that forces every decision:</p><p>&#8220;Dense information, zero noise &#8212; Bloomberg terminal meets Notion.&#8221;</p><p>&#8220;A hand-written letter from your smartest friend.&#8221;</p><p>&#8220;The UI equivalent of a perfectly tailored black suit.&#8221;</p><p>When you give an agent a conviction instead of a style, the output has a point of view. The typography choice follows from the conviction. The spacing follows. The color palette follows. The micro-copy follows. Everything coheres because everything serves the same sentence.</p><div><hr></div><h2>The 5-star tells</h2><p>Here is how you know your AI agent is building at 5 stars. Every one of these is a default pattern that agents reach for when you do not intervene:</p><ul><li><p><strong>Generic hero sections</strong> with &#8220;Welcome to [Product]&#8221; and a gradient background</p></li><li><p><strong>Centered spinners</strong> with no context. The user stares at a circle and wonders if the app is broken.</p></li><li><p><strong>&#8220;Something went wrong&#8221;</strong> as an error message. No specificity. No fix. No warmth.</p></li><li><p><strong>Gray placeholder rectangles</strong> as empty states. The user sees nothing and learns nothing.</p></li><li><p><strong>Purple-to-blue gradients</strong>. The unofficial color scheme of AI-generated interfaces. You have seen it a thousand times. So has everyone else.</p></li><li><p><strong>Uniform card grids</strong> with no visual hierarchy. Everything is the same size, the same weight, the same importance. Nothing leads. Nothing recedes.</p></li></ul><p>These are not bugs. They are the natural output of an agent that was asked for a functional solution and delivered one. The problem is not the agent. The problem is the spec.</p><div><hr></div><h2>Making it operational</h2><p>This is not a philosophy you apply once. It is a filter you run on everything.</p><p>Every time you open a chat with your coding agent, ask three questions before you type:</p><ol><li><p>What does the user feel right now, before they touch this?</p></li><li><p>What should they feel after?</p></li><li><p>What star level am I about to accept?</p></li></ol><p>If you cannot answer those, you are not ready to prompt. You will accept whatever the agent gives you, and it will give you 5 stars.</p><p>Here is the operating rule I use: <strong>never ship the first output.</strong> The first output is always the 5-star version. It is the default. Treat it as a sketch, not a deliverable. Push the agent to 7 by naming the emotional gap. Push to 8 by adding the design conviction. Get close to 9 by specifying the sad paths, what happens when the screen is empty, when the data fails to load, when the user has 1,000 items instead of 10.</p><p>The agents are fast enough that three rounds of refinement still takes less time than one round of manual coding. Use that speed to raise the floor, not to ship the default.</p><div><hr></div><h2>Why this compounds</h2><p>A 5-star experience retains users at the baseline rate. They use it when they need it. They forget about it when they do not.</p><p>An 8-star experience creates a moment the user did not expect. They text someone. They leave a review that says something specific instead of &#8220;works great.&#8221; They open the app when they do not strictly need to, because it made them feel something.</p><p>That difference, between functional and felt, is the difference between a product that grows linearly through acquisition spend and a product that compounds through word of mouth. The 11-star framework is not about perfection. It is about building the habit of imagining the impossible version first, then working backwards to the version that is shippable and still makes someone pause.</p><p>Every week you ship 5-star output, you are training yourself, and your agents, to accept the default. Every week you push to 8, you are compounding craft. And craft, unlike features, does not depreciate.</p><p>Brian Chesky did not invent a design methodology. He invented a question: <em>What would the impossible version look like?</em> The answer to that question, trimmed back to reality, is always better than the answer to &#8220;What is the functional version?&#8221;</p><p>Mar&#237;a in Medell&#237;n is still staring at her closet screen. The grid loads. The thumbnails are fine. She deletes the prompt and types a new one. This time she starts with how the user should feel.</p><p>The grid looks different now.</p><div><hr></div><p>The principle is free. The skill file that makes your AI agent build at 8 stars by default, the full prompt engineering framework, the star-mapping template, and the design system it enforces, is below.</p><p>For paid subscribers: the full 11-Star UX/UI Experience Builder skill, the implementation checklist, and the design system architecture below.</p>
      <p>
          <a href="https://www.thecompoundingfounder.com/p/your-ai-agent-is-building-5-star">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Your harness is too small]]></title><description><![CDATA[Designing an integrated business environment]]></description><link>https://www.thecompoundingfounder.com/p/your-harness-is-too-small</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/your-harness-is-too-small</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Fri, 13 Mar 2026 17:21:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/youtube/w_728,c_limit/Kfbc_PmzK34" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There is a scene in Succession where Matteson asks Roman what he is worst at. Roman deflects &#8212; &#8220;who, me?&#8221; &#8212; and Matteson answers the question himself. &#8220;Success doesn&#8217;t really interest me anymore,&#8221; he says. &#8220;Analysis plus capital plus execution &#8212; anyone can do that.&#8221; Then he leans in: &#8220;Failure &#8212; that&#8217;s a secret. Just as much failure as possible, as fast as possible. Burn that shit out. That&#8217;s interesting.&#8221; He says it like a man who has already solved the easy problem and moved on.</p><div id="youtube2-Kfbc_PmzK34" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;Kfbc_PmzK34&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/Kfbc_PmzK34?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>LLMs just made success cheaper. Given contextualized data, a model can research a problem, analyze it, execute on it, and iterate until the outcome occurs. The formula Matteson found trivial is now available to anyone who can state a goal clearly. Which means Matteson was right about what becomes interesting next. The failure mode. The thing that still burns. The constraint that did not disappear when the inputs got cheaper. That constraint is context rot.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.thecompoundingfounder.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">The Compounding Founder is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Context rot is not a metaphor. Humans set ambitious goals. They know the best practices. And then time passes, attention moves, the context decays. Not because they stopped caring. Because context does not persist reliably in humans. We forget. We drift. We apply yesterday&#8217;s understanding to today&#8217;s problem and call it experience.</p><p>Here is what I find both ironic and almost serendipitous: we built our artificial intelligence with a version of the same flaw. Not identical. Human context rot is organizational, cultural, the slow drift of priorities over months. AI context decay is technical, session-scoped, attention fading across a long prompt. Different failure modes. But the same symptom: the model loses the thread. We replicated our own cognitive limitation in the tools we made to augment us. The interesting part is that the fix for the AI problem (injecting context deliberately at the start of every session) also forces you to confront the human problem. You cannot write program.md without deciding what you are actually trying to do. The act of building the injection mechanism disciplines the goal.</p><p>But unlike humans, you can deterministically re-inject context into an AI. Every session. Mechanically. Karpathy&#8217;s AutoResearch project, released this month, makes the mechanism concrete. There is a file called program.md. It contains the goal, the constraints, the research direction. The agent reads it at the start of every run. You do not touch the Python. You program the program.md. The agent re-grounds itself to your intent on every single loop, not because it remembers, but because the context is injected fresh each time. That is the fix. Not memory. Injection. The Socratic question: are we still moving toward what we said we were trying to do? It gets asked mechanically, before anything executes. Humans cannot do this reliably. A well-designed system can. That is what the harness needs to be built around.</p><div><hr></div><h2>The funnel is not the only option</h2><p>What changed is not that context became more available. Organizations have always produced enormous amounts of context &#8211; metrics, feedback, goals, decisions, postmortems. What changed is the cost of routing it to the moment it matters.</p><p>That cost is now near zero.</p><p>An agent can pull last week&#8217;s conversion data and attach it to a feature spec before a line of code is written. It can surface the three support tickets that describe this exact problem in user language, not product language. It can check whether the goal that originally motivated this work still exists &#8211; or got quietly deprioritized while everyone was looking at their tickets.</p><p>This is not a vision. This is plumbing. Every piece of that context already exists somewhere in the stack. The question is whether your harness routes it, or lets it stay siloed.</p><p>Most harnesses let it stay siloed.</p><p><strong>The harness that only covers the development lifecycle is not a small problem. It is a fundamental mismatch between what is now possible and what most builders have built.</strong></p><div><hr></div><h2>How organizations will run</h2><p>The organizations that figure this out first will not look like they have better tools. They will look like they think differently &#8211; like they make fewer decisions based on stale context, like they ship things that actually move the numbers they were trying to move, like they notice problems before the metrics force them to.</p><p>What they will actually have is a harness that extends beyond code.</p><p>I use the word <em>builder</em> deliberately. Not developer. Because the person making product decisions in a two-person startup, or a solo founder shipping an app between Slack messages, or a non-technical operator wiring together AI agents &#8211; that person is building. Their harness should not assume they write code. Their harness should assume they have goals.</p><p>That distinction changes everything about what the harness needs to know.</p><p>A developer needs autocomplete and debugging.</p><p>A builder needs that &#8211; plus: <strong>Is this the right thing to build? What happens to the number if we do not ship it? Who asked for it, in what words, and how recently?</strong></p><div><hr></div><h2>The sidecar that watches everything</h2><p>I am building this for myself with Clueless Clothing as a guinea pig. Not from theory. From watching how much context disappears between the moment I understand what the business needs and the moment I sit down to build something.</p><p>The architecture I landed on is not just a wider harness. It is a harness with a governing layer above it &#8211; an agent that observes what is actually happening, analyzes it against what I said I was trying to do, and surfaces improvements for me to review and act on.</p><p>The observe-analyze-action loop is not new. Every retrospective, every postmortem, every weekly metrics review is some version of it. What is new is that it does not have to be periodic anymore. It does not have to wait for the meeting.</p><p>But here is the part that took me a while to see clearly.</p><p><strong>The sidecar does not improve processes. It improves the processes that move your specific goal.</strong></p><p>That sounds like a minor distinction. It is not. If your goal is to grow to $50k MRR, the sidecar optimizes toward revenue-moving leverage points &#8211; shipping frequency, conversion bottlenecks, activation rate. If your goal is to build the most efficient engineering team, the sidecar optimizes toward a completely different set of processes. The same codebase, the same activity, the same signals &#8211; and a different set of improvements surfaces.</p><p>The goal is the lens. Without it, you are just generating observations. But this is also where the thesis gets uncomfortable. The sidecar only fights context rot on the goal you gave it. If the goal is wrong, if you are faithfully optimizing toward $50k MRR when the actual opportunity is a different product entirely, the system will not tell you. It will optimize diligently toward the wrong destination. AutoResearch is instructive here, but not in the way I originally thought. The agents own train.py entirely. They modify it, iterate it, discard what does not work. But program.md, the file that contains the research direction, the goal, the constraints, stays human. Deliberately. The design enforces it. Which means the question of whether you are optimizing toward the right thing never gets delegated. The next version of the sidecar is not one that updates your goal automatically. It is one that makes the cost of ignoring a wrong goal high enough that you actually revisit it.</p><div><hr></div><h2>The IDE moment, one level up</h2><p>This has happened before in a smaller domain.</p><p>Before the IDE existed, developers used separate tools for editing, compiling, and debugging. Each tool did its job. The cost was not in any single tool &#8211; it was in the constant switching, the context that leaked at every seam, the mental overhead of holding state across systems that did not talk to each other.</p><p>Then someone merged them. The productivity gain was not from better features. It was from eliminating the seams.</p><p>We are at that moment again, one level up.</p><p>The separate systems now are not editor, compiler, and debugger. They are product management, development, analytics, marketing, support, and documentation. Each one has its own tool, its own data model, its own workflow. The seams between them leak context every day &#8211; the same context that used to require a meeting to route, that now could be routed automatically.</p><p>This is what an Integrated Business Environment actually is. Not a better IDE. Not an IDE with more integrations bolted on. A different assumption about what the harness is for &#8211; not just executing work, but <strong>continuously orienting work toward the goal, based on what is actually happening.</strong></p><p>The IDE helped developers write better code.</p><p>The IBE helps builders make better decisions about what to build next.</p><div><hr></div><h2>The architecture</h2><p>I am building this in layers. Not because I love architecture diagrams, but because the tooling landscape is moving too fast to couple everything together. The best AI coding tool right now will not be the best in four months. If your business rules are tangled with your adapter logic, every tool switch means starting over.</p><p><strong>Five layers:</strong></p><p><strong>Core</strong>: the rules that should still make sense in a different repo, a different agent, a different stack entirely. Think of it as the constitution. It does not care what language you write in.</p><p><strong>Adapter</strong>: environment-specific translations. This is the disposable layer. When the tool changes, rewrite the adapter. Leave everything else alone.</p><p><strong>Stack</strong>: technology-specific defaults. React Native looks different from Rails. This layer knows that.</p><p><strong>Overlay</strong>: the layer most harnesses do not have. Product rules. Business context. Metrics. Goals. User segments. The actual reason any of this work is happening.</p><p><strong>Split</strong>: not a layer but a discipline. Any file that mixes concerns gets split over time. This is what keeps the architecture honest as it evolves.</p><p>The overlay layer is where the IBE lives. It is also where I spend most of my time now. That is the right place to spend it.</p><div><hr></div><p>I am open-sourcing this structure. Not because it is finished, but because the pattern is clear enough to share before it is polished.</p><p>One thing worth saying plainly before you go build it: this is not a software idea.</p><p>Look at what AutoResearch actually does to understand why. The primitive Karpathy&#8217;s agents operate on is not software. It is experiments. Train for five minutes. Check if val_bpb improved. Keep or discard. Repeat. The code is almost incidental. The agent rewrites it, yes, but that is not the work. The work is the experiment loop oriented toward a measurable goal. Software is just what happens to be the medium. A restaurant optimizing table turns has a different medium. A consulting firm improving proposal win rates has a different medium. The loop is the same. The medium changes. The harness needs to know the difference.</p><p>The IDE framing is a useful on-ramp. It is not the destination. The destination is any operation, any goal, any set of processes, with a governing layer that fights context rot and keeps everything oriented toward what you said you were trying to do.</p><p>There is one more place where the harness is too small. Not the tooling. Not the lifecycle. The goal itself.</p><p>I have been running $50k MRR as a target because it feels concrete. Measurable. Safe to write in a file. But $50k MRR is not a goal. It is a metric that might serve a goal, if I chose the right one. The actual goal, the one worth writing in program.md, is something like: enough margin to be present. To travel when it matters. To not make decisions from fear at 11pm. Or read more. Or be more present with the people who matter. The harness does not require the goal to be economic. It requires the goal to be real. But if you give it the metric instead of the goal, you will hit the number and wonder why the thing you were actually trying to do did not happen. The sidecar optimizes toward what you said. Not toward what you meant. That gap is yours to close, before you write anything down.</p><p>Matteson was right. Success got cheap. The failure mode is what stayed interesting. Context rot is the failure mode. The only question is whether you build a harness that fights it, or one that lets it win quietly while you keep shipping.</p><p><em>Next: The Athenians did not need productivity software. They had something better: a society that treated economic output as infrastructure, not identity. The citizen class was freed from subsistence not so they could optimize, but so they could think, argue, govern, and live. We are building the infrastructure that could do the same thing. We just have not figured out what we are supposed to do once it works. The Greeks knew. Next piece is about what they got right, and what it means for how we should be spending the hours the agents are buying us.</em></p><p>What is actually in your program.md right now, and is it a goal, or a metric?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.thecompoundingfounder.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">The Compounding Founder is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Setup That Changes How You Work With Claude]]></title><description><![CDATA[Every guide starts with the prompt. The one that worked for me started with the goal.]]></description><link>https://www.thecompoundingfounder.com/p/the-setup-that-changes-how-you-work</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/the-setup-that-changes-how-you-work</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Sat, 07 Mar 2026 18:47:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ay6o!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F365e2f47-28e8-428d-9c67-44413c616ce1_1287x1287.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I did not set out to automate my social media content.</p><p>I opened a Claude cowork session in January with one goal written down: grow paid subscribers for the Clueless Clothing app. That was it. No prompt about social media. No instruction to connect any accounts. I just told Claude what I was trying to build and started working backwards from there.</p><p>What happened over the next two sessions is what I want to show you.</p><p>Claude started asking me questions. Not the questions I expected &#8212; not &#8220;what do you want to post?&#8221; &#8212; but questions about where my subscribers were coming from, what signals I had, what data it would need to actually help me move that number. It asked me to provide it access to sources I had not thought to connect. It mapped a growth loop I had not drawn yet. By the end of session two, it had proposed automating my social distribution as a subscriber acquisition channel &#8212; and built the workflow to do it.</p><p>I did not tell it to do that. It got there because I started with the goal.</p><p>That is goal-first setup. And it is the one thing no Claude guide I found told me to do first.</p><div><hr></div><h2>The Conventional Argument</h2><p>Every guide I found ran the same sequence. Pick a model. Write better prompts. Use Projects for memory. Maybe add a custom system prompt. The advice is not wrong. It is just aimed at the wrong question.</p><p>Those guides assume you want Claude to write faster. What I wanted was for Claude to think with me &#8212; to hold context across sessions, to remember what I had already decided, to get closer over time instead of resetting.</p><p>That is a different tool. Same software. Different setup. One step separates them, and I did not find it in any guide.</p><div><hr></div><h2>The Dismantle</h2><p>Here is what was happening before I rebuilt my setup: I was optimizing the prompt before I had clarity on the problem.</p><p>Each session, I opened a blank thread and started from scratch. Claude had no idea what we had worked through the session before. No idea what the constraint was. No idea what a bad recommendation looked like in my context. I was treating a thinking partner like a search engine with better grammar.</p><p>The result was clean output that did not accumulate. Every session produced something. Nothing compounded. I was doing the connective work manually &#8212; in my head, in a doc somewhere, in the friction between sessions &#8212; and calling it a workflow.</p><p>I stopped when I realized I was spending more time re-orienting Claude than I was spending on the actual work.</p><div><hr></div><h2>The Core Idea</h2><p><strong>Start with the goal, not the prompt. That one inversion turns Claude from a session tool into a compounding asset.</strong></p><div><hr></div><h2>What I Changed</h2><h3>1. I wrote the goal before I opened Claude</h3><p>Before I rebuilt my setup, I wrote one paragraph offline &#8212; in Notes, not in Claude &#8212; that answered three things: what am I trying to accomplish in the next 60 to 90 days, what does done actually look like, and what do I know that Claude does not.</p><p>For the subscriber growth goal: I needed paid subscribers for the Clueless Clothing app. Done meant a number &#8212; a specific acquisition target with a date attached. What Claude did not know: my list size, my current conversion rate, where my traffic was coming from, and the fact that social distribution was an untapped channel I had not systematized.</p><p>That paragraph took eight minutes to write. It changed what Claude asked me in the first session. It stopped generating output and started asking what it needed to actually help.</p><h3>2. I built Projects around outcomes, not topics</h3><p>My original Projects were organized like a filing cabinet. Writing. Strategy. Research. Useful for finding things. Useless for compounding work.</p><p>I rebuilt them around outcomes. <em>App subscriber growth. Product roadmap for Q2. Distribution systems.</em> Every conversation inside an outcome-scoped Project is already aimed at something that ends. Claude does not have to guess what matters. The sessions inherit direction automatically.</p><p>The subscriber growth Project is where Claude eventually proposed the social automation &#8212; because every session inside it was pointed at the same number. The proposal did not come from a prompt. It came from context accumulating until Claude had enough to make a connection I had not made myself.</p><h3>3. I wrote instructions that named constraints, not just context</h3><p>The old version of my Project instruction: <em>I&#8217;m the founder of Clueless Clothing, a mobile app. Help me grow.</em></p><p>The new version: <em>I&#8217;m the solo founder of Clueless Clothing, a mobile app. My goal is growing paid subscribers. My constraint is time &#8212; I have no team, so any system I build has to run without me. If a recommendation requires manual execution more than once a week, flag it. I want compounding distribution, not one-off campaigns.</em></p><p>The second version changed what Claude recommended. It stopped suggesting tactics that required daily attention. It started building systems. The social automation came out of that constraint &#8212; Claude understood that a workflow I had to run by hand every day was not a solution.</p><h3>4. I loaded the artifact, not the summary</h3><p>Before the first serious session on a Project, I paste in the actual document &#8212; the analytics export, the campaign brief, the current conversion data. Not a summary. The source.</p><p>Summaries introduce my interpretation. My interpretation introduces drift. Three sessions in, I am working from Claude&#8217;s model of my model of the situation. Loading the source document removes one layer of inference. The output quality difference is not subtle.</p><h3>5. I close the loop once a week</h3><p>Friday afternoon, five minutes in the active Project. One line: what changed this week. One line: whether the goal shifted. One line: what Claude got wrong that I corrected.</p><p>This is not journaling. It is calibration. Claude does not update its model of the project automatically. I have to push the new information in. Five minutes a week prevents months of accumulated drift. Without it, the Project becomes a transcript of what you used to be working on.</p><div><hr></div><h2>What Happened After</h2><p>By the second week of running goal-first setup, the sessions were noticeably shorter. Not because Claude had gotten better. Because the Project already held the context. I was not re-anchoring. I was working.</p><p>The social automation workflow Claude proposed is now running. It pulls content signals, formats posts, and queues distribution &#8212; without me prompting it to do any of that on a given day. I did not ask for a social media automation tool. I asked Claude to help me grow subscribers. It got there on its own because the goal was loaded, the constraints were named, and the context had been accumulating for two weeks.</p><p>That is the thing no guide I found described. The Project does not just remember. It reasons forward. The work you do in session one changes what session eight looks like. That is different from a good prompt. That is a different tool.</p><div><hr></div><p><em>The principle is free. The system I use to enforce it &#8212; the goal-framing worksheet, the Project instruction template, the Clueless Clothing worked example, and the </em><code>/retro</code><em> skill that keeps the whole thing from decaying &#8212; is below.</em></p>
      <p>
          <a href="https://www.thecompoundingfounder.com/p/the-setup-that-changes-how-you-work">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Taste Bottleneck]]></title><description><![CDATA[AI can generate faster than you can judge. That is the actual problem.]]></description><link>https://www.thecompoundingfounder.com/p/the-taste-bottleneck</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/the-taste-bottleneck</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Mon, 02 Mar 2026 04:35:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ay6o!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F365e2f47-28e8-428d-9c67-44413c616ce1_1287x1287.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last Tuesday at 2am, I opened my app and scrolled through the screens I had shipped that month. Thirty-six of them. AI wrote most of the code. Every screen worked. Nothing crashed. The buttons went where buttons go.</p><p>I felt nothing.</p><p>Not bad. Not good. Just &#8212; flat. The spacing was fine. The typography was fine. The colors were fine. Fine is the word you use when you cannot explain what is wrong but you know something is.</p><p>I closed the laptop and went to bed. In the morning I opened it again. The screens were still fine.</p><div><hr></div><p>The conventional argument about AI and quality goes like this: AI is a tool, and like any tool, the output reflects the operator. If the design is mediocre, that is a skill gap. Learn design. Get better at prompting. Ship with more intention.</p><p>The problem is that this argument assumes the bottleneck is generation. It is not.</p><p>I can generate a screen in four minutes. A good prompt, a decent component library, and Claude will produce something functional before my coffee cools. That part works. That part has never been easier.</p><p>The bottleneck is judgment. Specifically, my judgment, applied thirty-six times, at a quality that does not degrade when I am tired, distracted, or moving fast.</p><p>I am one person. The app does not care.</p><div><hr></div><p>Here is what I mean by judgment.</p><p>When I look at a screen and something feels off, I am actually running five evaluations simultaneously. I had never separated them before. It took sitting down and forcing myself to articulate what my eyes were doing.</p><p>Spatial rigor. Is every margin, every padding, every gap a multiple of four? The 4px grid is not a preference. It is what makes things feel aligned without the user knowing why. I ran a grep across all thirty-six screens. Eleven had spacing values like 15, 18, or 22 &#8212; values that are not on the grid. Eleven of thirty-six. Almost a third were subtly misaligned, and I had not caught it because each screen, in isolation, looked close enough.</p><p>Typographic hierarchy. One screen, one clear entry point. You squint at the screen and you should immediately know where to look first. Fifteen of the twenty-six style learning screens had no display-weight typography at all. Every heading was the same weight as the body text. The visual hierarchy was not wrong &#8212; it was absent.</p><p>Restraint. For every border, every badge, every icon &#8212; does removing it break the task? I started counting visual elements. Most screens had three to five things that existed for decoration, not function. Removing them did not hurt comprehension. It helped.</p><p>Motion. This one was the worst. Nineteen of twenty-six screens had no entrance animation at all. Static content appearing from nothing, like a light switch instead of a door opening. The screens that did have motion used hardcoded durations and cubic curves instead of spring physics. Nothing felt alive.</p><p>Information ergonomics. Can I reach the primary action with my thumb? Are there fewer than seven interactive elements in the initial viewport? Basic stuff. The kind of thing you check once and then forget to check on screen number twenty-three.</p><p>Five evaluations. Each measurable. Each something I could check by reading the code, not by looking at a screenshot. Each something I had been doing unconsciously &#8212; and therefore inconsistently.</p><div><hr></div><p>So I wrote the rubric down.</p><p>Not as guidelines. Not as principles. As a scoring system. Each dimension, one to five. Weighted: spatial and typographic at 25% each, restraint at 20%, motion and ergonomics at 15% each. Calibrated against my own gut.</p><p>The calibration was the hard part.</p><p>Left alone, AI will score its own work 4.5 out of 5 every time. &#8220;The spacing is clean and the typography is well-structured.&#8221; It says this the way a student who did not read the book still tries to sound confident in the essay. Technically defensible. Actually meaningless.</p><p>I defined what a 3 looks like: it works, nothing is broken, it feels generic. That is the expected starting point for AI-generated UI. Not a failure. The baseline. Most of what ships in this industry is a 3.</p><p>I defined what a 5 looks like: this could be a screenshot in Apple&#8217;s Human Interface Guidelines. I have not seen my code produce a 5 yet. That is fine. It means the scale is honest.</p><p>And I added one rule: when in doubt between two scores, pick the lower one. It is easier to discover a screen is better than expected than to ship something that needed more work.</p><div><hr></div><p>Then I pointed the system at my own app.</p><p>The wardrobe feature &#8212; ten screens, the part of the app I had spent the most time on &#8212; averaged 3.4 out of 5. Two screens were ready to ship. Seven needed polish. One was at the threshold.</p><p>The style learning feature &#8212; twenty-six screens, the part I thought was in good shape &#8212; also averaged 3.4. But the distribution was worse. Three screens shipping. Seventeen needing polish. Six needing complete reworks. Six.</p><p>The weakest dimension across both features: motion. Average score of 2.0 to 2.5. Almost every screen was missing entrance animations, haptic feedback, and stagger patterns. The code was architecturally sound and visually dead.</p><p>I sat with that for a while.</p><div><hr></div><p>The conventional response here would be to frame this as a success story. I found the problems. I built the system. Now I can fix them efficiently. Quality will compound.</p><p>That is partly true. The system does compound. Run the audit after fixing six screens, and the new ones score higher because the patterns are in place. Run it again a month later, and regressions get caught immediately. The bar does not lower. It ratchets.</p><p>But I want to be honest about what this actually reveals.</p><p>I shipped thirty-six screens. I was proud of them. An automated system I built in a weekend told me that six of them need to be rewritten, seventeen need meaningful polish, and the best ones &#8212; the ones I had spent the most time caring about &#8212; are still only 3.8 out of 5. Not bad. Not great. Close enough to good that I had convinced myself they were good.</p><p>The system did not tell me anything I did not already know. It told me things I had been choosing not to see.</p><div><hr></div><p>The pattern underneath this is not about design.</p><p>Any judgment you make repeatedly can be decomposed into measurable dimensions. You can score those dimensions. You can calibrate the scores against your own taste. You can teach a system to apply them. And then the system will tell you the truth even when you would rather not hear it.</p><p>The uncomfortable question is whether the rubric captures taste or merely approximates it. Whether a system that scores 3.4 out of 5 is telling me the truth or telling me a useful lie.</p><p>I think the answer is both. The rubric does not replace judgment. It makes judgment auditable. It turns &#8220;this feels off&#8221; into &#8220;the spatial score on screen fourteen is 2.5 because of three off-grid padding values on lines 47, 82, and 114.&#8221; That is not taste. That is measurement.</p><p>But measurement is what scales. Taste does not.</p><div><hr></div><p>Design was the first domain. API quality is next. Content quality after that. Each one follows the same five-step process, and each one starts with the same hard question: what am I actually evaluating when I say &#8220;this is good&#8221;?</p><p>The process works because it is a forcing function, not a scoring matrix. It forces you to articulate what you have been doing unconsciously. The articulation changes how you see.</p><h2>The Philosophy Behind the Skills</h2><p>Apple does not publish a rubric. But if you study their apps long enough &#8212; not the guidelines document, the actual apps &#8212; you notice they encode five things into every screen.</p>
      <p>
          <a href="https://www.thecompoundingfounder.com/p/the-taste-bottleneck">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Options Are Cheap. Conviction Is Rare.]]></title><description><![CDATA[Marco has twelve browser tabs open.]]></description><link>https://www.thecompoundingfounder.com/p/options-are-cheap-conviction-is-rare</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/options-are-cheap-conviction-is-rare</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Thu, 26 Feb 2026 20:41:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ay6o!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F365e2f47-28e8-428d-9c67-44413c616ce1_1287x1287.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Marco has twelve browser tabs open.</p><p>He has been building his app for fourteen months. Shipped it in October. Four hundred users. He has been thinking about the paywall for three weeks, which means he has been not thinking about the paywall, which means he has been producing tabs.</p><p>Soft paywall. Hard paywall. Feature gating. Usage gating. Seven-day trial. Fourteen-day trial. Freemium with limits. Freemium without. The $9.99 tier. The $14.99 tier. The &#8220;most popular&#8221; badge on the middle option.</p><p>He has a spreadsheet. It has conversion rate benchmarks. It has twelve rows.</p><p>He is not stuck because he lacks information.</p><p>He is stuck because he has too much of it.</p><p>This is the new problem. Not the blank page. The full one.</p><p><strong>AI can generate options. It cannot generate conviction.</strong></p><div><hr></div><h2>Building vs Crafting</h2><p><strong>Building is</strong> making something that works.</p><p><strong>Crafting is</strong> making something that feels inevitable to the right person.</p><p>Building outputs features.</p><p>Crafting outputs belief.</p><p>And belief is what turns:</p><ul><li><p>a product into a habit</p></li><li><p>a landing page into a conversion</p></li><li><p>a tool into &#8220;this feels like me&#8221;</p></li></ul><p>When building was expensive, scarcity forced decisions. You could not afford twelve options. You picked one and lived with it.</p><p>Now the cost of &#8220;plausible&#8221; has collapsed.</p><p>And the discipline of choosing has not kept up.</p><h2>Taste (Defined Without the Vibes)</h2><p>Taste has a reputation problem. It gets treated like a personality trait &#8212; something you either have or you do not.</p><p>But taste is not a mood.</p><p>It is not an aesthetic.</p><p>It is not &#8220;I like this.&#8221;</p><p>Taste is a set of skills that let you navigate abundance without drowning in it.</p><p>Here is a definition that holds up under pressure:</p><p><strong>Taste is the ability to consistently choose what matters.</strong></p><p>It shows up as:</p><ul><li><p><strong>Signal detection:</strong> noticing what is real and what is noise</p></li><li><p><strong>Coherence:</strong> decisions that reinforce a point of view</p></li><li><p><strong>Editing:</strong> removing &#8220;almost right&#8221; things that dilute the whole</p></li><li><p><strong>Timing:</strong> knowing when &#8220;good enough&#8221; is actually wrong</p></li><li><p><strong>Empathy:</strong> understanding what the user is trying to become</p></li></ul><p>A punchier frame:</p><p><strong>Taste is compression. Direction is selection. Empathy is calibration.</strong></p><h3>Taste is compression</h3><p>You take a messy reality and compress it into something simple without being simplistic.</p><p>You are not adding clarity. You are removing confusion.</p><h3>Direction is selection</h3><p>Direction is the ability to say: &#8220;We are doing <em>this</em>, not that.&#8221;</p><p>Not because you are stubborn.</p><p>Because you have a point of view, and you are willing to be misunderstood by the wrong audience to be unforgettable to the right one.</p><h3>Empathy is calibration</h3><p>Empathy is not &#8220;being nice.&#8221;</p><p>It is the skill of mapping what people desire, what they fear, and what they will actually do on a Tuesday morning when they are tired and their alarm just went off.</p><p>Empathy lets you calibrate craft to the human on the other side.</p><h2>Why This Matters Now</h2><p>AI does not just speed you up.</p><p>It multiplies the number of plausible paths you could take.</p><p>That abundance sounds like freedom.</p><p>But it creates a quiet failure mode:</p><p>You keep moving.</p><p>You keep generating.</p><p>You keep shipping.</p><p>And you stop committing.</p><p>Options become a substitute for decisions.</p><p>That is why crafting gets harder.</p><p>Not because we lost tools.</p><p>Because we gained too many.</p><h2>Where This Gets Painful: Paywalls</h2><p>Paywalls are where &#8220;options vs conviction&#8221; stops being abstract.</p><p>The conventional wisdom says: run tests, track conversion, optimize the flow. The data will tell you what works.</p><p>This is not wrong. It is incomplete.</p><p>Because monetization is not just a pricing tactic.</p><p>It is product philosophy.</p><p>It teaches users what the product is.</p><p>It shapes how safe they feel exploring it.</p><p>It decides whether the relationship starts with trust or pressure.</p><p>And in the long run, <strong>trust compounds harder than conversion rate.</strong></p><p>I have run this calculation myself. I was building Clueless Clothing, an AI wardrobe app &#8212; a product where the user shows up every morning already tired, already making dozens of small decisions before they even open the app. I had the same spreadsheet Marco has. Every paywall option was technically defensible. None of them felt obviously right. The question that eventually cut through the tabs was not &#8220;which one converts best?&#8221;</p><p>It was: <em>What kind of relationship am I building?</em></p><p>A trust-first answer tends to sound like:</p><ul><li><p>&#8220;We will not trick you.&#8221;</p></li><li><p>&#8220;We will not punish curiosity.&#8221;</p></li><li><p>&#8220;We will not create regret.&#8221;</p></li></ul><p>Because the best monetization is not extracting value.</p><p>It is <em>aligning value</em>.</p><p>It is the user thinking: &#8220;I&#8217;m paying because this helps me. And it feels fair.&#8221;</p><p>That feeling cannot be generated by more data.</p><p>It requires conviction.</p><p>So how do you actually build it? I have been using a five-step framework I call the Conviction Loop for every decision that matters, features, onboarding, positioning, paywalls. It is not a scoring matrix. It is a forcing function. And it comes with a worksheet you can copy into any decision you are staring at right now.</p><p><em>For paid subscribers: the full Conviction Loop framework, the copy-paste worksheet, and a worked example using a real paywall decision below.</em></p>
      <p>
          <a href="https://www.thecompoundingfounder.com/p/options-are-cheap-conviction-is-rare">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[A solo founder building in public. One artifact at a time.]]></description><link>https://www.thecompoundingfounder.com/p/coming-soon</link><guid isPermaLink="false">https://www.thecompoundingfounder.com/p/coming-soon</guid><dc:creator><![CDATA[Eduardo]]></dc:creator><pubDate>Thu, 26 Feb 2026 14:34:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ay6o!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F365e2f47-28e8-428d-9c67-44413c616ce1_1287x1287.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most founder content tells you what worked.</p><p>This one tells you what it actually feels like while it is happening, not after the outcome made it look obvious.</p><p>I am Eduardo. I am building Clueless Clothing, an AI wardrobe app that helps people love the clothes they already own. Solo. No team. Fourteen months in. Still going.</p><p>The Compounding Founder is where I think out loud about the decisions that do not have clean answers.</p><p>Not the retrospective. The middle.</p><p>What it feels like to stare at twelve paywall options and know that data will not tell you which one is yours. What it means to build craft into a product when AI can generate every plausible alternative in four seconds. What it costs to commit to something when the cost of hedging has never been lower.</p><p>Every post either includes a tool, a skill file, or a framework you can actually use, or it does not get published. I do not write about building. I write from inside it.</p><p>Paid subscribers get the working files: including every SKILL.md, build artifacts, prompts, etc&#8230; If I am using it then you get it also.</p><p>Free subscribers get the essay.</p><p>The calculation I run every morning is simple: if you are going to build something, you should be able to decide something. That is what this is about.</p><p>&#8594; Subscribe if you are in the middle of it too.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.thecompoundingfounder.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.thecompoundingfounder.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>