<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[BuilderLab.ai]]></title><description><![CDATA[0→1→100 scaling, strategy, and No BS AI — for people who actually build.]]></description><link>https://www.builderlab.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!BoiJ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0e09a97-71ce-40b8-85cd-9976fe218d89_268x268.png</url><title>BuilderLab.ai</title><link>https://www.builderlab.ai</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 21:15:23 GMT</lastBuildDate><atom:link href="https://www.builderlab.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Pranav Pathak]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[buildersbriefingroom@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[buildersbriefingroom@substack.com]]></itunes:email><itunes:name><![CDATA[Pranav Pathak]]></itunes:name></itunes:owner><itunes:author><![CDATA[Pranav Pathak]]></itunes:author><googleplay:owner><![CDATA[buildersbriefingroom@substack.com]]></googleplay:owner><googleplay:email><![CDATA[buildersbriefingroom@substack.com]]></googleplay:email><googleplay:author><![CDATA[Pranav Pathak]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Pricing AI: What Actually Works]]></title><description><![CDATA[How to price AI products and features without destroying your margins or your customer relationships.]]></description><link>https://www.builderlab.ai/p/pricing-ai-what-actually-works</link><guid isPermaLink="false">https://www.builderlab.ai/p/pricing-ai-what-actually-works</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 05 Apr 2026 05:01:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/33a2bf94-b820-4f97-9fca-39e4fcd81555_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>SaaS pricing was a solved problem for fifteen years. Charge per seat, per month. Predictable for the buyer, predictable for the seller, beautiful for the board deck. Gross margins north of 80%. Everyone was happy, and nobody had to think very hard.</p><p>AI broke the math.</p><p>Not the model, the math. The cost to serve a traditional SaaS user is close to zero. The cost to serve an AI user is not. Every inference call, every agentic loop, every long-context request costs real money. AI companies run gross margins of 50-60%, compared to 80-90% for traditional SaaS. And those are the <em>good</em> AI companies. Early-stage ones can run as low as 25%.</p><p>But here&#8217;s what most pricing analyses get wrong: the margin compression isn&#8217;t the hard problem. The hard problem is that AI inverts the relationship between product success and pricing power. When your AI agent does the work of five people, charging per seat penalizes you for the exact outcome your product delivers. If your product works spectacularly, the customer needs <em>fewer</em> of the thing you charge for.</p><p>That&#8217;s a business model crisis.</p><p>Seat-based pricing dropped from 21% to 15% of companies in just twelve months. Hybrid models surged from 27% to 41%. Credit-based models grew 126% year-over-year. The PricingSaaS 500 Index tracked over 1,800 pricing changes across the top 500 B2B and AI companies in 2025 alone. That&#8217;s 3.6 changes per company in a single year. The industry is running experiments in public, at scale, and nobody has converged on the answer.</p><p>This piece isn&#8217;t a survey of those experiments. There are good ones out there. This is a framework for thinking about the structural question underneath all of them: what is your product actually worth, who knows it, and can you measure it cleanly enough to charge for it?</p><div><hr></div><h2><strong>The Litmus Test That Matters More Than Any Framework</strong></h2><p>Before we get into models, one question separates companies that will get pricing right from companies that will iterate forever:</p><p><strong>If your product succeeds spectacularly, does the customer need fewer of the thing you charge for?</strong></p><p>If yes, you&#8217;ve picked the wrong metric. Full stop. It doesn&#8217;t matter how elegant your pricing page looks or how well your sales team can pitch it. You&#8217;re building a machine that destroys its own revenue as it improves. Every product improvement is a pricing headwind.</p><p>This is the single most important question in AI pricing, and most companies get it wrong because they inherit their pricing metric from whatever they charged for before they added AI. The seat was the unit because bodies were the input. AI changes the input. If you don&#8217;t change the unit, you&#8217;re optimizing a dashboard that&#8217;s lying to you.</p><div><hr></div><h2><strong>Four Models, and When Each One Kills You</strong></h2><h3><strong>Seats: The Bridge, Not the Destination</strong></h3><p>Per-seat pricing is comfortable. Buyers understand it. CFOs can budget it. Procurement can approve it. It worked for Salesforce, it worked for Slack, it worked for every enterprise SaaS company of the last decade.</p><p>The problem: AI inverts the logic. Per-seat pricing was magical for Salesforce because more sales reps drove more revenue. For an AI that writes the emails those reps used to write, charging per seat penalizes you for the efficiency gain you just delivered.</p><p>Microsoft Copilot charges $30 per user per month on top of an existing Microsoft 365 license. It works as a bridge because Microsoft has distribution and lock-in that makes the per-seat premium tolerable. Most companies don&#8217;t have that luxury.</p><p><strong>Use it when:</strong> AI is a lightweight enhancement to an existing workflow, usage is relatively uniform across users, and you&#8217;re adding AI features to a product that already charges per seat. Think: AI-assisted search, smart suggestions, grammar checking. The AI makes the seat more valuable without replacing the person in it.</p><p><strong>Kill it when:</strong> Your P90 user costs 10x your P50 user in compute. When that happens, you&#8217;re subsidizing power users with light users, and your margins are eroding invisibly. The thing nobody tells you about AI cost distributions is that they aren&#8217;t bell curves. They&#8217;re power laws. Your heaviest users don&#8217;t cost a little more. They cost orders of magnitude more. And you won&#8217;t see it in your averages until it&#8217;s already in your margin.</p><h3><strong>Usage-Based: Honest but Dangerous</strong></h3><p>Usage-based pricing aligns revenue with cost. That&#8217;s its strength and its trap. You charge for tokens, API calls, credits, compute minutes, whatever maps to the resource being consumed. By 2025, 85% of SaaS companies were using some form of usage-based pricing.</p><p>The appeal is real. It lets small customers start cheap and grow. It protects your margins because revenue scales with cost. It&#8217;s transparent.</p><p>But transparency cuts both ways, and in AI, the variance between a simple request and a complex one can be 100x. A user doesn&#8217;t know that their three-sentence prompt triggered a six-step agentic loop that burned through their monthly allocation.</p><p>Cursor is the cautionary tale everyone should study. In June 2025, Cursor switched from a flat 500-requests-per-month plan to a credit pool tied to API rates. The economics were sound. Frontier models were getting more expensive, and Cursor was eating the cost under flat pricing. They had reportedly reached over $500 million in ARR while spending 100% of revenue on AI costs. Something had to change.</p><p>But the rollout was a catastrophe. Users who expected predictable billing got surprise charges. Some burned through their monthly allocation in a few prompts. The community backlash was so severe that the CEO published a public apology and offered refunds.</p><p>The lesson isn&#8217;t that usage-based pricing is wrong. The lesson is that <em>unpredictable</em> usage-based pricing is lethal. If a customer&#8217;s bill is ever a surprise, you&#8217;ve already lost their trust. And in AI, where the customer can&#8217;t see the compute their request triggers, surprise is the default unless you build aggressively against it: usage dashboards, threshold alerts, spending caps, model-cost previews. These aren&#8217;t features. They&#8217;re survival infrastructure.</p><p><strong>Use it when:</strong> You&#8217;re selling API access, infrastructure, or developer tools where the buyer expects metered billing. Your cost scales linearly with consumption. Your customers are technical enough to understand and monitor usage.</p><p><strong>Kill it when:</strong> Your buyers are business users who budget monthly and hate surprises. Or when your product encourages the kind of exploration and experimentation that makes metered billing feel punitive. Nothing kills adoption faster than a user thinking, &#8220;should I really ask this question, or will it cost too much?&#8221;</p><h3><strong>Outcome-Based: The Ideal That Most Products Can&#8217;t Reach</strong></h3><p>Outcome-based pricing is where incentives align perfectly. The customer pays when the AI delivers something valuable. If the AI fails, the customer pays nothing. It sounds like the obvious future of AI pricing. It isn&#8217;t, for most companies. But in specific domains, it&#8217;s devastating.</p><p>Intercom&#8217;s Fin is the reference case. When Intercom launched its AI agent in 2023, they charged $0.99 per successful resolution instead of per seat. Not per conversation. Not per API call. Per resolved issue, confirmed by the customer. Fin now resolves over 56% of conversations autonomously and generates tens of millions in annual revenue. The per-resolution model doubled adoption rates compared to seat-based alternatives.</p><p>The reason Fin works on outcome-based pricing isn&#8217;t that Intercom is clever about billing. It&#8217;s that customer service has a structural property most AI use cases don&#8217;t: the outcome is binary, the attribution is clean, and the counterfactual is obvious. Either the ticket got resolved or it didn&#8217;t. There&#8217;s no committee required to decide if value was delivered. No attribution model to argue over. No six-month lag between action and result.</p><p>Having worked in customer service automation, I can tell you this clarity is rare and precious. Advertiser retention (another area I have scars from) has the same structural property: either the churning advertiser came back or they didn&#8217;t. The line from AI action to business result is unmistakable. You can draw a box around the interaction, measure what went in, measure what came out, and price the delta.</p><p>That&#8217;s the prerequisite for outcome-based pricing: a binary outcome, clean attribution, and a value the customer already quantifies in their own business. Support resolutions. Completed transactions. Retained accounts. Verified lead qualifications. These domains share a structural shape that makes outcome-based pricing not just possible but natural.</p><p>Most AI products don&#8217;t operate in these domains. Most AI products operate where the outcome is soft, shared, or delayed. &#8220;Better decisions.&#8221; &#8220;Improved productivity.&#8221; &#8220;Faster workflows.&#8221; These are real, but you can&#8217;t put a price on them per instance because you can&#8217;t isolate the AI&#8217;s contribution from everything else that happened.</p><p>Salesforce learned this with Agentforce. They launched at $2 per conversation. Customers immediately started asking what counted as a &#8220;conversation.&#8221; Was a multi-turn exchange one conversation or several? What if the AI handled part and a human finished? The problem wasn&#8217;t the price. It was that &#8220;conversation&#8221; isn&#8217;t a clean outcome. It&#8217;s an activity metric dressed up as a value metric. Salesforce has since introduced three different pricing models in eighteen months: per-conversation, flex credits at $0.10 per action, and per-user licenses at $125 per month. They&#8217;re running all three simultaneously because they still haven&#8217;t found the answer.</p><p>Outcome-based pricing isn&#8217;t a maturity stage every company will eventually reach. It&#8217;s a structural fit. It depends on the measurability and attributability of the outcome in your specific domain. Kyle Poyar&#8217;s data shows that only about 5% of companies can actually pull it off. The other 95% aren&#8217;t behind. They&#8217;re in domains where the outcome can&#8217;t be cleanly isolated. And that&#8217;s fine, as long as they stop pretending they&#8217;re on a journey toward outcome-based pricing and start building the model that actually fits their product.</p><p><strong>Use it when:</strong> Your outcome is measurable, binary, and unambiguously attributable to your product. The value of that outcome should already have a dollar figure in the customer&#8217;s mind. If you have to convince them of the value, you&#8217;re not ready.</p><p><strong>Kill it when:</strong> Attribution is murky, outcomes are soft, or your model&#8217;s performance varies significantly. If your AI has bad weeks, your revenue has bad weeks. And if the customer has to <em>trust</em> that value was delivered rather than <em>see</em> it, you&#8217;re one bad quarter away from a churn event.</p><h3><strong>Hybrid: Where Most Companies Should Be (For Now)</strong></h3><p>Hybrid pricing combines a base subscription with usage or outcome tiers. It&#8217;s the pragmatic middle ground: predictable for the buyer, margin-protective for the seller, and it lets you learn.</p><p>Over 60% of SaaS companies now use hybrid models. The structure is typically a monthly base fee that covers a defined level of usage, with overages or premium tiers beyond that.</p><p>The base fee gives the CFO a number to budget. The usage tier gives you upside when the customer scales. The customer feels safe starting, and you don&#8217;t subsidize power users forever.</p><p>The design principle matters: your base tier should cover 70-80% of your users comfortably. The usage tier should kick in for the top 20-30% who are getting disproportionate value. If most users hit the usage tier, your base is too low and you&#8217;re functionally usage-based with extra steps. If almost nobody hits it, your usage tier is a rounding error and you&#8217;re functionally subscription with an illusion of upside.</p><p>Anthropic&#8217;s own tiering is instructive. Free, Pro at $20, Max at $100, Team at $30 per seat. Each tier serves a genuinely different user with different behavior patterns. A casual user and a developer running Claude Code all day aren&#8217;t &#8220;light&#8221; and &#8220;heavy&#8221; versions of the same persona. They&#8217;re different products serving different jobs.</p><p><strong>Use it when:</strong> You&#8217;re uncertain about the right model. Hybrid lets you collect data on usage patterns, cost distributions, and willingness to pay while maintaining customer predictability. Start here, run it for 90 days, and let the data tell you where to evolve.</p><p><strong>Kill it when:</strong> Complexity is killing your sales cycle. If you can&#8217;t explain your pricing in one sentence, it&#8217;s too complicated. Every layer of complexity adds friction to the purchase decision, and friction compounds across the sales cycle in ways that don&#8217;t show up until your win rates start declining.</p><div><hr></div><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/pricing-ai-what-actually-works?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Enjoying this? Spread the word, it helps me keep writing!</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/pricing-ai-what-actually-works?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/pricing-ai-what-actually-works?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h2><strong>Why the Margin Problem Is Structural, Not Temporary</strong></h2><p>Before you pick a model, you need to accept something uncomfortable: AI margins are structurally lower than SaaS margins, and they&#8217;re not going back up.</p><p>Traditional SaaS has near-zero marginal cost per user. AI does not. Every inference call costs money. Every agentic loop burns tokens. Every long-context request hits your GPU bill. The industry runs 50-60% gross margins versus 80-90% for traditional SaaS.</p><p>This isn&#8217;t a temporary phase that better models or cheaper compute will fix. Even as inference costs decline, product complexity increases. Models get more capable, which means customers use them for harder tasks, which means more compute per request. The cost curve flattens but it doesn&#8217;t collapse. Plan for 50-65% gross margins for the foreseeable future.</p><p>Three things follow from this:</p><p><strong>Track your cost distribution, not your average.</strong> What does your P90 user cost versus your P10? The average is meaningless if your distribution is skewed, and in AI, it&#8217;s almost always skewed. The cost per user you measure in month one is also meaningless because your users haven&#8217;t discovered the expensive features yet. The real cost curve doesn&#8217;t stabilize for at least 90 days.</p><p><strong>Run the litmus test.</strong> Does product success reduce the thing you charge for? If yes, change the metric before you have 10,000 customers locked into the wrong model.</p><p><strong>Model your margins at scale.</strong> If the math doesn&#8217;t work at 10 customers, it won&#8217;t work at 10,000. AI costs don&#8217;t improve as much with scale as SaaS costs did. The per-unit economics of your hundredth customer look a lot like your tenth.</p><div><hr></div><h2><strong>Adding AI to a Product That Already Has Pricing</strong></h2><p>Most of us aren&#8217;t building AI-native companies from scratch. We&#8217;re adding AI features to products that already have pricing, customers, and expectations. This is harder than greenfield pricing because you have constraints and relationships to protect.</p><p>Four options, each with a sharp edge:</p><p><strong>Absorb it.</strong> Bundle the AI feature into your existing plan at no extra charge. Google did this with Workspace. The bet: AI makes the core product stickier, reduces churn, and justifies future price increases. This works when the AI feature is lightweight, broadly used, and compute cost per user is low enough to absorb. It doesn&#8217;t work when usage varies widely. You&#8217;ll subsidize power users with light users and watch your gross margins compress without knowing why.</p><p><strong>Add-on.</strong> Sell the AI feature as a separate line item. Notion charges $8 per member per month for its AI add-on. Microsoft charges $30 per user for Copilot on top of existing licenses. This works when the AI feature is clearly differentiated and not every user needs it. It doesn&#8217;t work when the add-on feels like a tax on features that should have been included. Watch your NPS after launching an AI add-on. If it drops, you&#8217;ve mispriced the relationship, not the feature.</p><p><strong>New tier.</strong> Create a new plan tier that includes AI features. This works when AI features naturally cluster with other premium capabilities, and you can draw a line between tiers that feels like a product boundary rather than an arbitrary usage cutoff. It doesn&#8217;t work when the only difference between tiers is &#8220;more AI.&#8221; If your tiers are just usage gates with different labels, customers will see through it.</p><p><strong>Usage gate.</strong> Give everyone access, meter it, and charge for heavy usage. The freemium-to-usage pipeline. This works when you want maximum adoption and your free-to-paid conversion is healthy (above 2-3%). It doesn&#8217;t work when your free tier is too generous. OpenAI reportedly burned $8 billion annually on compute in 2025 while serving 900 million weekly free users. Most companies can&#8217;t afford that bet. If conversion is below 2%, you&#8217;re running a charity.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I write frameworks you can use Monday morning, insights on building with AI, and deep dives into the most successful AI companies.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>Five Things That Kill AI Pricing</strong></h2><p><strong>Pricing on cost instead of value.</strong> Your inference cost is your problem, not your customer&#8217;s. Customers don&#8217;t care what a token costs. They care what the outcome is worth. A support resolution worth $15 to the customer should be priced on that value, not on the $0.03 it cost you in compute. Cost sets your floor. Value sets your price. Most companies confuse the two.</p><p><strong>Inheriting the old metric.</strong> Per-seat pricing was designed for a world where marginal cost per user was near zero and more users meant more value created. AI products don&#8217;t live in that world. The number of people logging in tells you nothing about the value being delivered or the cost being incurred. Using the old metric because it&#8217;s familiar is how you wake up to negative margins.</p><p><strong>Surprise bills.</strong> The Cursor debacle came down to one failure: customers couldn&#8217;t predict their spend. In AI, where a single complex request can trigger a chain of expensive operations, prediction is hard. Build the transparency infrastructure before you need it. Usage dashboards, threshold alerts, spending caps. If you wait until customers complain, you&#8217;ve already lost them.</p><p><strong>Treating pricing as a launch decision.</strong> The market tracked 3.6 pricing changes per company in 2025. Credit-based models grew 126% year-over-year. The best companies are iterating quarterly. If you&#8217;re revisiting pricing once a year, you&#8217;re falling behind the companies that are treating pricing as a continuous experiment.</p><p><strong>Soft ROI positioning.</strong> Copilots that offer advice without closing the loop live in dangerous territory. &#8220;Are we really getting value?&#8221; is the question that kills renewals. As AI pilots from 2025 hit their renewal cycles, pricing must reflect actual value delivered, not potential or promise. Measure outcomes even when you don&#8217;t price on them. Build the dashboards, establish baselines, create feedback loops. This builds the trust that sustains pricing power. And when you&#8217;re ready to move to outcome-based pricing (if your domain supports it), you&#8217;ll have the data to make the case.</p><div><hr></div><h2><strong>The Agentic Frontier</strong></h2><p>Everything above applies to AI features and copilots. Agents are a different animal.</p><p>An AI agent doesn&#8217;t assist a user. It replaces a workflow. It resolves the ticket, books the meeting, drafts the contract, processes the claim. Autonomously. Agents don&#8217;t log in. They don&#8217;t hold licenses. They can complete thousands of tasks in the time a human completes one.</p><p>Charging per-seat for an agent is like charging per-parking-space for a self-driving car fleet. The unit of value isn&#8217;t the seat. It&#8217;s the work.</p><p>This is why &#8220;digital labor&#8221; pricing is emerging: per action, per resolution, per completed workflow. Salesforce now lets enterprises convert user licenses into flex credits and back, treating human and AI labor as interchangeable budget lines.</p><p>The market hasn&#8217;t converged on the right model for agents, and it won&#8217;t for at least another 12-18 months. If you&#8217;re building agents, price on the work they do, not on access to them. Apply the structural test: is the outcome of the work binary and measurable? If yes, you have a real shot at outcome-based pricing. If not, start with hybrid and instrument everything so you can evolve.</p><p>Nobody has this figured out. Not Salesforce, not Intercom, not your competitors. The companies that learn fastest will win.</p><div><hr></div><h2><strong>The Decision Sequence</strong></h2><p>If you take one thing from this piece, make it this sequence:</p><p><strong>First:</strong> Know your cost structure. Not the average. The distribution. What does your P90 user cost? If you don&#8217;t know, stop and go find out. You cannot price what you cannot measure.</p><p><strong>Second:</strong> Identify the value metric your customer already uses. Not what you want to charge for. What they believe they&#8217;re buying. If they&#8217;re buying resolutions, don&#8217;t charge for tokens. If they&#8217;re buying productivity, don&#8217;t charge for seats.</p><p><strong>Third:</strong> Run the litmus test. If your product succeeds, does the customer need fewer of the thing you charge for? If yes, you&#8217;ve picked the wrong metric.</p><p><strong>Fourth:</strong> Match the model to the buyer. Enterprise CFOs want predictable line items. Developers expect metered billing. SMB founders want simple monthly fees. Your pricing model must fit how your buyer purchases, not how your product works internally.</p><p><strong>Fifth:</strong> Start hybrid if uncertain. Base subscription plus usage tiers. Collect data for 90 days. Then optimize. Don&#8217;t try to be clever with pricing before you have usage data. Cleverness without data is just guessing with confidence.</p><div><hr></div><h2><strong>The Bottom Line</strong></h2><p>AI pricing forces you to confront things SaaS pricing let you ignore: real marginal costs, wildly variable usage patterns, and a value metric that might not be the one you inherited.</p><p>The industry is in the messy middle of hybrid models, quarterly iterations, and learning in public. That&#8217;s fine. The point isn&#8217;t to pick the perfect model. The point is to understand the structural question your product sits on top of: Is the outcome binary? Is attribution clean? Does the customer already quantify the value?</p><p>If yes to all three, price on outcomes. You&#8217;re in the 5% with a structural advantage.</p><p>If no, start hybrid, instrument everything, measure outcomes even if you don&#8217;t price on them, and iterate faster than your competitors.</p><p>The companies winning at pricing right now aren&#8217;t the ones who picked the perfect model on day one. They&#8217;re the ones who understood the structure of their problem, picked a starting point, and learned faster than everyone else.</p><p>ROI or die.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for practical strategy on pricing, shipping, and scaling AI products.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Builder Bubble]]></title><description><![CDATA[Why the Next 2-3 Years Are the Biggest Asymmetric Bet of Your Career]]></description><link>https://www.builderlab.ai/p/the-builder-bubble</link><guid isPermaLink="false">https://www.builderlab.ai/p/the-builder-bubble</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 29 Mar 2026 09:56:03 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/828af10e-a125-4121-87ea-e94ce78a8c8d_960x540.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you read BuilderLab, you are probably in a bubble. The bubble is your advantage.</p><p>You (or your team) shipped a feature last week that would have taken a team of three a month to build two years ago. You did it in a day, mostly by talking to an AI. And the strangest part was not the speed. It was that nobody around you seemed to notice how abnormal that is.</p><p>This is the feeling that builders have right now, and struggle to articulate. Something has shifted. Quietly. Practically. Daily. You can do things that you could not do before. Lots of things. Fast. And when you look up from your screen and talk to people outside your bubble, you realize they have no idea.</p><p>16.3% of the world&#8217;s working-age adults have used an AI tool. Of the people who use AI at work, only 5% use it in ways that actually transform their output. Meanwhile, builders in the 95th percentile of AI adoption are producing 6-17x more than the median worker using the same tools.</p><p>Completely different realities.</p><p>The conversation about AI is stuck on whether it will replace jobs. That question is years premature. The real story is simpler: a small group of people is building at a pace the general population cannot comprehend.</p><h1>The Ability Bubble</h1><p>Call it what it is: an ability bubble. A temporary period where a specific cohort of builders can do dramatically more than everyone else, because of tool adoption and workflow design.</p><p>Solo founders are shipping products that used to require teams of ten. A developer with Claude Code or Cursor is writing in a day what used to take a sprint. A product manager who never wrote code is deploying production software. Structural shifts in what one person can accomplish.</p><p>And the people doing this are pulling further ahead every week.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I write about building with AI every week. Strategy, tools, frameworks, and the occasional uncomfortable truth.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Why the Bubble Expands (Not Contracts)</h2><p>Your intuition says this should not last. Early adopter advantages shrink as tools go mainstream. That is how it worked with email, spreadsheets, smartphones. Everyone catches up. The advantage normalizes.</p><p>AI does not work this way. Three forces keep the bubble expanding.</p><p><strong>AI skills compound.</strong> The distance between using ChatGPT for summarization and orchestrating multi-agent workflows across research, code, and deployment is not a weekend workshop. It is months of daily iteration. Every hour you spend with these tools teaches you something that makes the next hour more productive. The builder who started in 2024 has two years of compounded intuition that cannot be shortcut. This is like learning to code: the skill itself creates the capacity to learn more of the skill.</p><p><strong>The rest of the world faces structural barriers.</strong> The 55-year-old operations manager at a mid-market company has access to ChatGPT. What they lack is the workflow context, the prompt intuition, the tool-chaining mental models, and the organizational permission to rebuild how they work. Only 35% of enterprises have a mature AI upskilling program. The training infrastructure itself is the bottleneck.</p><p><strong>Builder tools are accelerating faster than consumer tools.</strong> The tools getting dramatically better every quarter are Claude Code, Cursor, Windsurf, Devin, Replit Agent. Coding assistants now write 41% of all code. 73% of engineering teams use AI coding tools daily, up from 18% two years ago. Every improvement makes builders more productive while doing nothing for the person who uses AI to rewrite emails.</p><p>Tool velocity times adoption intensity. Builders have both.</p><h1>The 2-3 Year Window</h1><p>The gap will close eventually. But not soon.</p><p>AI interfaces will become invisible, embedded into every tool as default behavior rather than a separate thing you choose to use. Most enterprise software still treats AI as an add-on. That transition takes 2-4 years across the long tail. Education will catch up, but most corporate L&amp;D programs are still running 2023-era prompt workshops. Organizational culture will shift from permission mode, where employees need approval to use AI, to expectation mode, where not using it requires justification. But culture change in large organizations is measured in years.</p><p>Add these timelines together: 2-3 years where builders operate at a fundamentally different level than everyone else.</p><p>After that, the advantage does not disappear. It normalizes. The edge shifts from &#8220;I can do what you cannot&#8221; to &#8220;I can do it slightly faster.&#8221; Slightly faster is a far less valuable position than categorically different.</p><h1>The Three Psychological Barriers Nobody Talks About</h1><p>The data says the bubble is real. The structural forces say it is expanding. The timeline says you have 2-3 years. So why are so many people inside the bubble still not building? Why are you not building - and shipping? And not monetising?</p><p>Because the biggest barriers are not technical. They are psychological. And they hit differently depending on which side of the code divide you sit on.</p><p><strong>Barrier 1: &#8220;I&#8217;m not a real developer.&#8221;</strong></p><p>If you do not code for a living, building with AI feels like trespassing. You are a product manager, a designer, a marketer, a founder with a business background. And here you are, shipping actual software. It feels like an outside-in move. Like you snuck into someone else&#8217;s domain through a side door that was not supposed to be open.</p><p>This is the primary reservation for most people who have not shipped yet. Not that they cannot figure out the tools. That they feel like they should not be allowed to. That somewhere there is a credential they are missing, a gatekeeping exam they never sat, a permission they were never granted.</p><p>I know this because it happened to me. The first time I shipped something built with AI coding tools, the dominant feeling was not pride. It was a quiet suspicion that I was getting away with something. That a &#8220;real&#8221; developer would look at what I built and see through it.</p><p>Here is the truth: nobody is checking your credentials at the deploy button. The user does not care whether you have a CS degree. The customer does not care whether you wrote the code by hand or with Claude. They care whether the product solves their problem. That is the only qualification that has ever mattered, and AI just made it accessible to a much larger group of people.</p><p>The mindset shift is simple but hard: you are not trespassing. The door is open because the walls came down.</p><p><strong>Barrier 2: &#8220;I&#8217;m cheating.&#8221;</strong></p><p>If you do code for a living, the barrier is the opposite. Using AI feels like cutting corners. You built your career on craftsmanship: understanding systems deeply, writing clean code, catching edge cases, knowing why something works and not just that it works. Now an AI writes 80% of the code and you are left wondering what you missed.</p><p>Obsessing over details becomes a pattern. Did the AI introduce a subtle bug? Is there a security hole I did not catch? Am I still a good engineer if I did not write this myself?</p><p>This is an identity crisis disguised as quality control. The anxiety is not really about bugs. It is about what it means to be a developer when the core act of writing code is no longer the bottleneck. The answer is that your value was never in the typing. It was in the judgment: knowing what to build, how to architect it, where the failure modes live, and when to override the machine. AI makes that judgment more valuable, not less.</p><p><strong>Barrier 3: &#8220;If I can build this in a week, it must be worthless.&#8221;</strong></p><p>This one is the most insidious, because it feels like logic. You built something in a week with AI tools. Your immediate thought: if it was that easy for me, anyone can do it. So why would anyone pay for it?</p><p>This is where the bubble shows up as a psychological blind spot. You are inside the bubble. You have the tool fluency, the workflow intuition, the domain knowledge, and the prompt engineering instincts that make a week-long build feel effortless. You forget that most people are still struggling with barriers one and two, if they have gotten that far at all. Many have not even started. And even among those who have, most lack the domain depth to know what is worth building in the first place.</p><p>The ease of your build is not evidence that the product is worthless. It is evidence that you are inside the bubble. The 16.3% number is your answer: the vast majority of the working world cannot do what you just did. Not yet. Not for another 2-3 years. And in that window, the thing you built in a week has real, defensible value to the people outside the bubble who need it.</p><p>Stop discounting your output by the effort it took. Start pricing it by the problem it solves.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/the-builder-bubble?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Know a builder who needs to hear this? Share this article with them. The barriers are quieter than the hype, and most people fight them alone.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/the-builder-bubble?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/the-builder-bubble?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1>How to Capitalize: Five Moves for Builders</h1><p>The question is not whether you have an advantage. If you are reading this, you almost certainly do. The question is whether you are extracting maximum value from a temporary asymmetry. Most builders are not. They are using AI to do their existing job 30% faster instead of using AI to build things that were previously impossible.</p><p>Here are five moves that separate builders who capitalize from builders who coast.</p><p><strong>Move 1: Build products, not just productivity.</strong></p><p>The most common mistake: using AI exclusively to speed up your current work. Faster emails. Faster code reviews. Faster slide decks. This is real value, and you should capture it. But the asymmetric opportunity is somewhere else.</p><p>The asymmetric opportunity is building new things. Solo-founded startups surged from 23.7% of all new companies in 2019 to 36.3% by mid-2025. A solo founder with the right AI stack can now operate at a level that previously required a team of five to ten. Operating margins of 60-80% on a tool stack that costs $100-500 per month.</p><p>Danny Postma runs HeadshotPro solo at $3.6 million ARR. Maor Shlomo built Base44 to 250,000 users and sold it to Wix for $80 million within six months. These are not outliers of talent. They are outliers of timing and tool adoption.</p><p>You do not need to quit your job. But you should be building something on the side. A product, a tool, a service, a system. Something that converts your AI fluency into an asset that exists outside your employer&#8217;s org chart.</p><p><strong>Move 2: Invest in workflow design, not just tool selection.</strong></p><p>The 6x productivity gap comes from workflow architecture, not tool choice. The power users and the median employees are often using the same tools. The difference is how tasks are decomposed, how AI is chained across steps, how outputs from one model become inputs to another, and how human judgment is inserted at the right checkpoints.</p><p>The real question is: &#8220;how do I redesign this entire process so that AI handles 80% and I handle the 20% that requires judgment?&#8221;</p><p>The builders who compound fastest are the ones who spend time designing workflows, not just executing tasks. Every workflow you build is a template you can reuse, share, sell, or license. Your workflow library is your moat.</p><p><strong>Move 3: Ship publicly and build reputation now.</strong></p><p>In 2-3 years, everyone will claim AI expertise. Right now, the proof is thin. The people who are visibly building, shipping, and sharing their work in public during this window will own the credibility when the market matures.</p><p>Write about what you are building. Show the process, not just the output. Document your failures and your iterations. The builders who establish authority during the bubble will be the ones hired, funded, and followed when the rest of the market catches up.</p><p>This is about converting a temporary skill advantage into a durable reputation asset.</p><p><strong>Move 4: Build your data and distribution moats early.</strong></p><p>AI tools are commoditizing. The models get cheaper every quarter. The interfaces get easier. The competitive advantage lives in proprietary data that makes AI more valuable and distribution channels that give your AI-powered products an audience.</p><p>Every product you build should be designed to collect unique, defensible data with every use. Every workflow you automate should connect to an audience or customer base you own. The AI advantage fades. The data and distribution advantages do not.</p><p>This is the difference between someone who used AI to build a tool and someone who used AI to build a business. The tool can be copied. The business cannot.</p><p><strong>Move 5: Hire or partner to fill your gaps, while the talent is underpriced.</strong></p><p>Here is a counterintuitive implication of the bubble: the people who have AI building skills are in an expanding bubble, but the market has not fully priced this in yet. The 56% wage premium is growing but still lags the actual productivity differential. A developer who can build AI-native workflows is generating 6-17x more output, but they are not being paid 6-17x more.</p><p>If you are hiring, this is the window to lock in AI-fluent talent before the market fully prices the skill. If you are a solo builder, this is the window to partner with other builders who complement your gaps, before the talent pool gets priced up to match the productivity premium.</p><h1>The Uncomfortable Truth</h1><p>There is one more thing that does not get said enough in AI commentary, because it sounds elitist and uncomfortable.</p><p>The bubble is also a compounding inequality.</p><p>The people who benefit most from AI tools are the people who were already high-performers: educated, technically literate, in knowledge-work roles, in high-income countries, at companies that provide access and permission. Microsoft&#8217;s data shows it clearly: AI usage is concentrated among younger, educated, higher-income workers in the Global North.</p><p>Worth acknowledging, because ignoring it leads to bad strategy. If you are in this cohort, your advantage is real but not earned purely through merit. It is partially a function of circumstance. And the ethical response to that is not guilt. It is urgency. Use the advantage while it exists. Build things that create value beyond your personal gain. And recognize that the window where this kind of asymmetry is possible is historically rare and structurally temporary.</p><h1>The Close</h1><p>Let me return to where we started. Two populations, one timeline.</p><p>16.3% of the world&#8217;s working-age adults have used AI. 5% of employees are maximizing it. And inside that 5%, there is a smaller group still: the people who are not just using AI to be more productive, but building with it. Creating products, systems, workflows, and businesses that did not exist two years ago.</p><p>That group is in a bubble. A bubble of capability that the rest of the world has not caught up to yet, and will not catch up to for another 2-3 years.</p><p>The question is not whether you are in the bubble. The question is what you are building while you are inside it.</p><p>The tools are temporary. The models will change. The interfaces will evolve. But the products you ship, the reputation you build, the data you collect, and the distribution you own will outlast the window.</p><p>Build now. The bubble is real. And it will not last forever.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>If this piece shifted how you think about the next 2-3 years, <a href="https://claude.ai/claude-code-desktop/link">subscribe</a> so you don't miss what comes next.</em></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Most AI UX Is Lazy]]></title><description><![CDATA[And why we love to hate chatbots]]></description><link>https://www.builderlab.ai/p/most-ai-ux-is-lazy</link><guid isPermaLink="false">https://www.builderlab.ai/p/most-ai-ux-is-lazy</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 08 Mar 2026 09:34:44 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2498cde6-09f7-4693-8f67-3d8241f10354_360x360.avif" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Every AI feature shipped in the last two years looks the same.</p><p>A chat box, a loading spinner, a disclaimer at the bottom, and the interface is a text field that says &#8220;Ask me anything.&#8221;</p><p>The problem is not the AI. The problem is that product builders don&#8217;t ask the right question. They ask &#8220;how do I add AI to our product?&#8221; They should be asking: &#8220;What user problem can I solve now that was previously unsolvable?&#8221;</p><p>When you skip that question, you get the default: a chatbot. When you answer it, you get a lot more success. My most successful launches using AI were not chat interfaces. Some of them were text fields that turn into checkboxes. Some were descriptions and summaries. Some others were just high performing classifiers. More on that later.</p><p>Lazy AI UX comes from lazy AI thinking. Teams start with the interface instead of the capability.</p><h3><strong>The Four Lazy Defaults</strong></h3><p><strong>The Chatbot.</strong> Everyone&#8217;s first instinct. &#8220;Chat with your data.&#8221; Notion, Salesforce, HubSpot, Shopify. They all shipped it as their flagship AI feature. Chat shifts cognitive load to the user. &#8220;Figure out what to ask&#8221; is work for the user, that you&#8217;ve offloaded from the work you need to be doing. Your product should already know what the user needs. Chat is one AI primitive forced onto every problem regardless of fit. When your travel app, your CRM, your project management tool, and your email client all have the same chat interface, something has gone wrong with the thinking.</p><p><strong>The Magic Wand Button.</strong> &#8220;AI Generate&#8221; dropped next to every text field. No context about what it will generate. No way to guide it. No refinement loop. Click, get a blob of text, accept or regenerate. This is another single primitive (generation) with zero UX thought applied to it. The lottery machine approach to product design. Pull the lever, see what comes out, pull it again if you don&#8217;t like it.</p><p><strong>The Loading Spinner of Trust Destruction.</strong> Eight-second wait. No progress indication. Then streaming text appears character by character. Impressive in a demo, anxiety-inducing in production. Users cannot verify something that is still being written. They sit there watching words appear, not knowing whether to trust what they are reading or wait for the next sentence to contradict it. The latency is not a performance problem. It is a trust problem.</p><p><strong>The Disclaimer.</strong> &#8220;AI-generated content may contain errors.&#8221; Plastered at the bottom of every output as a substitute for designing verification into the experience. If your UX strategy for handling AI errors is a legal footnote, you did not design UX. You designed liability coverage.</p><h3><strong>Why Teams Default to Lazy</strong></h3><p>The problem is structural. These are not bad designers making bad decisions. These are reasonable people responding to bad incentives.</p><p><strong>Demo-driven development.</strong> Chat demos beautifully in board meetings. Executive types a question. AI answers. Room applauds. The demo becomes the product strategy. But what demos well and what gets adopted are different things. A chatbot that impresses a VP in a conference room can still get low adoption from actual users who have faster ways to get the same information.</p><p><strong>Model-out thinking instead of user-back thinking.</strong> &#8220;We have GPT-4, what can we do with it?&#8221; instead of &#8220;Users struggle with X, can AI help?&#8221; The capability drives the UX instead of the problem driving the capability. This is the root of lazy AI UX. Starting with the tool instead of the user.</p><p><strong>The API makes chat easy and everything else hard.</strong> The default API interface is messages in, text out. Chat is the path of least engineering resistance. Structured output, proactive surfacing, inline suggestions: all harder to build. So teams ship the easy thing and call it v1.</p><p><strong>No established patterns, so teams copy each other.</strong> When Notion shipped a chatbot, everyone shipped a chatbot. The industry is cargo-culting UX from whoever shipped first, regardless of whether it worked. Copying someone else&#8217;s lazy default does not make it less lazy.</p><p><strong>The Right Question: What Can You Do Now That You Could Not Before?</strong></p><p>AI gives you a set of primitives. Things that were previously impossible or prohibitively expensive to do at scale. So the question to ask is, you have a capability now that you did not have before. What problems  were previously unsolvable, that have now become possible to solve?</p><p>Finding the answer to this question requires a deep understanding of the capabilities the tech unlocks. I&#8217;m going to list down my mental model here. Yours may be slightly different, but as long as the broad strokes capabilities converge, you have a tool set that you can use to build a foundation on.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">BuilderLab publishes practical strategy for people who ship AI products. Frameworks, teardown, and real examples from products in production</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3><strong>The LLM Primitives</strong></h3><h4><strong>1. Intent Detection</strong></h4><p><strong>What it unlocks:</strong> understanding what users mean, not just what they type. Interpreting unstructured input and mapping it to structured actions at scale. This was previously impossible without building thousands of hand-coded rules that broke every time someone phrased something differently.</p><p><strong>Example:</strong> Booking.com Smart Filters. User types &#8220;hotels with pool and gym.&#8221; The AI maps that to existing structured filters: Swimming Pool and Fitness Center checkboxes, pre-selected. Property type set to Hotels. 42 results. The AI detected intent and translated it to structured action. No chat. No generated text. Full story below.</p><h4><strong>2. Generation</strong></h4><p><strong>What it unlocks:</strong> creating content that did not exist, at zero marginal cost. First drafts, suggestions, completions, variations. Previously, every piece of content required a human to write it from scratch.</p><p><strong>Example:</strong> inline drafts with refinement loops. GitHub Copilot suggests a line of code. Tab to accept. Keep typing to reject. The suggestion disappears with zero friction. Cursor takes it further: it shows diffs you can accept or reject line by line. The output is not text to read. It is a decision to make. The user stays in control throughout.</p><h4><strong>3. Summarisation</strong></h4><p><strong>What it unlocks:</strong> compressing large volumes of information so humans do not have to process all of it. Distilling hundreds of pages, thousands of reviews, or hours of conversation into the parts that matter.</p><p><strong>Example:</strong> contextual highlights and progressive disclosure on a set of reviews or user generated content. Surface the three things that matter. Let the user drill into detail on the ones they care about. The AI does not just compress information. It prioritises it, based on what the user is trying to do right now. A summary of a 50-page legal contract should look different for the CFO than for the head of legal.</p><h4><strong>4. Classification</strong></h4><p><strong>What it unlocks:</strong> sorting, routing, labelling, and triaging automatically. Categorising unstructured content accurately without hand-built rules for every scenario.</p><p><strong>Example:</strong> Linear&#8217;s Triage Intelligence. When a new issue comes in, the AI suggests a team, assignee, labels, and related issues. The suggestions are visible. Hover over them to see the reasoning. Accept with one click or override. The interesting thing about classification is that the lazy default is not a bad interface. It is no interface at all. Teams use classification behind the scenes but never surface it as something the user can see, trust, and refine. That is a missed opportunity. Making classification visible turns it into a feature users value.</p><h4><strong>5. Extraction</strong></h4><p><strong>What it unlocks:</strong> pulling structured data from unstructured input. Reliably extracting specific fields from messy documents, images, or conversations at scale.</p><p><strong>Example:</strong> auto-populated fields from uploaded documents. Upload an invoice, get line items extracted and pre-filled. Upload a resume, get candidate fields populated. The user reviews and corrects. They are editors, not transcribers. The shift from &#8220;type everything&#8221; to &#8220;verify and fix&#8221; is enormous. It turns a ten-minute task into a thirty-second review.</p><h4><strong>6. Memory</strong></h4><p><strong>What it unlocks</strong>: accumulating context over time and using it to make every other primitive smarter. A system with memory knows that this user always filters for pet-friendly hotels. It knows that this team routes billing bugs to the payments squad. It knows that this reader skips the methodology section.</p><p>Memory is the enabling layer. Intent detection with memory gets better at understanding what this specific user means. Classification with memory routes more accurately based on how similar items were handled before. Summarisation with memory knows what this person cares about and prioritises accordingly.</p><p><strong>Example</strong>: a support agent that already knows your order history, your past issues, your account tier. It does not ask you to explain your problem from scratch. It picks up where the last interaction left off. Every interaction makes the next one faster.</p><p>Without memory, every AI interaction starts from zero. With it, the product compounds. This is also why memory is the hardest primitive to copy. Your competitor can use the same model. They cannot replicate what your product has learned about your users.</p><p><em>These six primitives naturally combine.</em> Intent detection figures out what the user wants. Memory recalls what it already knows about them. Classification routes the request. Extraction pulls the relevant data. Summarisation compresses it. Generation produces the response. Orchestrate all six well and you get a conversation. That is why chat makes for impressive demos. It exercises every primitive at once. It is also why chat is the laziest default. You are using all six primitives to do a job that one or two of them could handle better.</p><h1><strong>The Agent Primitives</strong></h1><p>These are emerging. The UX patterns are still being invented.</p><h4><strong>1. Planning and Decomposition</strong></h4><p><strong>What it unlocks:</strong> the user states an outcome. The AI figures out how to get there. Breaking an ambiguous goal into concrete, executable steps.</p><p><strong>The lazy version:</strong> an agent that narrates its chain-of-thought in a chat window. &#8220;First, I will search for... Now I am analysing... Let me think about this...&#8221; The user watches the AI think out loud, step by step, for ninety seconds. This is performative reasoning. It looks impressive. It wastes the user&#8217;s time.</p><p><strong>The good version:</strong> a proposed plan the user can review, edit, and approve before execution starts. Show the steps. Let the user reorder, remove, or add to them. Then execute. The user wants to approve the plan, not watch the thinking.</p><h4><strong>2. Autonomous Execution</strong></h4><p><strong>What it unlocks:</strong> completing multi-step tasks without human involvement at every step. The AI does the work, not just the thinking.</p><p><strong>The lazy version:</strong> a chat-based agent that asks permission at every single step. &#8220;Should I do this? Okay. Should I do this next? Okay.&#8221; This turns autonomy into a slower version of doing it yourself. The whole point of an agent is that it handles the steps. If you have to approve each one, you are just operating a very slow remote control.</p><p><strong>The good version:</strong> background execution with confirmation checkpoints at meaningful decision points. The agent does the work. You approve the result. Not every intermediate step. Just the ones that matter.</p><h4><strong>3. Environment Interaction</strong></h4><p><strong>What it unlocks:</strong> the AI does not just suggest. It acts. It books, sends, files, updates. It reaches into external systems and changes things.</p><p><strong>The lazy version:</strong> &#8220;I have drafted an email, would you like me to send it?&#8221; Repeated for every individual action. Twelve confirmation dialogs for twelve actions. The friction of asking permission one action at a time destroys any efficiency the agent was supposed to create.</p><p><strong>The good version:</strong> batch actions with review, clear undo capability, and sandbox-then-commit patterns. This is where trust design matters most. The gap between &#8220;AI that suggests&#8221; and &#8220;AI that acts&#8221; requires UX that gives users confidence without demanding constant supervision. Show the user what will happen. Let them approve the batch. Make it reversible.</p><h4><strong>4. Monitoring and Reaction</strong></h4><p><strong>What it unlocks:</strong> the AI pays attention when you are not. Watching for conditions and acting proactively when they are met.</p><p><strong>The lazy version:</strong> notification spam. Alerting on everything because the system cannot distinguish signal from noise. Fifteen notifications a day, fourteen of which are irrelevant. The user learns to ignore all of them.</p><p><strong>The good version:</strong> smart triggers with context and action. &#8220;Your flight price dropped $200. I held the lower rate for 24 hours.&#8221; The notification is not just information. It is a completed action the user can confirm or undo. Event-driven, not attention-demanding. The AI earns trust by acting on the user&#8217;s behalf with clear, reversible actions.</p><p>The agent primitives will define the next generation of AI UX decisions. The lazy defaults are already forming. Mostly chat windows narrating agent behaviour. The teams that design better patterns now will own the category.</p><h1><strong>Smart Filters: Intent Detection Done Right</strong></h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!efi0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!efi0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 424w, https://substackcdn.com/image/fetch/$s_!efi0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 848w, https://substackcdn.com/image/fetch/$s_!efi0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 1272w, https://substackcdn.com/image/fetch/$s_!efi0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!efi0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png" width="1456" height="1020" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1020,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:728572,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/190265961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!efi0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 424w, https://substackcdn.com/image/fetch/$s_!efi0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 848w, https://substackcdn.com/image/fetch/$s_!efi0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 1272w, https://substackcdn.com/image/fetch/$s_!efi0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b7d9ade-3122-4c4f-a55c-1fc240cf8ae5_3048x2136.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The most successful AI feature I shipped at Booking.com has no chat, no generated text, and no disclaimer.</p><p>The experience: you are searching for hotels in Paris. On the search page, there is a Smart Filters text field. You type &#8220;hotels with pool and gym.&#8221; You tap &#8220;Find properties.&#8221; What comes back is not a paragraph of AI-generated recommendations. It is the existing filter UI. Checkboxes for Swimming Pool and Fitness Center, pre-selected. Property type set to Hotels. 42 results.</p><p>That is it. The AI understood what you meant and translated it into the structured actions the product already supports. Here is why it works:</p><p><strong>The AI does not replace the interface. It accelerates it.</strong> The filters already existed. Users already knew how to use them. The AI just fills them in faster than the user could manually. The user ends up in the exact same place, with the same controls, the same mental model. Just faster.</p><p><strong>The output is verifiable at a glance.</strong> Did it understand &#8220;pool&#8221;? Swimming Pool is checked. &#8220;Gym&#8221;? Fitness Center is checked. Verification takes one second. Compare that to a chatbot response: &#8220;Based on your preferences, I recommend these hotels...&#8221; followed by a generated list you have no way to audit for completeness or accuracy.</p><p><strong>It is editable without starting over.</strong> Wrong filter? Uncheck it. Missing one? Add it. The user stays in control without having to rephrase a question and hope the AI gets it right the second time. Compare this to the regeneration loop of a chatbot, where your only option when the output is wrong is to try again with different words.</p><p><strong>Failure is cheap.</strong> If the AI misinterprets something, the user adjusts a checkbox. No dead end. No trust collapse. No disclaimer needed, because the output is not generated content. It is a filter selection the user can see and fix instantly.</p><p><strong>It uses the right primitive.</strong> This is intent detection. The AI&#8217;s job is to understand what the user means and translate it to structured action. A chatbot would have used dialog management and generation. Those are the wrong primitives for this problem. The user does not want a conversation about hotels. They want 42 results that match their criteria.</p><p>The question that led to this feature was not &#8220;how do we add AI to search?&#8221; It was &#8220;users know what they want, but finding the right combination of filters from a list of 200+ takes too long. Can we solve that?&#8221; The primitive (intent detection) matched the problem (translating fuzzy preferences into structured filters). The UX followed from there.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/most-ai-ux-is-lazy?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Think your design team could use this?</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/most-ai-ux-is-lazy?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/most-ai-ux-is-lazy?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h1><strong>The Real Problem</strong></h1><p>The lazy UX problem is a lazy thinking problem. Teams jump from &#8220;we have AI&#8221; to &#8220;ship a chatbot&#8221; without asking what capability they have and what user problem it solves.</p><p>Eleven primitives across two tiers. Each one unlocks UX patterns that were previously impossible. Each one has a lazy default that teams reach for because it is easy, not because it is right.</p><p>The builders who win will match the primitive to the problem and design the interface that primitive deserves.</p><p>A text field that turns into checkboxes. Intent detection with the right UX. No chat. No generated paragraphs. No disclaimer. Just the existing interface, faster.</p><p>Start with the user problem. Identify the primitive. Design the UX it deserves.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>BuilderLab</strong> publishes practical strategy for people who ship AI products. If this was useful, subscribe to get the next one.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Growth Machines: The Cursor Story]]></title><description><![CDATA[Lessons in scale you can steal from the fastest growing SaaS ever]]></description><link>https://www.builderlab.ai/p/growth-machines-the-cursor-story</link><guid isPermaLink="false">https://www.builderlab.ai/p/growth-machines-the-cursor-story</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 01 Mar 2026 14:32:04 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a824e2e1-dcaa-4198-a8b8-d7f562d3d2e0_960x540.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Four MIT students fork a code editor in 2022. Two years later, they are running the fastest-growing SaaS company in history.</p><p>$1 million in revenue in 2023. $100 million in 2024. $1.2 billion in 2025. A valuation of $29.3 billion by November 2025. An 11x increase in twelve months.</p><p>Here is the detail that should bother every startup founder reading this: they spent zero dollars on marketing to reach $100M ARR. Not &#8220;low marketing spend.&#8221; Not &#8220;efficient growth.&#8221; Zero. If you are building a product right now and your growth plan starts with &#8220;demand gen,&#8221; this story is about to rearrange your priorities.</p><p>Cursor is the most instructive growth story in software since Notion. It made five specific product decisions that any builder can study, steal, and adapt, whether you are building an AI-native startup, a developer tool, or a productivity product in any category.</p><p>This is a teardown of those five decisions. What they chose, what they rejected, and what made each choice compound into the kind of growth that makes VCs rewrite their models.</p><div><hr></div><p>Michael Truell, Sualeh Asif, Arvid Lunnemark, and Aman Sanger met studying computer science and math at MIT. They collaborated at MIT CSAIL, MIT&#8217;s Computer Science and Artificial Intelligence Laboratory, before founding Anysphere in 2022..</p><p>Everyone was saying AI is the future of coding, but their bet was more specific and more interesting: the right place to embed AI is not on top of the code editor as a plugin. It is inside the code editor itself. That distinction, on top vs. inside, is the entire story of Cursor&#8217;s product.</p><p>They raised $3.38 billion in total funding from Accel, Coatue, Thrive Capital, Andreessen Horowitz, Google, and NVIDIA. Today, Cursor has over 1 million daily active users, 360,000 paying customers, and more than 50,000 enterprise accounts. It generates nearly a billion lines of code daily. Major customers include Stripe, OpenAI, and Spotify.</p><p>The numbers are staggering. The decisions behind them are useful.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I write about patterns like this every month, where product depth turns into strategic advantage.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Decision 1: Fork, Don&#8217;t Build</h1><p>The first decision Cursor made saved them two years. They forked VS Code. They took the open-source codebase of the most widely used code editor in the world and made it theirs.</p><p>This gave them three things instantly:</p><p><strong>Muscle memory</strong>. Every developer who uses VS Code (hundreds of millions) already knew how to navigate Cursor on day one. Keyboard shortcuts, extension ecosystem, settings, themes &#8212; all familiar. The switching cost was almost zero.</p><p><strong>Maturity</strong>. VS Code is decades of engineering work. The file system, the debugging tools, the terminal, the Git integration, and everything that makes it work,  Cursor did not have to build any of it. They inherited a production-grade foundation.</p><p><strong>Depth permission</strong>. A plugin sits on top of VS Code, it can suggest a line of code. It can autocomplete. But it cannot fundamentally change how the editor understands your project. A fork on the other hand can rewrite the way context flows between files. It can introduce Composer mode, a multi-file, project-aware AI that understands the relationships between your modules, your tests, and your config.</p><p>The lesson is clean: when a mature, open-source platform exists in your domain, forking it is not lazy, it is strategic. You skip the years of infrastructure work and go straight to the layer where your actual differentiation lives. Every month you spend building what already exists is a month your competitor spends building what doesn&#8217;t.</p><h1>Decision 2: Custom Models for Workflow, Foundation Models for Reasoning</h1><p>The second trap Cursor avoided was the &#8220;API wrapper&#8221; trap.</p><p>In 2023 and 2024, thousands of startups launched products that were thin interfaces around OpenAI or Anthropic API calls. The problem is, when the model provider ships a better default UI, your product becomes redundant overnight.</p><p>Cursor built custom AI models trained specifically for coding workflow tasks. Things like advanced autocomplete that understands multi-file context, predictive edits based on your recent changes, and code generation that respects your project&#8217;s patterns and conventions. These custom models handle the workflow-specific tasks. Foundation models like GPT and Claude handle the general reasoning.</p><p>This is a deliberate architectural split: proprietary models for the 80% of interactions that are workflow-specific, foundation models for the 20% that require open-ended reasoning.</p><p>By the time Cursor 2.0 shipped, it included a multi-agent architecture and a proprietary Composer model that could orchestrate complex, multi-step coding tasks across an entire project. This was not something you could replicate by calling an API.</p><p>If your only moat is a nice UI around someone else&#8217;s model, you do not have a moat. Build custom intelligence for the workflows that are specific to your domain. Use foundation models for everything else. The custom layer is where your product becomes irreplaceable.</p><h1>Decision 3: Serve Power Users First, Everyone Else Later</h1><p>GitHub Copilot has over 20 million users. Cursor has around 1 million. Cursor makes roughly four times the revenue.</p><p><strong>20x fewer users, 4x more revenue.</strong></p><p>How? Copilot optimized for the broadest possible audience. Single-line autocomplete suggestions works fine for casual code. It&#8217;s a great convenience tool.</p><p>Cursor optimized for professional developers doing the hardest work. Multi-file edits. Project-wide context. Agent-mode workflows that can scaffold entire features. Composer sessions that understand the relationships between your API routes, your database schema, and your frontend components.</p><p>When you serve power users first, three things happen:</p><p><strong>They pay more</strong>. Cursor Pro at $20/month and Business at $40/month convert at higher rates because the value proposition is clear: this tool does not help you write code a little faster. It fundamentally changes how much you can ship in a day.</p><p><strong>They evangelize harder</strong>. A casual user who saves 10 minutes a day tells nobody. A senior engineer who ships a feature in an afternoon that used to take a week tells everyone on their team. And their team tells other teams. This is how you get $100M ARR with $0 marketing spend.</p><p><strong>They are harder to displace</strong>. Once a power user has rewired their workflow around your tool, switching costs are enormous because of habit, muscle memory, and accumulated context.</p><p>By October 2025, Cursor had captured approximately 40% of all AI-assisted pull requests despite having a fraction of Copilot&#8217;s user base. The power users were producing the majority of the output.</p><p>The lesson is counterintuitive for founders raised on &#8220;total addressable market&#8221; thinking: a smaller market of high-intent users is more valuable than a larger market of casual users.</p><h1>Decision 4: Let the Product Be the Marketing</h1><p>Cursor&#8217;s growth was entirely product-led. No ads. No content marketing. No sales team (until enterprise). No growth hacks. The product itself was the distribution mechanism.</p><p>Here is how this worked:</p><p>A developer tries Cursor&#8217;s free tier (2,000 monthly completions). They experience the difference between single-line autocomplete (Copilot) and project-aware AI editing (Cursor). They upgrade to Pro. They mention it in a team standup. Their teammates try it. Their manager notices the team is shipping faster. The company starts a Business plan. That company&#8217;s engineers mention it at a conference. The loop repeats.</p><p>This is not a new idea. Slack did it. Figma did it. Notion did it. But Cursor executed it in a category (developer tools) where the word-of-mouth dynamics are unusually potent. Developers are opinionated about their tools. When one of them switches editors, everyone notices and everyone asks why.</p><p>The free tier was not a lead-gen funnel in disguise. It was genuinely useful. 2,000 completions per month is enough to get real work done. This matters. A free tier that feels crippled creates resentment. A free tier that feels generous creates evangelists.</p><p>Revenue doubled approximately every two months during the hypergrowth phase. That does not come from marketing campaigns. That comes from a product that makes people want to tell other people about it.</p><p>The principle for builders: if your product requires a sales pitch, it is not ready. The product should be the pitch. The users should be the sales team.</p><h1>Decision 5: Usage-Based Pricing for the AI Layer</h1><p>In August 2025, Cursor shifted to usage-based pricing for its agent features. The logic: the more the AI does for you, the more you pay. If Cursor&#8217;s agent writes 500 lines of code that would have taken you two hours, you pay proportionally more than when it autocompletes a function name. The price reflects the value delivered, not the time spent in the editor.</p><p>This matters for two reasons.</p><p>First, it makes the economics self-correcting. As AI capabilities improve, users get more done per session, which means they pay more per session, which means revenue grows with capability, not just with headcount.</p><p>Second, it signals confidence. A usage-based pricing model says: &#8220;We believe you will use this so much that you will be happy to pay for what you consume.&#8221; A flat-rate model says: &#8220;We hope you use this enough to justify the price.&#8221;</p><p>The pricing tiers: free (2,000 completions), Pro ($20/month), Business ($40/month), plus usage-based agent pricing create a natural upgrade path. You start free. You hit the limit. You upgrade. You start using agent features. You pay for what you consume. At every step, the value is obvious before the payment is required.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I break down the strategic decisions behind AI-native companies: what they chose, what they rejected, and why it compounds.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1>The Copilot Counterpoint</h1><p>What Happens When You Go Broad Instead of Deep</p><p>No teardown is useful without a counterexample. Copilot is the counterexample.</p><p>GitHub Copilot launched with Microsoft&#8217;s distribution muscle. It is bundled with GitHub. It is integrated into VS Code as a plugin. It has the backing of OpenAI. It has over 20 million users and penetration in 90% of the Fortune 100. It generates roughly $300 million in ARR. Cursor, with 5% of the user base, generates $1.2 billion.</p><p>Why? Because Copilot is constrained by the plugin architecture. A VS Code plugin cannot do what a VS Code fork can. It cannot rewrite how the editor processes project context. It cannot introduce multi-agent workflows that operate across files. It cannot build Composer mode. The plugin model inherently limits how deeply AI can integrate with the development workflow. Copilot optimized for breadth: every developer, every use case, every company.</p><h1>The Moat Question: Is This Sustainable?</h1><p>Intellectual honesty demands we address this. Cursor&#8217;s moat is real, but it is thinner than the valuation implies.</p><p>Three threats worth watching:</p><p><strong>Microsoft&#8217;s bundling power</strong>. GitHub Copilot is already included in enterprise contracts. When your competitor can offer &#8220;free with GitHub Enterprise,&#8221; you need your product to be dramatically better, not marginally better. So far, Cursor clears that bar. But the bar moves every quarter.</p><p><strong>Agent platforms targeting non-developers</strong>. Tools like v0 and Bolt.new are not competing with Cursor for professional developers. They are competing for the &#8220;vibe coder&#8221;, the non-developer who can now build software with AI assistance. If that market becomes larger than the professional developer market, Cursor&#8217;s power-user strategy becomes a ceiling, not an advantage.</p><p><strong>Model providers building vertically</strong>. OpenAI, Google, and Anthropic all have incentives to build their own coding environments. They have direct model integration advantages that third parties cannot match. If Claude or GPT ships a native IDE experience, Cursor&#8217;s custom model layer becomes less differentiating.</p><p>The current valuation of $29.3 billion sits at roughly 58x ARR. That prices in sustained hypergrowth. If any of these three threats materializes faster than Cursor&#8217;s product evolves, that multiple compresses quickly.</p><p>There is also a sentiment signal worth noting: developer favorability toward AI coding tools dropped from over 70% in 2023-2024 to 60% in 2025, even as adoption rates rose to 91%. Developers are using these tools more but trusting them less.</p><div><hr></div><h1>The Fork-and-Deepen Playbook: What You Can Steal</h1><p>Strip away the specifics of code editors and AI models. What is the replicable playbook?</p><p><strong>At 0 to 1</strong>:  Fork, don&#8217;t build. Find a mature, open-source or widely used platform in your domain. Fork it or build deeply on top of it. Skip the infrastructure years. Go straight to the differentiation layer. Your users get instant familiarity. You get instant focus on what matters.</p><p><strong>At 1 to 10</strong>: Serve the 10x user. Do not build for the average user. Build for the most demanding user in your category. The user who will restructure their workflow around your product if it is good enough. These users pay more, evangelize more, and are harder to displace. They are your growth engine.</p><p><strong>At 10 to 100</strong>: Let the product be the distribution. If you need a sales deck to explain your product&#8217;s value, the product is not ready. Build until the experience is so obviously better that users become your marketing department. Free tiers should be genuinely useful, not crippled demos.</p><p><strong>At every stage</strong>: Build custom intelligence for your domain. Do not be an API wrapper. Foundation models are commodities. Your custom models trained on your domain&#8217;s specific workflows are your advantage.</p><p><strong>When scaling:</strong> Align pricing with value delivery. Usage-based pricing for AI features is a signal of confidence in your own product. If your AI delivers real value, you should be willing to price it based on how much value it delivers, not on a flat subscription that averages out the power users with the casual ones.</p><h1>Where you should exercise caution</h1><p>One caveat worth naming: playbooks extracted from hypergrowth stories are seductive precisely because the outcomes are so clean. Cursor&#8217;s decisions look inevitable in retrospect, but each one carried real risk that happened to break their way. </p><p>Forking creates a dependency on an upstream project you do not control. </p><p>Serving power users first means saying no to revenue from a larger, easier market while you burn cash. </p><p>Product-led growth with zero marketing spend works until it doesn&#8217;t, and you have no distribution muscle to fall back on. </p><p>Usage-based pricing can suppress adoption in cost-sensitive segments. </p><p>The value of this piece is in the underlying logic: choose depth over breadth, align your architecture with your differentiation, and price to the value you deliver. When the logic fits your context, the decisions follow naturally. When it doesn&#8217;t, copying the moves will hurt you.</p><h1>The Number That Matters Most</h1><p>Here is the data point I keep coming back to: Copilot has 20 million users. Cursor has 1 million. Cursor makes 4x the revenue. Depth beats breadth. Power users beat mass users, integration depth beats surface-level convenience, a fork beats a plugin.</p><p>Every founder I talk to worries about total addressable market. The Cursor story says the opposite: worry about depth of value per user. The market follows the product.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/growth-machines-the-cursor-story?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Useful? Pass it on.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/growth-machines-the-cursor-story?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/growth-machines-the-cursor-story?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div>]]></content:encoded></item><item><title><![CDATA[Your Product Now Has Two Users]]></title><description><![CDATA[And one of them is AI]]></description><link>https://www.builderlab.ai/p/your-product-now-has-two-users</link><guid isPermaLink="false">https://www.builderlab.ai/p/your-product-now-has-two-users</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 22 Feb 2026 18:50:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/737398a9-b35f-462d-9cf2-555631f55486_2112x1188.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Julie wants running shoes for wet pavement. Two years ago, she would have typed &#8220;best waterproof running shoes&#8221; into Google, clicked through ten blue links, bounced between review sites, and eventually bought something from the brand that ranked highest and had the most convincing landing page.</p><p>Today she asks ChatGPT. It recommends three shoes. Your product is not one of them.</p><p>Your shoe isn&#8217;t bad. It&#8217;s not that expensive either. It&#8217;s just invisible. The AI intermediary that now sits between your product and your customer could not understand what your product does, who it is for, and where it falls short.</p><p>Welcome to the era of AI-mediated discovery. Your product now has two users. The human who uses it, and the AI that decides whether the human ever sees it.</p><h1>The Numbers That Should Scare You</h1><p>This is not a hypothetical future. It is already here, and the data is brutal.</p><p>Over 80% of all searches now end without a single click. When a Google AI Overview appears, that zero-click rate climbs to 83%. Click-through rates drop from 15% to 8% when an AI Overview is present, a 58% reduction. The click, once the atomic unit of digital discovery, is becoming an endangered species.</p><p>But that data isn&#8217;t indicative alone. Sure, information seekers clicking on fewer links adds noise. Not everyone is looking for product recommendations after all. Some of us just want to know if Koalas are vegan. So here&#8217;s some more well rounded data that proves my point:</p><p>SOCi analyzed over 350,000 locations across 2,751 multi-location brands and found that ChatGPT recommends just 1.2% of them. Google&#8217;s local three-pack shows 35.9%. That makes AI recommendation roughly 30 times more selective than traditional search.</p><p>Read that again: 30 times more selective. If traditional search was a funnel, AI-mediated discovery is a needle.</p><p>Meanwhile, the traffic that does come through AI channels is surging. AI-driven traffic to US ecommerce sites increased 758% year over year by November 2025. According to IBM&#8217;s study with NRF, 45% of consumers now turn to AI during buying journeys: 41% to research products, 33% to interpret reviews, 31% to hunt for deals. About a third of US consumers say they would let an AI make purchases for them entirely.</p><p>This is not a trend that might matter someday. This is the new distribution channel. And your product is either legible to it or invisible within it.</p><h1>The Old Playbook Is Broken</h1><p>For two decades, the discovery playbook was straightforward. Rank for keywords. Drive traffic. Convert on-site. The entire machine was built on the assumption that discovery happens through a list of links, and the winner is whoever sits highest on that list.</p><p>AI-mediated discovery breaks every piece of that machine.</p><p>Non-brand, awareness-driven B2B traffic has declined by up to 60% as AI search reduced click-through behavior. These companies watched something deeply disorienting happen: their search rankings stayed stable, but the clicks disappeared. They were still &#8220;winning&#8221; at SEO. The traffic just stopped arriving. The game changed around them while the scoreboard kept showing the old score.</p><p>When a user asks an AI a question, the AI does not return a list of links for the user to evaluate. It returns an answer. It synthesizes, filters, compares, and recommends. This is the fundamental shift: your product is no longer competing for a user&#8217;s attention. It is competing for an AI&#8217;s comprehension.</p><h1>What AI Intermediaries Actually Reward</h1><p>The brands ChatGPT recommends are not the biggest. They are the most legible.</p><p>Recommended locations average 4.3-star ratings. AI platforms heavily weight sentiment. Not volume of reviews, but consistency of positive signal. More importantly, AI rewards cross-channel consistency over strength in a single channel. A brand that is excellent on its website but thin on Google, Yelp, and social profiles performs worse than a brand that is consistently good across all of them.</p><p>This is a fundamentally different optimization target than traditional SEO. Search engine optimization rewarded you for being strong in one place: Google&#8217;s index. AI-mediated discovery rewards you for being coherent everywhere.</p><p>The second pattern is even more counterintuitive. Constraint-focused product content outperforms feature-focused content for AI recommendations. &#8220;Best for daily commuters; water-resistant for light rain, not heavy downpours&#8221; beats &#8220;advanced waterproof technology with moisture-wicking lining.&#8221; The AI intermediary is not looking for marketing language. It is looking for specificity about what your product does, who it serves, and, critically, where it does not work.</p><p>Think about why. When a user asks an AI for a recommendation, the AI is trying to match a specific need to a specific product. Feature lists are noise. Constraints are signal. The AI needs to know what your product is not good at just as much as what it is good at, because that is how it avoids recommending the wrong thing to the wrong user.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I write about these structural shifts every so often at BuilderLab for product leaders navigating AI-native markets.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The Second-Order Consequences</h1><p>This shift does not stop at &#8220;optimize your product data.&#8221; The incentives underneath AI-mediated discovery are different from search, and that difference compounds.</p><p>Google monetized clicks. Large language models monetize sessions. Their economic incentive is to keep users inside the interface, deliver a high-confidence answer, and reduce friction. That means fewer options, higher precision, and lower cognitive load. A list of ten links was acceptable when the click was the business model. A single synthesized answer is required when retention is.</p><p>Right now, most AI interfaces recommend three to four products. In many categories, eight to twelve products could legitimately satisfy the user&#8217;s constraints. The compression is already severe. When sponsored placements enter the system, surface area shrinks further. If one slot is paid and one is reserved for a trusted integration, you are effectively competing for one or two organic positions. This is not SEO with better copy. It is a high-precision, low-inventory discovery economy.</p><p>Here, discoverability becomes winner-takes-most. Marginally legible products do not receive marginal traffic. They disappear. From the AI provider&#8217;s perspective, monetization must remain tightly coupled to relevance. If AI platforms degrade recommendation quality for ad revenue, users leave. Unlike search engines, they cannot fall back on ten blue links. Trust is the product.</p><p>There is another layer. Today, structured constraint data is an advantage. Tomorrow, it becomes table stakes. When every brand expresses &#8220;best for daily commuters, not ideal for heavy rain&#8221; in clean, machine-readable form, differentiation moves elsewhere. The AI must choose among equally legible products.</p><p>What does it optimize for then? Likely the lowest probability of user regret. Consistent satisfaction signals. Return rates. Review stability. Cross-channel coherence. In that world, operational excellence beats marketing brilliance. The brands that win are not those with the loudest features, but those with the most reliable outcomes.</p><h1>The Trust Layer Underneath</h1><p>Before you rush to &#8220;optimize for AI,&#8221; there is a critical nuance.</p><p>Only 46% of shoppers fully trust AI recommendations. And 89% still check information before buying based on an AI recommendation. The AI intermediary is not a decision-maker. It is a filter. It narrows the consideration set from everything to three or four options. Then the human takes over.</p><p>This means the AI is deciding who gets into the room. But the human is still deciding who wins.</p><p>Your product needs to pass two tests now, in sequence. First, can the AI understand it well enough to recommend it? Second, once the human arrives (from the AI recommendation), does the product experience close the deal?</p><h1>What This Means for Builders</h1><p>The implications are immediate. Your product&#8217;s discoverability is no longer just a marketing problem. It is a product data problem. How your product describes itself in its attributes, constraints, use cases, and limitations, determines whether it passes through the AI filter.</p><p>Does your product data describe constraints and use cases, or does it describe features and benefits? Does your product information maintain consistency across every surface it appears on, or is it fragmented across channels with different messaging on each?</p><p>If you are a startup founder, there is a genuine opportunity here. AI recommendation is more meritocratic than traditional search in one specific way: it rewards data density and accuracy over brand size and ad spend. A small brand with thorough, constraint-focused product data can outperform a Fortune 500 brand with thin product pages. You do not need a massive SEO team or a big advertising budget. You need clear, specific, honest product data that an AI can parse and match to user intent.</p><p>Here is the practical checklist.</p><p><strong>Audit your product data for AI legibility.</strong> Take your top 10 product pages. Ask ChatGPT, Gemini, and Perplexity to recommend a product in your category for a specific use case. Are you in the results? If not, look at who is and study their product data structure.</p><p><strong>Shift from features to constraints.</strong> Rewrite your product descriptions to explicitly state who the product is for, what specific problems it solves, and where it falls short. &#8220;Works best for X. Not ideal for Y.&#8221; This feels counterintuitive. Why would you highlight limitations? Because the AI intermediary uses limitations to make accurate recommendations, and accurate recommendations build trust in the AI&#8217;s suggestions.</p><p><strong>Ensure cross-channel consistency.</strong> Your product data should tell the same story on your website, Google Business Profile, Amazon listing, social profiles, and review platforms. The AI is synthesizing across all of them.</p><p><strong>Monitor AI recommendations, not just search rankings.</strong> Your weekly check should now include &#8220;what does ChatGPT recommend for [your category] + [specific use case]?&#8221; Track this the way you track keyword rankings. Build it into your product analytics workflow.</p><p><strong>Design your product information architecture for machines, not just humans.</strong> Structured data, consistent taxonomy, explicit attribute-value pairs. The prettier your product page is for humans, the less useful it may be for AI intermediaries if the underlying data is trapped in marketing copy rather than structured fields.</p><h1>The Opportunity Behind the Disruption</h1><p>Retailers who have already adapted to this shift are seeing results. Those adopting conversational, intent-driven search internally reported a 35% drop in zero-result queries and a lift in first-click conversion. Companies implementing AI-powered discovery report conversion lifts of 20-40%.</p><p>Major retailers like Walmart, Target, and Etsy are integrating directly with ChatGPT, Gemini, and Copilot, allowing purchases within AI conversations. The transaction layer is moving into the AI intermediary&#8217;s environment. If your product is legible to these systems, you are not just getting discovered, you are getting purchased without the user ever visiting your site.</p><p>The irony is beautiful: in a world of AI intermediaries, the competitive advantage is radical clarity about your product.</p><h1>What Comes Next</h1><p>We are still early. The AI-mediated discovery layer is clumsy, inconsistent, and sometimes wrong. The 46% trust number tells you that consumers know this. The 89% verification rate tells you humans are not blindly following AI recommendations.</p><p>But the behavioral shift is locked in. The zero-click trend, the 758% traffic surge, the one-third of consumers willing to let AI buy for them. These are not numbers that reverse. The intermediary layer between your product and your user is only going to get thicker, more capable, and more selective.</p><p>The products that start designing for two users now, the human who uses the product and the AI that recommends it, will have a compounding advantage. Not because they cracked some new marketing hack. Because they did the hardest, most boring, most valuable thing in product development.</p><p>They made their product easy to understand.</p><p>That running shoe for wet pavement? The one the AI did not recommend? The product was fine. The data was not. Same shoe. Same quality. Same reviews. But one had structured, constraint-specific product data that an AI could parse, and the other had a beautiful landing page full of marketing copy that a human loved and an AI could not decode.</p><p>Your product now has two users. Design for both of them, or the one that matters most will never know you exist.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you&#8217;re building products in a world of AI intermediaries, I write about the structural shifts shaping distribution, monetization, and product design.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>PS: My most favorite electronics e-commerce brand is CoolBlue. I&#8217;ve bought all my laptops and phones from them, all my home appliances, even an entire kitchen. Most of it online. They&#8217;ve been doing the Pros and Cons list (and a fantastic product page) for ages.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5zat!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5zat!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 424w, https://substackcdn.com/image/fetch/$s_!5zat!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 848w, https://substackcdn.com/image/fetch/$s_!5zat!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 1272w, https://substackcdn.com/image/fetch/$s_!5zat!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5zat!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png" width="698" height="880" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:880,&quot;width&quot;:698,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:165097,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/188819522?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5zat!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 424w, https://substackcdn.com/image/fetch/$s_!5zat!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 848w, https://substackcdn.com/image/fetch/$s_!5zat!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 1272w, https://substackcdn.com/image/fetch/$s_!5zat!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdae6ef72-d01f-47e1-8a0f-049bbe99e76d_698x880.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[ROI or Die.]]></title><description><![CDATA[Why the AI accountability gap is the most expensive problem in tech right now.]]></description><link>https://www.builderlab.ai/p/roi-or-die</link><guid isPermaLink="false">https://www.builderlab.ai/p/roi-or-die</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 08 Feb 2026 06:02:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4ba7d3b3-ee8e-4d2d-b5c0-fc2436546bc9_960x540.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In 2024, corporations spent $252 billion on AI. That same year, several studies found that the majority of enterprise AI pilots delivered zero measurable impact on the P&amp;L. Not &#8220;underperformed expectations.&#8221; Zero. No measurable bottom-line impact whatsoever.</p><p>By 2025, 42% of companies had abandoned most of their AI initiatives, up from 17% the year before. The average organization now scraps nearly half its AI proof-of-concepts before they reach production. Gartner predicted that at least 30% of generative AI projects would be abandoned after proof-of-concept by end of 2025. RAND Corporation found that AI projects fail at twice the rate of non-AI IT projects.</p><p>Read those numbers again. A quarter-trillion dollars deployed. Ninety-five percent failure rate.</p><p>And yet, at the same time, others are making AI work. Decisively.</p><p>AI-powered personalization is lifting ecommerce conversion rates by up to 23%. Contact centers using AI see 30% operational cost reductions. A Harvard Business School study found AI users completed tasks 25% faster at 40% higher quality. Early GenAI adopters report $3.70 in value for every dollar invested, with top performers hitting $10.30 per dollar. McKinsey found that the 6% of organizations they classify as &#8220;AI high performers&#8221; attribute over 5% of their entire EBIT to AI.</p><p>Developers using GitHub Copilot code 55% faster in controlled studies. 88% of accepted Copilot suggestions stay in the codebase. Teams across the software development lifecycle report 15%+ velocity gains from AI tooling, and coding has become a $4 billion AI category in 2025, up from $550 million the year before. That&#8217;s not a pilot. That&#8217;s a market.</p><p>Same technology. Same year. Same models. Radically different outcomes.</p><p>So what separates the 5% that extract millions in value from the 95% that produce nothing? It&#8217;s not the model. It&#8217;s talent, it&#8217;s data, and&#8230;</p><p>It&#8217;s accountability. And the lack of it has a name.</p><div><hr></div><h1><strong>The Lie</strong></h1><p>Two words: &#8220;strategic investment.&#8221;</p><p>This is how AI projects avoid accountability. At every altitude. In every organization. And it&#8217;s the most expensive phrase in corporate vocabulary right now.</p><p>Think about something. If a VP of engineering walked into a budget review and said &#8220;I need $3 million for a new microservices architecture, no measurable business outcome expected for 18 months, and I can&#8217;t tell you what metric it will move&#8221;, that request would be dead before lunch. The CFO would laugh. The board would never see it.</p><p>Now replace &#8220;microservices architecture&#8221; with &#8220;AI.&#8221; Same request. Same missing numbers. Different outcome. The room nods. The budget gets approved. The CFO might even smile. &#8220;We need to be investing in AI.&#8221;</p><p>This is the exemption. And it operates at every altitude.</p><p>In the boardroom, the pattern is remarkably consistent. An investor sits on four boards. Every board meeting, the CEO presents the AI roadmap. She asks about ROI. The CEO says &#8220;it&#8217;s a strategic investment, we&#8217;ll see returns in phase two.&#8221; She doesn&#8217;t push harder because she doesn&#8217;t want to be the partner who &#8220;doesn&#8217;t get AI.&#8221; Multiply this by every firm, every fund, every seat in the portfolio. That&#8217;s how billions get deployed without a number attached.</p><p>One level down, in the executive suite, the pattern mutates but survives. A VP of Product has four AI initiatives running. One is a zombie with no measurable impact, no clear path to impact, but it has executive sponsorship and &#8220;strategic&#8221; status. Killing it means a political fight with the exec who championed it. So the zombie lives, quietly consuming budget and engineering time that should be flowing to the projects that are actually working. He&#8217;s subsidizing failure with the budget meant for success. He knows it. His team knows it. Nobody says it.</p><p>On the product team, a PM is sitting in a review. Someone asks about ROI for the AI feature. The project lead deploys the magic words: &#8220;it&#8217;s a strategic investment.&#8221; Everyone nods. The question dies. The project lives another quarter. The PM notices the pattern over time: projects with numbers get funded aggressively or killed fast. Decisions happen. &#8220;Strategic&#8221; projects just linger, never killed, never funded properly, never held accountable. They exist in a permanent state of limbo, consuming resources and producing nothing.</p><p>Meanwhile, a solo founder with eleven months of runway doesn&#8217;t get to play &#8220;strategic.&#8221; Every dollar spent on AI that doesn&#8217;t convert to revenue, retention, or reduced costs is a dollar closer to death. But he&#8217;s watching competitors raise $20 million rounds on demos and narrative. AI-native. Foundation model. Platform play. Zero revenue. He&#8217;s wondering if he&#8217;s doing it wrong. He&#8217;s not. He&#8217;s the one doing it right. He just doesn&#8217;t know it yet.</p><p><strong>&#8220;Strategic investment&#8221; is not a strategy. It&#8217;s a hiding place.</strong> And its getting crowded.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This is one of my core pieces for AI Strategy. Subscribe to get the full series including the ROI Framework.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1><strong>You Already Know How to Do This</strong></h1><p>Here&#8217;s what makes the AI accountability gap so maddening: it&#8217;s not a knowledge problem. Every organization already has the tools to prevent it.</p><p>You don&#8217;t ship a feature without success criteria. You don&#8217;t approve headcount without a business case. You don&#8217;t launch a product without knowing what metric it&#8217;s supposed to move. You don&#8217;t hand an engineering team six months and a blank check and say &#8220;surprise me.&#8221; You run quarterly business reviews. You set OKRs. You kill projects that miss milestones. You&#8217;ve been doing this for decades.</p><p>Except with AI. With AI, you do all of the things you&#8217;d never do with any other investment. You fund without milestones. You launch without success criteria. You give &#8220;strategic&#8221; projects unlimited runway. And then you&#8217;re shocked that 95% of pilots produce nothing.</p><p>IBM spent $5 billion building Watson Health. Five billion. They acquired four health data companies, employed 7,000 people, and partnered with MD Anderson Cancer Center, which spent $62 million over four years on a Watson-powered oncology advisor. The system was never used on a single actual patient. MD Anderson killed the project citing cost overruns, delays, and procurement problems. IBM eventually sold Watson Health to a private equity firm for roughly $1 billion. A $4 billion write-down on a &#8220;strategic investment&#8221; that nobody held accountable until it was too late.</p><p>Zillow trusted its algorithm to buy homes. Their Zestimate model had been valuing homes for years. Good enough for browsing, not good enough for betting the company. But they bet the company anyway. In Q3 2021, Zillow&#8217;s iBuying division lost $304 million. Total write-downs exceeded $500 million. They laid off 2,000 people, 25% of their workforce. Their stock lost $9 billion in market cap. Their competitors, Opendoor and Offerpad, used similar models but with tighter risk controls and shorter feedback loops. They survived. Zillow didn&#8217;t.</p><p>The technology worked in both cases. The accountability didn&#8217;t. Normal investment discipline; milestones, kill criteria, measurable outcomes, would have caught the problem long before it became catastrophic. But the rules didn&#8217;t apply. Because AI.</p><p>Here&#8217;s the double standard made concrete. Zillow&#8217;s competitors, Opendoor and Offerpad, used similar pricing algorithms in the same markets during the same period. Opendoor reported $170 million in profits and 7.3% gross margins in the same quarter Zillow lost $304 million. The difference wasn&#8217;t the model. Opendoor had been in the iBuying business since 2014 and had built tighter risk controls, shorter feedback loops, and a healthy respect for what the algorithm could and couldn&#8217;t predict. They treated the algorithm as a tool with known limitations. Zillow treated it as an oracle and scaled aggressively on faith. Same technology, same market, but opposite discipline, opposite outcome.</p><p>IBM&#8217;s story is even more instructive. Watson Health wasn&#8217;t a bad idea. It was a badly managed investment. The MD Anderson project&#8217;s original budget was $2.4 million. It ballooned to $62 million over four years. A University of Texas audit found that contracts were deliberately structured just below the threshold for board approval, avoiding the very scrutiny that would have caught the problem. The system was built on a legacy medical records platform that MD Anderson had already replaced. Nobody stopped to ask whether the integration made sense. Nobody set a milestone that would have surfaced the misalignment. The project drifted for years under &#8220;strategic&#8221; cover until the audit blew it up. If someone had asked &#8220;what&#8217;s the dollar, and when do we see it?&#8221; at month six, the $62 million loss could have been a $5 million lesson.</p><p>Now look at the winners. McKinsey&#8217;s 2025 AI survey tested 25 organizational attributes to find what predicts bottom-line impact from AI. The single strongest predictor was <strong>workflow redesign with clear milestones before scaling.</strong> The 6% of organizations seeing real EBIT impact are three times more likely to have redesigned workflows and set measurable targets before they deployed. They didn&#8217;t have better technology. They had better questions. They asked &#8220;what&#8217;s the dollar?&#8221; before they wrote a line of code.</p><p>BCG&#8217;s 2025 research on AI in finance tells the same story from a different angle. The highest-ROI teams don&#8217;t just build AI, they focus relentlessly on value from day one. They prioritize quick wins over open-ended learning. They allocate dedicated budgets. They emphasize early impact. BCG found that emphasizing early impact increases the likelihood of AI project success by six percentage points. In a landscape where the baseline success rate is in the single digits, that&#8217;s enormous. The pattern is consistent: discipline first, technology second.</p><p>The discipline already exists. You just stopped applying it when someone attached the word &#8220;AI&#8221; to the budget request. Stop making the exception.</p><div><hr></div><h1><strong>The Framework</strong></h1><p>&#8220;ROI or die&#8221;. Here&#8217;s a reminder to use the framework you already have.</p><p>Before any AI initiative gets funded, run it through a test of value. Use the framework you use for everything else. I use PROVE-IT (<a href="https://www.builderlab.ai/p/how-to-spot-the-10-of-ai-projects">link here.</a>) If it doesn&#8217;t pass, it doesn&#8217;t get built. If it does, it enters the cycle.</p><p><strong>Explore &#8594; Exploit &#8594; Platformise.</strong></p><p><strong>Explore [20% of AI budget]. </strong>Tiny projects. Small segments. Limited exposure. Fast and cheap. The goal isn&#8217;t scale, it&#8217;s signal. Did we find a dollar worth chasing? Run the experiment in weeks, not months. Most things die here. That&#8217;s not failure, that&#8217;s the system working. The ROI of Explore is learning speed. The faster you can run cheap experiments and kill the losers, the faster you find the winners. A team running four small experiments a month will outperform a team running one big &#8220;strategic&#8221; initiative every quarter. Every time.</p><p><strong>Exploit [50% of AI budget]. </strong>This is where the money is made. The winners from Explore get scaled. Broad rollout. Full integration into existing funnels. Conversion optimization, support deflection, pricing, ops efficiency, workflow acceleration. AI applied to the places you already make or spend money. ROI-positive within 90 days or kill it. No exceptions. No &#8220;phase two.&#8221; No &#8220;strategic.&#8221; Show the dollar or free up the budget for something that will. This track funds everything else. If Exploit isn&#8217;t performing, you don&#8217;t have permission to explore or invest long-term. You have to earn the right to innovate by proving you can execute.</p><p><strong>Platformise [30% of AI budget]. </strong>The proven wins from Exploit get built into infrastructure. Distilled capabilities. Reusable systems. Self-service tools that make the next cycle of Explore faster and cheaper. This is where data flywheels live, where memory systems compound, where proprietary feedback loops create defensibility. Platformise still has ROI, it&#8217;s just measured in compounding velocity rather than quarterly P&amp;L.</p><p><strong>The cycle is the point. </strong>Explore feeds Exploit. Kill fast. Only the winners graduate. Exploit feeds Platformise. Double down. The proven wins become infrastructure. And Platformise feeds back into Explore. Better platform means faster, cheaper experiments. Each revolution of the flywheel gets faster than the last.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J_i5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J_i5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 424w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 848w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J_i5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png" width="1456" height="753" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:753,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:177609,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/187263977?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!J_i5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 424w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 848w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!J_i5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4594b42f-064a-4019-af2a-669afdb4505b_2080x1076.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>This is also how you prepare for the future you can&#8217;t predict. </strong>When agents arrived, teams that had been exploring weekly had the muscle to spin up an experiment in days. Teams that had put everything into Exploit were six months behind before they started. And the teams that had platformised their AI capabilities could plug new paradigms into existing infrastructure instead of building from scratch. The framework doesn&#8217;t predict the future, but it makes you fast enough that you don&#8217;t need to.</p><h2><strong>A concrete example</strong></h2><p>Take <strong>AI-powered support deflection</strong>. This is a classic use case because it touches real cost, has fast feedback loops, and fails loudly when it doesn&#8217;t work. Perfect for disciplined AI investment.</p><p><strong>Explore</strong></p><p>You don&#8217;t start by &#8220;reimagining support.&#8221; You start by isolating one narrow problem.</p><p>Pick a single issue category, say, password resets or order confirmation questions. These are high-volume, low-risk, and well understood by agents. You deploy an LLM in a constrained way: either drafting responses for agents or deflecting tickets to self service. No full automation, no broad rollout.</p><p>You cap exposure at ~5% of traffic. You run it for two weeks. What you measure is boring and explicit:</p><ul><li><p>Resolution rate vs baseline</p></li><li><p>CSAT delta</p></li><li><p>Cost per ticket</p></li><li><p>Agent intervention rate</p></li><li><p>Escalation frequency</p></li></ul><p>You&#8217;re not trying to prove AI is &#8220;the future of support.&#8221; You&#8217;re asking a much simpler question: <em>is there a dollar here worth chasing</em>?</p><p>Most experiments die at this stage. Maybe the model hallucinates edge cases. Maybe CSAT drops. Maybe agents hate it. That&#8217;s fine. You&#8217;ve spent weeks, not quarters, and thousands, not millions. If there&#8217;s no clear signal, you stop.</p><p><strong>Exploit</strong></p><p>Now assume the signal is real. The experiment shows a 12% reduction in tickets for that category, with no CSAT degradation and lower average handling time. That&#8217;s no longer a research question. It&#8217;s a business one.</p><p>You move into Exploit. You expand coverage to the top 5&#8211;10 issue categories that share similar characteristics. You integrate the capability directly into the help flow instead of running it as a sidecar. You invest in guardrails, fallback logic, and operational monitoring. You staff it properly.</p><p>And now the bar changes.</p><p>This phase has explicit expectations:</p><ul><li><p>Weekly cost savings tracked in dollars</p></li><li><p>Clear ownership</p></li><li><p>A defined 60&#8211;90 day window to prove the numbers hold at scale</p></li></ul><p>If the deflection rate collapses, or CSAT starts to drift, you don&#8217;t debate intent. You stop, surgically fix issues, rerun, or roll back. The project doesn&#8217;t get infinite runway just because it once worked in a pilot.</p><p>This is the critical shift most teams skip. They treat pilots as proof of inevitability rather than proof of value.</p><p>If Exploit works, this becomes one of the few AI projects that actually funds others.</p><p><strong>Platformise</strong></p><p>Only now does platform thinking start to make sense. You still don&#8217;t build a platform off a single use case. You look for <strong>reuse pressure</strong>.</p><p>Your support deflection product succeeds. Separately, a sales assist product shows promise. A fraud triage model is running into similar confidence and fallback questions. A content moderation workflow needs quality thresholds.</p><p>Now you have three or four production AI systems asking the same questions:</p><ul><li><p>How do we measure quality consistently?</p></li><li><p>How do we compare model versions safely?</p></li><li><p>How do we incorporate human feedback?</p></li><li><p>How do we decide when automation is allowed vs deferred?</p></li></ul><p>Only at that point do you invest in shared evaluation tooling: common metrics, logging standards, review workflows, offline test harnesses. Not because &#8220;AI needs a platform,&#8221; but because multiple profitable systems are being slowed down by the lack of one.</p><p>Platformisation exists to <strong>compress future Explore cycles</strong>, not to justify past ones.</p><p>The result is compounding speed. The next support experiment launches in days instead of weeks. The next workflow doesn&#8217;t need to reinvent evaluation logic. This discipline creates leverage.</p><div><hr></div><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/roi-or-die?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Know someone running AI with no kill criteria? Send them this.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/roi-or-die?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/roi-or-die?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div><hr></div><h2><strong>The Rules</strong></h2><p><strong>Exploit funds Explore and Platformise. </strong>No Exploit performance, no permission to explore or invest long. This is how you answer the question &#8220;what are we getting for this?&#8221;</p><p><strong>Every project has a track. </strong>Untracked AI work doesn&#8217;t get funded. If someone can&#8217;t tell you whether their project is Explore, Exploit, or Platformise, they haven&#8217;t thought hard enough.</p><p><strong>Track migrations are explicit decisions, not drift. </strong>Projects get promoted, killed, or demoted, never &#8220;let&#8217;s give it one more quarter.&#8221; That sentence has killed more AI projects than bad data ever has.</p><p><strong>The 20/50/30 is representative, not a target. </strong>Unspent exploration budget is a feature, not a failure. And the ratios flex by stage. A seed-stage startup might be 80% Exploit, 20% Explore, 0% Platformise, and that&#8217;s exactly right. An enterprise with $80 million in ARR might run the full 20/50/30. The principle is constant, but the allocation can adapt.</p><p><strong>Rebalance quarterly. </strong>Review the full portfolio. Which Explore experiments found signal? Which Exploit projects hit their 90-day milestones? Which Platformise investments are actually compounding velocity? Move budget toward what&#8217;s working. Kill what&#8217;s not. This is the portfolio discipline you already apply to every other investment. Apply it here.</p><div><hr></div><h1><strong>The Transition</strong></h1><p>If you&#8217;re reading this and thinking &#8220;we already have AI projects running with no tracks, no milestones, and no kill criteria&#8221;, you&#8217;re not alone. That&#8217;s the default state of most organizations right now.</p><p>You can&#8217;t walk in Monday and announce that everything is changing. Here&#8217;s the move instead.</p><p>Start with an audit. Map every active AI initiative. For each one, answer three questions: what&#8217;s the dollar it&#8217;s chasing? What&#8217;s the milestone for this quarter? What happens if it misses? If no one can answer those questions for a given project, that project goes on a 30-day clock. Define the answers or kill it.</p><p>If you&#8217;re running a portfolio of AI initiatives, this is the political cover you&#8217;ve been missing. You&#8217;re not killing someone&#8217;s pet project, you&#8217;re applying the same milestone discipline you apply to every other investment. Frame it that way. &#8220;We&#8217;re not cutting AI. We&#8217;re making AI accountable.&#8221; The zombie projects die. The budget they were consuming flows to the projects that are actually working. Your best teams get more resources, not less.</p><p>If you&#8217;re about to pitch the next AI feature, this is your competitive advantage. Don&#8217;t walk into the room with &#8220;we should build an AI feature.&#8221; Walk in with &#8220;This is an Exploit-track project. Here&#8217;s the funnel. Here&#8217;s the dollar. Here&#8217;s the 90-day milestone. Here&#8217;s what we kill if it misses.&#8221; That pitch gets funded every time.</p><p>If you&#8217;re watching funded competitors burn cash on AI features with no revenue model, keep going. Your ROI obsession is correct. The discipline you&#8217;re forced into by your runway is the same discipline that separates the 6% from the 94%. You&#8217;re not the small one. You&#8217;re the smart one.</p><div><hr></div><p>Every AI project should have a dollar sign attached. Every one.</p><p>Not because innovation doesn&#8217;t matter. It does. But because the projects that change the world are the ones that survive long enough to do it. And survival requires proof. IBM Watson had a $5 billion budget. It didn&#8217;t survive. Zillow Offers had the largest real estate dataset in America. It didn&#8217;t survive. The projects that survive are the ones that can answer the question: what did we get for that?</p><p>The accountability gap is closing. The era where &#8220;we&#8217;re investing in AI&#8221; was enough to satisfy boards, impress LPs, and justify headcount is ending. The firms that imposed discipline early, that tracked ROI from day one, that killed projects at 90 days, that required every initiative to name its dollar before writing a line of code, will outperform.</p><p>The firms that let &#8220;strategic&#8221; be a hiding place will write checks they can&#8217;t explain. Some already are.</p><p>You already know how to run disciplined investments. You already know how to set milestones and kill what doesn&#8217;t work. You already know that &#8220;strategic&#8221; without a number is just another word for &#8220;we don&#8217;t know.&#8221;</p><p>Apply what you know. AI doesn&#8217;t get special treatment.</p><p><strong>ROI or die.</strong></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">BuilderLab publishes practical strategy for people who ship AI products. Subscribe so you don't miss out.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p><strong><br><br>Claim Verifications and Citations</strong></p><h6><em>$252 billion corporate AI spending in 2024. <br>Source: Stanford HAI AI Index 2025 Annual Report.</em></h6><h6><em>95% of enterprise AI pilots delivered zero measurable P&amp;L impact. <br>Source: MIT Project NANDA, &#8220;The GenAI Divide: State of AI in Business 2025&#8221; (July 2025).</em></h6><h6><em>42% of companies abandoned most AI initiatives, up from 17%. <br>Source: S&amp;P Global Market Intelligence, &#8220;Voice of the Enterprise: AI &amp; ML, Use Cases 2025.&#8221;</em></h6><h6><em>Gartner: 30% of GenAI projects abandoned after POC by end of 2025. <br>Source: Gartner press release, Data &amp; Analytics Summit, Sydney (July 2024).</em></h6><h6><em>AI projects fail at twice the rate of non-AI IT projects. <br>Source: RAND Corporation, &#8220;Root Causes of Failure for AI Projects&#8221; (2024).</em></h6><h6><em>AI personalization lifting ecommerce conversion by up to 23%. <br>Source: Multiple industry reports (Envive AI, McKinsey retail research).</em></h6><h6><em>Contact centers using AI see 30% operational cost reductions. <br>Source: Multiple industry sources (Fullview, AmplifAI, McKinsey CX research).</em></h6><h6><em>AI users completed tasks 25% faster at 40% higher quality. <br>Dell&#8217;Acqua et al., &#8220;Navigating the Jagged Technological Frontier,&#8221; HBS Working Paper 24-013 (Sept 2023).</em></h6><h6><em>Early GenAI adopters report $3.70 per dollar; top performers $10.30. <br>Source: Widely cited (CIO Dive, BotsCrew, Fullview).</em></h6><h6><em>6% &#8220;AI high performers&#8221; attribute 5%+ of EBIT to AI. <br>Source: McKinsey, &#8220;The State of AI in 2025&#8221; (Nov 2025).</em></h6><h6><em>GitHub Copilot: 55% faster coding in controlled studies. <br>Source: GitHub Research (2022). Controlled experiment, P=.0017.</em></h6><h6><em>88% of accepted Copilot suggestions stay in the codebase. <br>Source: GitHub usage data, confirmed across multiple compilations.</em></h6><h6><em>Coding AI: $4B category in 2025, up from $550M. <br>Source: Menlo Ventures, &#8220;2025: The State of Generative AI in the Enterprise&#8221; (Jan 2026).</em></h6><h6><em>IBM spent ~$5B building Watson Health. <br>Source: IBM acquisition costs (Truven, Phytel, Explorys, Merge). Sold to Francisco Partners for ~$1B (2022).</em></h6><h6><em>MD Anderson spent $62M; system never used on a patient. <br>Source: University of Texas System audit report (Jan 2016). Confirmed by JNCI, Medscape, Forbes.</em></h6><h6><em>MD Anderson original budget $2.4M, ballooned to $62M. <br>Source: UT audit. IBM contract extended 12 times.</em></h6><h6><em>Contracts structured below board approval threshold. <br>Source: UT audit. &#8220;Fees consistently set just below the amount that would have required Board approval.&#8221;</em></h6><h6><em>System built on legacy EHR (ClinicStation) already replaced by Epic. <br>Source: UT audit.</em></h6><h6><em>Zillow iBuying lost $304M in Q3 2021; $500M+ write-downs; 2,000 laid off; $9B market cap loss. <br>Source: Zillow Q3 2021 earnings and public filings.</em></h6><h6><em>Opendoor: $170M profits, 7.3% gross margins same quarter. <br>Source: Opendoor financial filings, same period.</em></h6><h6><em>McKinsey: workflow redesign strongest predictor of AI impact. <br>Source: McKinsey, &#8220;The State of AI in 2025.&#8221; Note: McKinsey says &#8220;one of the strongest&#8221;.</em></h6><h6><em>High performers 3x more likely to have redesigned workflows. <br>Source: McKinsey, &#8220;The State of AI in 2025.&#8221;</em></h6><h6><em>BCG: emphasizing early impact increases success likelihood by six percentage points. Source: General thesis supported by BCG AI Radar 2025.</em></h6>]]></content:encoded></item><item><title><![CDATA[You're Not Building an ‘AI Product’.]]></title><description><![CDATA[Why model performance is a design constraint, and calibration is the real AI skill.]]></description><link>https://www.builderlab.ai/p/youre-not-building-an-ai-product</link><guid isPermaLink="false">https://www.builderlab.ai/p/youre-not-building-an-ai-product</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 01 Feb 2026 13:42:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/39261fb2-31c7-46a7-885d-c548d4b54aec_1432x1428.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>You&#8217;re an ecommerce app. You have an AI model that predicts which customers will cancel their order before it ships. Three scenarios: the model has 99% precision, 80% precision, or 40% precision. What do you build with each?</em></p><p><em>Precision means: when the model flags someone, it&#8217;s right that percentage of the time.</em></p><p><em>Most people answer like this:</em></p><ol><li><p><em><strong>99%:</strong> &#8220;Automate it, cancel their order before they do.&#8221;</em></p></li><li><p><em><strong>80%:</strong> &#8220;Good start, but let&#8217;s improve the model before we ship.&#8221;</em></p></li><li><p><em><strong>40%:</strong> &#8220;Not useful. Too many false positives.&#8221;</em></p></li></ol><p><em>Wrong. Wrong. Wrong.</em></p><p><em>The right answer reveals the skill that actually matters when you&#8217;re building with AI. Almost nobody has it.</em></p><div><hr></div><h2><strong>The Answer</strong></h2><p>Notify the customer: &#8220;Confirm or cancel your order&#8221;. Same model. Same email. Different intensity.</p><p><strong>99% precision:</strong> Aggressive CTA. &#8220;Still want this? Click to confirm or we&#8217;ll cancel your order.&#8221; You&#8217;re almost certain they&#8217;ll cancel, better they do it now than you ship it and eat the return costs.</p><p><strong>80% precision:</strong> Medium CTA. &#8220;Your order ships tomorrow!&#8221; with a visible, easy cancel button. You&#8217;re right 4 out of 5 times. Give them a clear out. The 1 in 5 who weren&#8217;t going to cancel? Most of them feel informed. No harm done.</p><p><strong>40% precision:</strong> Soft CTA. &#8220;Your order is on the way!&#8221; with a subtle cancel link. You&#8217;re wrong more often than you&#8217;re right, but you&#8217;ve still got a meaningful signal over random. You can&#8217;t push hard. But you can quietly surface the option for those who need it.</p><p>One model. Three scenarios. Three calibrations.</p><p>The precision is the constraint. The intervention intensity is the variable.</p><p>This is the entire game. And most builders don&#8217;t even know they&#8217;re playing it.</p><div><hr></div><h2><strong>The Skill Nobody Teaches</strong></h2><p>Most builders ask: <em>Is the model good enough?</em></p><p>The rare ones ask: <em>Good enough for what?</em></p><p>That&#8217;s the skill gap. Not &#8220;how do I handle uncertainty.&#8221; But &#8220;how do I calibrate the product to the confidence level I actually have?&#8221;</p><p>Netflix understood this early. Their recommendation algorithm doesn&#8217;t need to be right, it needs to surface options you&#8217;ll scroll past if they&#8217;re wrong. 80% of Netflix watch time comes from recommendations. Not because they&#8217;re perfectly accurate, but because the cost of a bad recommendation is close to zero. You see a title you don&#8217;t want. You scroll. You pick something else. Maybe you hover, decide not to click, move on. The algorithm watches all of it: the hovers, the skips, the half-watched episodes, and learns. Netflix has said this saves them an estimated $1 billion per year in reduced churn. Not by being right. By making &#8220;wrong&#8221; cheap.</p><p>Spotify understood this. Discover Weekly gives you 30 songs you&#8217;ve never heard. Some you&#8217;ll hate. But the product isn&#8217;t &#8220;songs you&#8217;ll love.&#8221; It&#8217;s &#8220;songs you <em>might</em> love.&#8221; The intervention is calibrated perfectly: skipping costs 30 seconds and a thumb tap. No wasted money. No broken workflow. No angry customer. And your skips train the algorithm just as much as your saves do. Failure is feedback. The product gets better specifically because it&#8217;s allowed to be wrong.</p><p>GitHub Copilot understood this at a deeper level. Roughly 30% suggestion acceptance rate. Think about that. Most suggestions are wrong. Seven out of ten lines Copilot writes are rejected. By any traditional software metric, that&#8217;s a broken feature. But Copilot is one of the fastest-growing developer tools in history. Why? Because a wrong suggestion costs you nothing. Gray text appears, you keep typing, it vanishes. A right suggestion saves you minutes. Tab to accept, keep typing to reject. 88% of accepted code stays in the codebase. Users report coding 55% faster. The product works <em>because</em> failure is free, not in spite of its error rate.</p><p>The pattern is consistent across all three. Winners don&#8217;t wait for perfect precision. They calibrate the intervention to the confidence they have. They design the product around the error rate, not despite it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">I write about calibration, failure modes, and how real AI products actually ship. <strong>Build better AI products.</strong></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>The Cost of Getting It Wrong</strong></h2><p>Air Canada didn&#8217;t understand this.</p><p>Their chatbot gave a customer wrong information about bereavement fares. The customer relied on it, booked a flight, and applied for the discount. Air Canada refused the refund, then argued in a tribunal that the chatbot was &#8220;a separate legal entity responsible for its own actions.&#8221;</p><p>The tribunal called this a &#8220;remarkable submission.&#8221; Ordered Air Canada to pay $650. The chatbot was pulled from the website.</p><p>The product failed because nobody asked &#8220;what&#8217;s our precision, and what intervention is appropriate for that precision?&#8221; A chatbot that gives direct answers acts like a 99% precision product. Air Canada didn&#8217;t have 99% precision. They had a confident-sounding model and no fallback.</p><p>McDonald&#8217;s made the same mistake differently. They spent years building an AI drive-thru system with IBM. It misinterpreted orders, adding bacon to ice cream, ringing up 260 chicken nuggets. Discontinued in 2024 across more than 100 locations.</p><p>The precision might have been fine for a soft suggestion. &#8220;Did you mean a medium combo?&#8221; But they built a hard automation. Drive-thru failure isn&#8217;t like Netflix failure. You can&#8217;t scroll past a wrong order. The customer sees it. The kitchen makes it. Wasted food, frustrated customers, operational chaos. They mismatched the intervention to the precision.</p><p>Klarna made the boldest version of this mistake. In 2024, they replaced 700 customer service agents with AI. The CEO bragged publicly that he hadn&#8217;t hired a human in a year. By 2025, they were reassigning software engineers to handle customer calls because they couldn&#8217;t rehire fast enough. The CEO now says Klarna wants to be &#8220;best at offering a human to speak to.&#8221;</p><p>The pattern is just as consistent on the losing side. Every one of these companies built interventions that were too aggressive for their precision. No fallback. No calibration. No recovery path.</p><div><hr></div><h2><strong>The Frame</strong></h2><p>This is the divide:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X-fd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X-fd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 424w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 848w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 1272w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X-fd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png" width="1350" height="694" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:694,&quot;width&quot;:1350,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:136274,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/186497963?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X-fd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 424w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 848w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 1272w, https://substackcdn.com/image/fetch/$s_!X-fd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd24571e8-06a7-4981-a4d9-d8408c12dd93_1350x694.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The precision didn&#8217;t determine success. The match between precision and intervention did.</p><p>Every winner on this list has a cheap failure mode. Every loser has an expensive one. That&#8217;s not an accident. That&#8217;s a product decision someone either made deliberately or failed to make at all.</p><h2><strong>What This Looks Like in Practice</strong></h2><p>Way back at Booking.com, I had a model that detected the topic of customer service requests. Billing, cancellations, property issues, check-in problems.</p><p>The goal was routing. If you know what the customer needs, you send them to the right agent on the first try. No transfers. Transfers are expensive. Every handoff costs time, frustration, and money. A customer who gets transferred once is measurably less satisfied. Transferred twice and you&#8217;ve probably lost them. And you&#8217;ve wasted time for 3 customer service agents.</p><p>So we tuned for high precision. When the model predicted a topic, it had to be right. And it was. But recall was 5%. The model barely fired. Out of every hundred customer requests, it confidently classified about five. The other ninety-five got the same routing they always had.</p><p>It sat there, almost perfectly correct and almost perfectly useless. Didn&#8217;t move metrics. Couldn&#8217;t move metrics. A model that&#8217;s right 95% of the time but only makes a prediction 5% of the time has almost no product impact.</p><p>Same model. We lowered the precision threshold. Recall jumped to 80%+. Now it caught most incoming requests, but it was wrong more often. Meaningfully wrong.</p><p>The old intervention, routing, couldn&#8217;t tolerate that. If you send a customer with a billing issue to the cancellation team, you&#8217;ve created the exact problem you were trying to solve. The customer gets transferred. They wait again. They&#8217;re angrier than when they started. Wrong routing is worse than no routing.</p><p>So we changed the intervention.</p><p>Instead of routing to an agent, we surfaced self-service options based on the predicted topic. <em>Before</em> the customer ever reached a human. &#8220;It looks like you might need help with your cancellation. Here are some options.&#8221;</p><p>Prediction right? Customer solves it themselves. Never reaches an agent. Cost to serve: near zero.</p><p>Prediction wrong? Customer ignores the suggestions, closes the panel, and connects to an agent anyway. No harm done.</p><p>30% reduction in contact rates. Same model. Same underlying precision.</p><p>The precision didn&#8217;t change the outcome. The match between precision and intervention did. We didn&#8217;t make the model better. We made the product smarter about how it used the model.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This is the kind of product judgment most teams never develop. I break these decisions down in BuilderLab.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>The 70% Mindset</strong></h2><p>The rare builders share a mental model I call the 70% mindset.</p><p>They internalize that the model will be wrong some percentage of the time, and they calibrate the intervention to that percentage. Not hide it. Not wait to fix it. <em>Design for it.</em> They treat the error rate as a first-class input to product design, not as a bug to be squashed before launch.</p><p>The simulation they run:</p><ol><li><p>&#8220;I have 70% precision. What intervention is appropriate for that?&#8221;</p></li><li><p>NOT &#8220;I need 95% precision. How do I get there?&#8221;</p></li></ol><p>The difference sounds subtle. It&#8217;s not. The first question leads to a product you can ship this quarter. The second leads to a roadmap that never converges.</p><p>They start with the constraint. They calibrate inside it.</p><p>Gmail Smart Compose works maybe 30% of the time. But it&#8217;s explicitly designed not to be authoritative. It&#8217;s gray text that disappears as you type. Wrong guesses vanish silently. You never feel wrong. You never feel corrected. The model&#8217;s error rate is invisible because the intervention is so soft that failure literally doesn&#8217;t exist from the user&#8217;s perspective. Soft intervention for low precision. Perfectly calibrated.</p><p>Perplexity has variable precision across topics. A question about recent events might be less reliable than a question about established science. But every claim has a numbered source. Wrong answer? One click to verify. They didn&#8217;t try to eliminate error. They built verification into the product itself. The intervention <em>assumes</em> the model will be wrong sometimes and gives the user the tool to catch it. That&#8217;s calibration.</p><p>Waze has roughly 85% precision on surface street timing. But the product isn&#8217;t &#8220;perfect ETA.&#8221; It&#8217;s &#8220;less traffic.&#8221; Wrong prediction? You still arrive. Maybe a few minutes later than expected. And your drive, the actual data from your actual route, makes the next prediction better for everyone. Soft intervention that self-corrects with feedback. The product improves <em>because</em> it ships with imperfect precision, not in spite of it.</p><p>This is the mindset shift that separates builders who ship from builders who wait. Stop asking &#8220;how do I make the model better?&#8221; Start asking &#8220;what product can I build with the model I have?&#8221;</p><p>A startup with 40% precision can beat an incumbent with 95%&#8212;if the startup calibrates the intervention correctly and the incumbent doesn&#8217;t. This happens more often than you think.</p><div><hr></div><h2><strong>Why Almost Nobody Gets This Right</strong></h2><p>Nothing in conventional product training develops this skill. And it&#8217;s not because training is bad, it&#8217;s because the fundamental paradigm changed and the training hasn&#8217;t caught up.</p><p>Traditional software is deterministic. The spec says X, the engineer builds X, and X either works or it doesn&#8217;t. The button either saves the file or it throws an error. The search either returns results or it returns nothing. There&#8217;s no &#8220;the button saves your file 80% of the time.&#8221; You don&#8217;t ship that. You fix the bug.</p><p>This is how every PM, designer, and engineer has been trained to think. Define the requirement. Build to the requirement. Test against the requirement. Ship when it passes.</p><p>AI is probabilistic. The model says X, and X is right some percentage of the time. There is no &#8220;fix the bug.&#8221; The error rate <em>is</em> the product. The intervention has to match the confidence. &#8220;Ship it&#8221; isn&#8217;t enough. &#8220;Ship it with the right calibration&#8221; is the actual job, and nothing in the traditional playbook prepares you for that.</p><p>The entire pipeline: CS education, PM bootcamps, design programs, agile certifications, produces builders who think in certainties. Then hands them probabilistic tools and expects them to figure it out. Most don&#8217;t. They either wait for certainty that will never come, or they ship with certainty they don&#8217;t have.</p><p>DPD&#8217;s chatbot started swearing at customers and writing poems mocking the company after a system update. Disabled immediately. Nobody had asked: &#8220;What&#8217;s our precision on edge cases, and what intervention is appropriate when the model goes off the rails?&#8221; They built a product that assumed the model would behave predictably. The model didn&#8217;t. There was no fallback, no graceful degradation, no circuit breaker.</p><p>That&#8217;s not an AI failure. That&#8217;s a calibration failure. The skill that would have prevented it isn&#8217;t technical, it&#8217;s judgment. Matching intervention intensity to model confidence. Designing the failure state as carefully as the success state.</p><p>My team at Booking.com reviews 2,000 to 2,500 applications per year. We hire 3, maybe 5. That&#8217;s a 0.2% acceptance rate. As selective as Google, more selective than Y Combinator, 18x more selective than Harvard.</p><p>The rare ones don&#8217;t have better AI knowledge. They have better calibration instincts. They hear &#8220;40% precision&#8221; and immediately start designing interventions. Everyone else hears &#8220;40% precision&#8221; and says &#8220;not good enough.&#8221; The rare ones ask &#8220;good enough for what?&#8221; Everyone else asks &#8220;how do we get to 95%?&#8221;</p><p>That&#8217;s the gap. And no amount of &#8220;AI PM&#8221; certifications will close it.</p><div><hr></div><h2><strong>The Calibration Checklist</strong></h2><p>Before you build anything, answer these six questions:</p><ol><li><p><strong>What&#8217;s the precision?</strong> Not what you hope for, what you actually have right now. When the model flags something, how often is it right?</p></li><li><p><strong>What intervention does that precision support?</strong> 99% supports hard automation. 80% supports clear suggestions. 40% supports soft nudges. Different precisions demand different intervention intensities.</p></li><li><p><strong>What happens when it&#8217;s wrong?</strong> Describe the failure state as vividly as you describe the success state. If you can&#8217;t, you don&#8217;t understand your product yet.</p></li><li><p><strong>Who pays the cost of failure?</strong> The user? The company? A third party? The answer changes your risk tolerance entirely.</p></li><li><p><strong>How do you make failure cheap?</strong> Soft intervention? Easy reversal? Built-in verification? The best AI products are designed so that being wrong costs almost nothing.</p></li><li><p><strong>How do you detect failure?</strong> For high-stakes decisions, you need individual flags. For low-stakes, aggregate metrics. Match your monitoring to your risk.</p></li></ol><p>If you can answer all six, you might have a product.</p><p>If you can only answer the first two, you have a demo.</p><div><hr></div><h2><strong>The Exercise</strong></h2><p>Here&#8217;s how to develop this skill: run the 99/80/40 exercise on your own product.</p><p>Take whatever your model does, recommendations, predictions, classifications, generation. Now imagine three scenarios:</p><ol><li><p><strong>99% precision.</strong> What would you build? How aggressive would the intervention be? At this level, you can automate hard. You can take action on behalf of the user. You can make decisions that are expensive to reverse.</p></li><li><p><strong>80% precision.</strong> What changes? What do you pull back? At this level, you can suggest confidently. You can surface options prominently. But you need the user in the loop. You need a clear &#8220;no thanks&#8221; path.</p></li><li><p><strong>40% precision.</strong> What do you build now? Most people say &#8220;nothing.&#8221; The right answer is almost never nothing. At this level, you can still nudge. You can still surface. You can still learn. The intervention just has to be soft enough that being wrong is invisible.</p></li></ol><p>Then ask yourself: where is my product right now? And is my intervention calibrated to that precision&#8212;or am I pushing too hard?</p><p>If you&#8217;re building a 99%-confidence intervention on an 80%-precision model, you&#8217;re building the next Air Canada chatbot. If you&#8217;re waiting for 95% precision before shipping anything, you&#8217;re leaving the Copilot opportunity on the table. Both mistakes come from the same place: not treating the precision as a design input.</p><p>Start with the constraint. Calibrate inside it. Ship.</p><div><hr></div><h2><strong>The Point</strong></h2><p>You&#8217;re not building an AI product.</p><p>You&#8217;re building a product that has a capability which is right some percentage of the time. That percentage isn&#8217;t a limitation, it&#8217;s a design constraint. And like every design constraint, it produces better products when you respect it than when you fight it.</p><p>The question isn&#8217;t &#8220;how do I make the model more precise?&#8221; The question is &#8220;what intervention does this precision support?&#8221;</p><p>The 99% model and the 40% model aren&#8217;t better or worse. They&#8217;re <em>different</em>. Different precisions, different interventions, different products. A 40% model with the right intervention beats a 95% model with the wrong one. Every time. Copilot at 30% acceptance rate generates more revenue than McDonald&#8217;s AI drive-thru at 95% accuracy. Netflix at 80% relevance saves more money than Air Canada&#8217;s chatbot at 95% confidence. The precision doesn&#8217;t determine the outcome. The calibration does.</p><p>The AI arms race is misunderstood. It&#8217;s not about who has the best model. It&#8217;s about who matches the intervention to the precision. That&#8217;s a product skill, not an AI skill. And it&#8217;s the reason the &#8220;AI PM&#8221; job title is fake. The skill that matters here, calibrating intervention to confidence, is product judgment. It&#8217;s design sense. It&#8217;s understanding what happens when things go wrong and building for that from day one.</p><p>Saying you&#8217;re building an &#8220;AI product&#8221; is like saying you&#8217;re building a &#8220;C++ product.&#8221; The technology is an implementation detail. The market doesn&#8217;t care what&#8217;s under the hood. Users don&#8217;t care what&#8217;s under the hood. Your CFO definitely doesn&#8217;t care what&#8217;s under the hood.</p><p>What matters is what the product <em>does</em>.</p><p>Lead with that. &#8220;20% faster. 30% cheaper. 10x more scale.&#8221;</p><p>When they ask how, <em>then</em> you say AI.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If this changed how you think about AI products, you&#8217;ll like BuilderLab.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/youre-not-building-an-ai-product?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/youre-not-building-an-ai-product?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Growth Machines: The Notion Story]]></title><description><![CDATA[The growth framework that turned near-death into $10B.]]></description><link>https://www.builderlab.ai/p/growth-machines-the-notion-story</link><guid isPermaLink="false">https://www.builderlab.ai/p/growth-machines-the-notion-story</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 25 Jan 2026 16:28:56 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2893e37e-79aa-421c-a7e3-b0340b97e23b_944x944.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In 2015, Notion was dead. Out of money, team laid off, product too confusing for anyone to use.</p><p>Ten years later: 100 million users, $10B valuation, and a playbook every founder wants to copy.</p><p>The common explanation, &#8220;templates went viral&#8221;,  misses the point. Notion won because they solved a problem most builders ignore: <strong>how to make a product get deeper with use, not just wider with distribution.</strong></p><p>Most startups chase distribution first: more users, more channels, more reach. Notion did the opposite. They refused to scale until they&#8217;d solved a different problem: whether users would do <em>more</em> with the product over time, not just try it once.</p><p>That&#8217;s the difference between depth and distribution. Depth means users replace other tools, customize their setup, and eventually share what they&#8217;ve built, not because you incentivized them, but because the artifact is useful on its own. Distribution just brings people to the door. Depth determines whether they stay.</p><p>Notion&#8217;s insight: <strong>distribution only compounds when depth is already rising.</strong> Without depth, growth is just accelerated churn. With it, every new user makes the product stickier for the next.</p><p>Here&#8217;s how they built that depth, and the framework for knowing when you&#8217;re ready to scale.</p><h2><strong>The first Notion failed on legibility, not ambition</strong></h2><p>Early Notion was ambitious in exactly the way builders admire. It was flexible, expressive, and theoretically powerful. The problem wasn&#8217;t quality; it was interpretation. New users were confronted with too much possibility and too little guidance. The product demanded imagination before it delivered value, which meant only a narrow slice of users could succeed without help. Most didn&#8217;t churn because the product disappointed them. They churned because they didn&#8217;t know what to do.</p><p>Ivan Zhao later reflected on this failure: &#8220;We focused too much on what we wanted to bring to the world. We needed to pay attention to what the world wanted from us&#8221;[1]. That sentence matters because it reframes the problem. This wasn&#8217;t about marketing reach or onboarding polish. It was about legibility, about whether a user could look at the product and immediately recognize how it fit into their life.</p><p>The deeper realization came later: &#8220;People don&#8217;t want to build apps. They don&#8217;t think about programming, they simply want something that solves their problems quickly and efficiently&#8221;[2]. The first Notion had been built for toolmakers. Toolmakers, it turned out, are rare.</p><p>Legibility is a growth constraint, one often ignored. If users can&#8217;t see themselves in the product quickly, no amount of distribution fixes that. It only accelerates confusion.</p><h2><strong>The rebuild was about first value, not feature power</strong></h2><p>When Notion ran out of money, Ivan Zhao and Simon Last made a brutal calculation. They laid off their entire team of four, sublet their San Francisco office, and moved to Kyoto [3]. Kyoto housing was larger and cheaper, the cost of living was dramatically lower, and they could focus without distraction. In a rented two-story house with paper walls, no heating, and bedrooms separated only by shoji screens, they spent the next year rebuilding everything from scratch [4].</p><p>&#8220;Neither of us spoke Japanese and nobody there spoke English,&#8221; Ivan later said, &#8220;so all we did was code in our underwear all day&#8221; [5].</p><p>The central question wasn&#8217;t how to relaunch bigger. It was how to relaunch clearer. What does meaningful value look like in the first hour of use? That question forced a different kind of rebuild. The goal was no longer to showcase flexibility, but to reduce cognitive load. Every interaction was redesigned so the product felt usable before it felt powerful.</p><p>During this period, Ivan appeared at the top of Figma&#8217;s most active user list, spending 18+ hours a day iterating on the interface [6]. It was a war against ambiguity. As Ivan later put it, &#8220;Designers spend too much time on edge cases. Most of the time, what matters is the dumbest path&#8221;[7]. In Kyoto, they found that path.</p><p>This was Notion&#8217;s first major growth decision: scale didn&#8217;t matter until first value was obvious.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>Skip the BS. Build what matters. Join 2k+ builders.</em></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h2><strong>Templates were not a growth hack</strong></h2><p>When Notion relaunched, they needed traction fast. Ivan asked early investor Naval Ravikant to wake up at midnight, when Product Hunt&#8217;s daily list refreshed, and tweet to his followers to vote. Naval had, in Ivan&#8217;s words, &#8220;a bazillion Twitter followers&#8221; [8]. The gambit worked. Notion hit #1 for the day, week, and month.</p><p>But the product they launched wasn&#8217;t a blank canvas. It shipped with a set of internally created templates: project trackers, team wikis, meeting notes, reading lists. These templates are often described as a clever growth tactic. That misses the point. Templates were Notion&#8217;s solution to the legibility problem.</p><p>A blank page asks users to imagine value. A working system demonstrates it. Templates allowed users to start inside something that already worked, rather than designing from scratch. This collapsed time-to-value and removed the need for explanation. Users didn&#8217;t have to ask what Notion could be. They could see it operating in a familiar context.</p><p>Ivan later described this as &#8220;sugar-coated broccoli.&#8221; The original vision of letting anyone build their own software tools was the broccoli. Most people don&#8217;t want broccoli. But wrap it in something they already care about, like notes, docs, project management, and they&#8217;ll consume the underlying capability without resistance [9].</p><p>The critical move came next. Notion allowed users to publish and share templates themselves. This turned onboarding into a distributed system. New users increasingly encountered Notion not as a product, but as a solution, as a specific system someone else was already using. They didn&#8217;t arrive asking &#8220;what is Notion?&#8221; They arrived asking &#8220;how do I use this?&#8221;</p><p>Over two years, templates from Notion&#8217;s early gallery were downloaded over one million times [10].</p><h2><strong>Growth became possible once usage became shareable</strong></h2><p>At this point, Notion&#8217;s growth stopped being fragile. A user would customize a template, embed it into their workflow, and eventually share it because the artifact had value on its own. Sharing wasn&#8217;t incentivized, it was a byproduct of real work.</p><p>This is a crucial inversion. Most products try to engineer virality on top of shallow usage. Notion did the opposite. It engineered deep usage first, then let sharing emerge naturally. That&#8217;s why the template ecosystem could exist outside the company. Creators were selling on Gumroad, in newsletters, on YouTube, without collapsing into spam. The artifacts were useful even if you never signed up. Some creators have built entire businesses selling Notion templates, with top sellers earning over $1 million without writing a line of code [11].</p><h2><strong>Community amplified depth; it didn&#8217;t manufacture it</strong></h2><p>Notion&#8217;s community is often cited as the second reason for its growth. That gets causality backward. Community only worked because the product was already embedded in people&#8217;s workflows.</p><p>Ivan&#8217;s early behavior wasn&#8217;t a branding exercise. It was product development disguised as customer support. &#8220;In the early days, I spent a lot of time replying one-on-one to users on Twitter, making sure they knew there was a human behind Notion,&#8221; he explained. &#8220;We also logged every piece of feedback we got and tagged it so we could use it to develop features&#8221; [12].</p><p>When creators like Marie Poulin shared early frustrations with the product&#8217;s learning curve, Ivan didn&#8217;t just acknowledge the feedback. He watched how they used the product and shipped changes to remove friction [13]. That responsiveness turned critics into evangelists, because users could see their fingerprints on the product.</p><p>The results compounded. One ambassador now runs Notion&#8217;s subreddit with hundreds of thousands of members. Others built large regional communities in Korea, the Middle East, and Europe [14]. These were power users who became invested enough to build ecosystems around the tool.</p><p>Community didn&#8217;t create enthusiasm from nothing. It amplified usage density that already existed. Without that depth, community would have been noise.</p><h2><strong>Why Notion looked slow, and why that was the point</strong></h2><p>For years, Notion didn&#8217;t look impressive by startup standards. There were no viral spikes, no aggressive funnels, no growth-at-all-costs playbook. From the outside, it looked like restraint. Internally, it was discipline.</p><p>That discipline came at a cost. While Notion was rebuilding in Kyoto, Coda launched with $60M in funding. Airtable was scaling aggressively into enterprise. Slack had already become a verb. Investors knocked so persistently that Notion eventually moved offices and removed itself from Google listings. Ivan declined most conversations [15].</p><p>When Notion finally raised its Series A in 2020, it reached a $2B valuation with roughly 30 employees [16]. That ratio tells you what they were optimizing for: depth, not headcount.</p><p>And then that depth nearly broke them. During COVID, as millions of users flooded the platform, Notion&#8217;s infrastructure buckled. For years, the entire product ran on a single Postgres instance, upgraded again and again. Eventually, there were no larger machines left. There was a literal doomsday clock counting down to when the database would run out of space and the product would shut down [17].</p><p>They froze feature development. Every engineer pivoted to infrastructure. It was, in Ivan&#8217;s words, a close call [18].</p><p>They survived. And when Notion finally scaled, it did so cleanly because the system underneath could absorb load. Each cohort customized more deeply, relied on the product more centrally, and replaced more tools than the one before.</p><h2><strong>Notion and the AI Wave: Sticking to core principles.</strong></h2><p>When Notion launched AI features in February 2023, they faced a strategic choice: build a standalone AI tool, or embed intelligence into the workspace users had already built.</p><p>They chose depth over distribution again.</p><p>Notion AI is embedded directly into pages, databases, and workflows where users already work [19]. This matters because Notion had spent years getting users to build systems inside the product:  project trackers, team wikis, meeting notes, personal dashboards. That accumulated structure became AI&#8217;s context layer.</p><p>A standalone AI tool knows nothing about your work. Notion AI knows your page relationships, your database schemas, your team&#8217;s writing style, your project history. When you ask it to summarize last week&#8217;s meetings or draft a document in your company&#8217;s voice, it draws on depth that already exists [20].</p><p>Ivan described this as &#8220;consolidate before you automate&#8221; [21]. The years Notion spent getting users to replace fragmented tools with a unified workspace were AI groundwork. The more deeply users embedded their work in Notion, the more valuable AI assistance became, because context compounds.</p><p>This is sugar-coated broccoli, version two. The original vision of letting anyone build software tools was hidden inside productivity features people already wanted. The AI vision, giving everyone a reasoning engine for their work, is now hidden inside the workspace they&#8217;ve already built. Users don&#8217;t adopt &#8220;AI.&#8221; They just notice that Notion got smarter.</p><p>The pattern repeats: <strong>depth created the surface area for the next capability to land.</strong></p><p>Most companies bolt AI onto shallow products and wonder why adoption stalls. Notion waited until users had built something worth augmenting. That patience is compounding again.</p><h2><strong>The real lesson: depth earns distribution</strong></h2><p>Notion&#8217;s story collapses into a single growth law:</p><p><strong>If the product doesn&#8217;t get deeper with use, distribution just buys churn.</strong></p><p>Templates, community, and word of mouth worked because users were already doing more with the product over time. Most builders invert this order. They chase reach before depth, then wonder why growth feels fragile. The Notion story is instructive. But stories don&#8217;t scale. Frameworks do.</p><div><hr></div><h2><strong>From story to system: the Usage Density Framework </strong></h2><p>The Notion story collapses into a single constraint:</p><blockquote><p><strong>Growth compounds only when usage deepens faster than reach expands.</strong></p></blockquote><p>Everything that follows exists to help you answer one question honestly:</p><p><strong>When a new user shows up, does the product become clearer or more fragile?</strong></p><p>The Usage Density Framework answers that question using two lenses:</p><ol><li><p>A <strong>diagnostic</strong> (what the system looks like under load)</p></li><li><p>A <strong>depth ladder</strong> (why some users become distribution and others become churn)</p></li></ol><h3>1. Usage Density Diagnostic</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zArL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zArL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 424w, https://substackcdn.com/image/fetch/$s_!zArL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 848w, https://substackcdn.com/image/fetch/$s_!zArL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 1272w, https://substackcdn.com/image/fetch/$s_!zArL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zArL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png" width="1406" height="410" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:410,&quot;width&quot;:1406,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:67560,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/185731527?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zArL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 424w, https://substackcdn.com/image/fetch/$s_!zArL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 848w, https://substackcdn.com/image/fetch/$s_!zArL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 1272w, https://substackcdn.com/image/fetch/$s_!zArL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0ec5ba-80ff-49c5-beb4-c0b5ed2cef2d_1406x410.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This table is a <strong>stress test</strong>.</p><p>Each row is asking the same underlying question in a different way:</p><blockquote><p><strong>When new users arrive, do they converge toward a clear way of using the product, or do they fragment into edge cases?</strong></p></blockquote><p>If any row is firmly in the <em>Low Density</em> column, scaling acquisition will increase fragility.</p><p>Let&#8217;s unpack what each dimension actually means in practice.</p><h4><strong>Core feature usage</strong></h4><p><strong>Fragmented</strong> usage looks like engagement. Users click around, touch many features, and explore the surface area. This often shows up as healthy DAU/MAU, but it&#8217;s a false signal. Fragmentation usually means users are uncertain about what the product is for.</p><p><strong>Converging</strong> usage means retained users increasingly rely on the same subset of functionality. This is the product teaching users its intended shape.</p><p><strong>Dominant</strong> usage means there is an obvious default mode of use. This becomes the product&#8217;s center of gravity.</p><p><strong>How to test this</strong></p><ul><li><p>Take your top 20% most active users</p></li><li><p>Look at which features they use weekly</p></li><li><p>If you can&#8217;t describe a default usage pattern in one sentence, you&#8217;re still fragmented</p></li></ul><p><strong>Notion example<br></strong>Retained users converged on wikis, docs, and project systems, not every block type.</p><h4><strong>Sharing behavior</strong></h4><p>This dimension distinguishes <strong>earned distribution</strong> from <strong>forced distribution</strong>. <strong>Incentivized</strong> sharing requires prompts, rewards, or nudges. It scales poorly and degrades trust. <strong>Occasional</strong> sharing happens when users see value but don&#8217;t yet rely on it. <strong>Natural</strong> sharing happens because the artifact itself is useful. The user shares because it helps <em>them</em>, not you.</p><p><strong>How to test this</strong></p><ul><li><p>Look at shares that happen without prompts</p></li><li><p>Identify what&#8217;s being shared (pages, reports, templates)</p></li><li><p>Ask users why they shared, and if the answer is &#8220;because you asked,&#8221; it doesn&#8217;t count</p></li></ul><p><strong>Notion example<br></strong>Users shared templates because the template <em>was the value</em>.</p><h4><strong>Time-to-value</strong></h4><p>This row tells you how much imagination your product requires. <strong>Long</strong> time-to-value means users must configure, design, or understand before benefiting. This kills density. <strong>Improving</strong> means onboarding and defaults are doing some work. <strong>Immediate</strong> means success is obvious in the first session.</p><p><strong>How to test this</strong></p><ul><li><p>Measure time to first meaningful action (not signup)</p></li><li><p>Watch first-session recordings</p></li><li><p>Ask: <em>Could someone succeed without reading docs?</em></p></li></ul><p><strong>Rule<br></strong> If your product requires explanation before value, don&#8217;t scale acquisition.</p><h4><strong>How to interpret the diagnostic as a whole</strong></h4><p>You don&#8217;t need perfection across all rows.<br> You <em>do</em> need <strong>convergence</strong>.</p><blockquote><p><strong>Good growth makes users behave more similarly over time, not less.</strong></p></blockquote><p>If growth increases variance, confusion, or support load, you&#8217;re scaling uncertainty.</p><h3>2. The Usage Depth Ladder</h3><p>The diagnostic tells you <strong>whether</strong> depth exists. The ladder explains <strong>why some users become distribution and others become churn</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g41I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g41I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 424w, https://substackcdn.com/image/fetch/$s_!g41I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 848w, https://substackcdn.com/image/fetch/$s_!g41I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 1272w, https://substackcdn.com/image/fetch/$s_!g41I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g41I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png" width="1210" height="292" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:292,&quot;width&quot;:1210,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:57033,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/185731527?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g41I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 424w, https://substackcdn.com/image/fetch/$s_!g41I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 848w, https://substackcdn.com/image/fetch/$s_!g41I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 1272w, https://substackcdn.com/image/fetch/$s_!g41I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F543563c9-a3bb-4228-9ff1-de2ff334285a_1210x292.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h4><strong>Layer 1: Adopt</strong></h4><p><strong>What it means<br></strong>The user starts from something that already works. This is where templates, defaults, and opinionated setups matter.</p><p><strong>Failure mode<br></strong>Users start from a blank state and freeze.</p><p><strong>What to measure</strong></p><ul><li><p>% of users who start from a default/template</p></li><li><p>% who complete setup in first session</p></li></ul><p>If users don&#8217;t adopt a working system, everything downstream fails.</p><h4><strong>Layer 2: Embed</strong></h4><p><strong>What it means<br></strong>The product becomes part of a recurring routine. Weekly planning. Daily notes. Team updates.</p><p><strong>Failure mode<br></strong>Usage depends on reminders, novelty, or nudges.</p><p><strong>What to measure</strong></p><ul><li><p>Weekly active usage</p></li><li><p>Regular cadence actions</p></li><li><p>Drop-off when notifications stop</p></li></ul><p><strong>Rule<br></strong> If usage collapses without reminders, you don&#8217;t have embedding.</p><h4><strong>Layer 3: Export</strong></h4><p><strong>What it means<br></strong>The user creates something worth sharing. This is where depth turns into distribution.</p><p><strong>Failure mode<br></strong>Sharing requires incentives or explicit prompts.</p><p><strong>What to measure</strong></p><ul><li><p>Shares per retained user</p></li><li><p>Invites without referral rewards</p></li><li><p>Artifact reuse by others</p></li></ul><p><strong>Key insight<br></strong>Export is not virality. It&#8217;s <em>proof of depth</em>.</p><h2><strong>The operating rule (this is the whole point)</strong></h2><p>Here is the rule that ties both tables together:</p><blockquote><p><strong>You are allowed to scale acquisition only if newer cohorts climb the ladder more reliably than older ones.</strong></p></blockquote><p>If newer cohorts:</p><ul><li><p>adopt less</p></li><li><p>embed less</p></li><li><p>export less</p></li></ul><p>then growth is degrading the system.</p><p>Stop. Fix depth. Resume later.</p><h2><strong>Why this matters</strong></h2><p>Most teams use dashboards to justify growth. These tables exist to <strong>prevent premature growth</strong>.</p><p>They force you to answer one uncomfortable question honestly:</p><blockquote><p><strong>Will more users make this product clearer, or will they expose what&#8217;s still broken?</strong></p></blockquote><p>Notion waited until the answer was &#8220;clearer.&#8221;</p><p>That patience is why their growth compounded.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>BuilderLab is for people who ship. Playbooks on scaling. No-BS AI. Frameworks you can use Monday Morning.</em></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p><h5>References</h5><h6>[1]: Figma Blog, &#8220;Design on a deadline: How Notion pulled itself back from the brink of failure,&#8221; March 2019. https://www.figma.com/blog/design-on-a-deadline-how-notion-pulled-itself-back-from-the-brink-of-failure/</h6><h6>[2]: Kitrum, &#8220;The Phenomenal Journey of Ivan Zhao, Notion&#8217;s Founder,&#8221; August 2025. https://kitrum.com/blog/the-phenomenal-journey-of-ivan-zhao-notions-founder/</h6><h6>[3]: Sequoia Capital, &#8220;Notion CEO Ivan Zhao: Augmenting Human Intellect,&#8221; April 2024. https://sequoiacap.com/article/notion-spotlight/</h6><h6>[4]: Sequoia Capital, ibid. (&#8221;a rented two-story house so small that only a traditional Shoji screen separated their bedrooms&#8221;)</h6><h6>[5]: Figma Blog, ibid.</h6><h6>[6]: Figma Blog, ibid. (&#8221;He suddenly popped to the top of our most active user list&#8212;spending upwards of 18+ hours a day in our design tool.&#8221;)</h6><h6>[7]: DigidAI, &#8220;Ivan Zhao: Notion&#8217;s $10B Productivity Vision,&#8221; November 2025. https://digidai.github.io/2025/11/23/ivan-zhao-notion-ceo-augmenting-human-intellect-deep-dive/</h6><h6>[8]: Sequoia Capital, ibid. (&#8221;And he has a bazillion Twitter followers&#8221;)</h6><h6>[9]: The Software Report, &#8220;The Kyoto Reboot: How Ivan Zhao Rebuilt Notion,&#8221; 2025. https://www.thesoftwarereport.com/the-kyoto-reboot-how-ivan-zhao-rebuilt-notion/</h6><h6>[10]: Notion Blog, &#8220;A brand new gallery for Notion templates and community creations,&#8221; November 2021. https://www.notion.com/blog/new-template-gallery</h6><h6>[11]: Simple.ink, &#8220;Notion Statistics: Growth, Revenue &amp; More,&#8221; 2024. https://www.simple.ink/blog/notion-stats</h6><h6>[12]: Ness Labs, &#8220;Building the world&#8217;s most customizable workspace with Ivan Zhao,&#8221; April 2023. https://nesslabs.com/notion-featured-tool</h6><h6>[13]: Sequoia Capital, ibid. (&#8221;They found a small group of really die-hard fans that carried the product with them.&#8221;)</h6><h6>[14]: Ness Labs, ibid. (&#8221;One of our ambassadors runs our subreddit with nearly 100,000 subscribers. Another runs our Facebook Group in Korea with over 20,000 folks. Another has an Arabic community with over 30,000 folks.&#8221;)</h6><h6>[15]: Nira, &#8220;How Ivan Zhao&#8217;s Notion Is Going After Atlassian and Why It Just Might Win,&#8221; March 2020. https://nira.com/notion-history/</h6><h6>[16]: The Launcher, &#8220;Notion - Reaching 1 million users with 18 employees,&#8221; May 2021. https://thelauncher.substack.com/p/notion-reaching-1-million-users-with</h6><h6>[17]: Lenny&#8217;s Podcast, &#8220;Notion&#8217;s lost years, its near collapse during Covid,&#8221; March 2025. https://www.lennysnewsletter.com/p/inside-notion-ivan-zhao (transcript: &#8220;there&#8217;s a Doomsday Clock that when we&#8217;re going to truly run out of this space to store everything in Notion&#8221;)</h6><h6>[18]: Lenny&#8217;s Podcast, ibid.</h6><h6>[19] Notion Blog, &#8220;Introducing Notion AI,&#8221; February 2023.<a href="https://www.notion.com/releases/2023-02-22"> https://www.notion.com/releases/2023-02-22</a> </h6><h6>[20] Kipwise, &#8220;Notion AI Features &amp; Capabilities,&#8221; 2025.<a href="https://kipwise.com/blog/notion-ai-features-capabilities"> https://kipwise.com/blog/notion-ai-features-capabilities</a> (&#8221;Unlike standalone AI tools that require constant switching between applications, Notion AI understands your page structures, database relationships, and project contexts&#8221;) </h6><h6>[21] PodCosmos, &#8220;How Notion Reimagined Productivity Tools | Ivan Zhao,&#8221; August 2025.<a href="https://www.podcosmos.com/kleiner-perkins/grit-podcast/how-notion-reimagined-productivity-tools-ivan-zhao"> https://www.podcosmos.com/kleiner-perkins/grit-podcast/how-notion-reimagined-productivity-tools-ivan-zhao</a> (&#8221;Consolidate Before You Automate - Build integrated tool foundations before attempting AI agent implementation&#8221;) </h6>]]></content:encoded></item><item><title><![CDATA[Enhancing Trust in AI Systems]]></title><description><![CDATA[Trust has quietly become the real bottleneck in AI adoption.]]></description><link>https://www.builderlab.ai/p/enhancing-trust-in-ai-systems</link><guid isPermaLink="false">https://www.builderlab.ai/p/enhancing-trust-in-ai-systems</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 19 Oct 2025 09:25:13 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/83cee87f-2667-47ba-ac3e-089a879a9a46_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Trust has quietly become the real bottleneck in AI adoption. Not compute. Not model size. Not latency. Trust.</p><p>The last two years have shown what AI can do. What comes next will depend on whether people believe it will do <em>what they expect it to do</em>, and nothing more. We&#8217;ve reached the stage where capability outpaces confidence. AI systems can reason, generate, and act, but too often they behave like overeager interns: smart, unpredictable, and hard to supervise.</p><p>If you&#8217;re building products with AI today, your biggest challenge isn&#8217;t getting better models. It&#8217;s <strong>engineering trust</strong> by design.</p><div><hr></div><h3><strong>The Real Trust Gap</strong></h3><p>When people say they don&#8217;t trust AI, they&#8217;re rarely questioning the math. They&#8217;re questioning <em>the behavior</em>.</p><p>This erratic behavior manifests in strange ways: a model writes perfect code in one moment, then refuses to recognize its own syntax five minutes later. A chatbot answers 100 customer queries flawlessly, then invents policy violations on the 101st. Or the ranking algorithm learns too aggressively, pushing click-bait and driving short-term engagement.</p><p>Technically, these systems are &#8220;working.&#8221; Statistically, their aggregate metrics look fine. Psychologically, they&#8217;ve failed.</p><p>In classical software, you can earn trust once and coast on it for years. With AI, you have to re-earn it daily. AI is probabilistic at its core. The same input doesn&#8217;t always yield the same output. To AI-native engineers, that&#8217;s normal variation. To users, it feels like instability.</p><p>One early GenAI customer-support pilot showed this perfectly. The bot handled 85 percent of questions without issue, but the remaining 15 percent were catastrophic contradictions: apologizing for errors that hadn&#8217;t happened, offering refunds that didn&#8217;t exist. The result wasn&#8217;t 85 percent satisfaction. It was <em>far lesser</em>, because one unpredictable failure erases the memory of many reliable moments.</p><p>The uncomfortable truth is that <em>trust is not proportional to capability</em>. Humans trust cars more than planes, even though planes are statistically safer, because they &#8216;feel&#8217; more in control in a car. The same principle applies to AI: people are more comfortable with a less capable system that they control than with a powerful one that they don&#8217;t.</p><div><hr></div><h3><strong>What Trust Actually Looks Like</strong></h3><p>Trust in AI is built through <em>calibration</em>: the alignment between what the user expects and what the system actually does. A self-driving car doesn&#8217;t need the driver to marvel at its intelligence. It needs to react in expected ways when it sees a stop sign or a child on the street. The same is true for AI systems.</p><p><strong>Predictable beats perfect</strong>: <em>Predictability creates permission.</em> When users can anticipate the next step, they relax into the system&#8217;s rhythm.</p><p><strong>Consistent interaction patterns: </strong>The best AI interfaces teach users what to expect through repetition. They turn stochastic behavior into a stable mental model:<br>&#8220;Click here to get a different angle.&#8221;<br>&#8220;See sources here to verify.&#8221;<br>&#8220;Expect four variations.&#8221;</p><p><strong>Explain enough, not everything: </strong>Another signal of trustworthy systems: <em>selective transparency.<br></em>Anthropic&#8217;s Claude often provides a short rationale (&#8220;Here&#8217;s why I think this answer fits&#8221;) instead of an overwhelming technical trace. That small explanation helps the user calibrate confidence without drowning them in details.</p><p>Too much explanation breeds fatigue; too little breeds suspicion. The art lies in showing just enough reasoning to make the system <em>feel steady</em>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoying this post? Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h3><strong>Three Layers of Trust</strong></h3><p>When you look at systems people truly rely on - air traffic control, autopilot, or payment infrastructure - they all earn trust in layers. AI products should do the same.</p><h4>1. Predictability</h4><p>Users should be able to anticipate what happens next. The model should behave the same way for the same inputs, and make its limits clear when it can&#8217;t.</p><p>In GenAI, unpredictability usually comes from <em>context drift</em>, or small variations in prompts or hidden memory states that change output meaningfully. This is why even capable systems lose user confidence: the same request feels &#8220;off&#8221; on a different day.</p><p>The best builders fix this not by chasing perfection, but by making behavior consistent.</p><ul><li><p><strong>ChatGPT&#8217;s &#8220;Custom Instructions&#8221;</strong> feature lets users explicitly define tone and context, anchoring the model&#8217;s behavior across sessions.</p></li><li><p><strong>Midjourney&#8217;s numbered variation grid</strong> teaches users that creative exploration has boundaries. The randomness becomes <em>structured discovery</em>, not chaos.</p></li></ul><p>Predictability gives users a stable mental model. They stop testing the system and start using it.</p><h4>2. Reversibility</h4><p>People take more risks when they know they can recover.<br>A reversible system is one that says: &#8220;Try this. If it fails, you can go back.&#8221;</p><p>In GenAI, that usually means <strong>containing the blast radius</strong> of automation.<br>GitHub Copilot doesn&#8217;t push to main. Rather, it suggests code inline, where the human remains the gatekeeper. Notion AI and Canva&#8217;s &#8220;Undo&#8221; and &#8220;Restore&#8221; options let users roll back any generated content.</p><p>When autonomy is paired with reversibility, users shift from fear to experimentation. They start asking, <em>What can this system do for me?</em> instead of <em>What could it mess up?</em></p><h4>3. Proof of Competence</h4><p>Trust comes from <em>small proofs</em>.</p><p>Before an AI system asks for full autonomy, it needs to earn credibility in low-stakes contexts.<br>Copilot earned developer trust by finishing lines, not writing files.<br>Notion AI started with rewriting and summarizing text, not running entire workflows.<br>Both systems built competence in narrow tasks, then expanded scope once reliability was visible.</p><p>Each visible success reinforces the belief that the system <em>knows its job</em>.</p><h4>Stacking the layers</h4><p>Predictability builds confidence.<br>Reversibility builds courage.<br>Proof of competence builds commitment.</p><p>Together, these layers create the scaffolding that lets intelligence scale without anxiety.</p><p>When users stop wondering <em>if</em> the AI will work and start focusing on <em>how best to use it</em>, trust has been achieved.</p><div><hr></div><h3><strong>Principles for Building Trust in AI Systems</strong></h3><p>Trust emerges from design. These principles translate directly into product choices that shape how people experience AI reliability day to day:</p><h4>1. Make behavior visible</h4><p>Users trust what they can see. Every action an AI takes, for example, what data it used, why it acted, and what it produced, should be observable. The goal isn&#8217;t radical transparency though; it&#8217;s <em>relevant transparency</em>.</p><p><strong>Example:</strong> Perplexity AI&#8217;s citation cards let users trace every claim back to a source. That single UX choice converts what would have been a &#8220;trust me&#8221; response into an evidence-based one.</p><p><strong>Insight:</strong> Visibility turns a black box into a glass box. When users can peek inside, even briefly, uncertainty drops and confidence rises.</p><h4>2. Build for reversibility</h4><p>Confidence grows when people know they can undo. Every AI action should carry an implied &#8220;escape hatch.&#8221; <br><strong>Example:</strong> In Notion AI, every edit has an immediate &#8220;Undo&#8221; state. It encourages experimentation because users know they can roll back. The same principle underpins GitHub Copilot&#8217;s inline suggestions: nothing executes without human review.</p><p><strong>Insight:</strong> Reversibility transforms fear into curiosity. When users can recover, they engage more deeply.</p><h4>3. Keep explanations human</h4><p>Clarity beats completeness.<br>Good explanations don&#8217;t unpack neural math; they help users decide what to do next. The test of a useful explanation isn&#8217;t accuracy, it&#8217;s <em>actionability</em>.</p><p><strong>Example:</strong> Anthropic&#8217;s Claude often prefaces responses with brief context (&#8220;Here&#8217;s how I approached this&#8230;&#8221;). It signals reasoning without dragging users into technical weeds.</p><p><strong>Insight:</strong> The point isn&#8217;t to prove the model&#8217;s intelligence. It&#8217;s to make its logic legible in human terms.</p><h4>4. Design safety margins</h4><p>A good system knows when to stop.<br>Overconfidence kills trust faster than errors do. When an AI isn&#8217;t sure, it should defer, clarify, or escalate.</p><p><strong>Example:</strong> Copilot for Microsoft 365 regularly prompts, &#8220;Would you like me to continue?&#8221;, a small checkpoint that prevents runaway generation. OpenAI&#8217;s function-calling APIs similarly force boundaries between thinking and doing.</p><p><strong>Insight:</strong> Predictable failure beats unpredictable success.</p><h4>5. Show a consistent track record</h4><p>Reliability compounds.<br>A system that behaves predictably builds more trust than one that dazzles sporadically. Users remember consistency far longer than brilliance.</p><p><strong>Example:</strong> ChatGPT&#8217;s monthly updates quietly improved memory, formatting, and speed without breaking prior behavior. That rhythm of steady improvement signals maturity, not volatility.</p><p><strong>Insight:</strong> Every stable interaction becomes a micro-deposit in the user&#8217;s trust account. Over time, those deposits turn into loyalty.</p><div><hr></div><h3><strong>The Economics of Trust</strong></h3><p>Trust is an economic principle. When users trust a system, they delegate more, churn less, and experiment faster. When they don&#8217;t, everything slows down: adoption, retention, and even internal confidence within the team.</p><p><strong>Predictability drives speed: </strong>Teams that invest in predictability ship faster.<br>Because when the system behaves consistently, engineers spend less time firefighting edge cases, and product managers spend less time negotiating stakeholder anxiety.</p><p>Predictable systems scale faster inside organizations because every function&#8212;legal, design, marketing&#8212;knows what to expect.</p><p><strong>Reversibility increases retention: </strong>Users stay longer when they can recover from mistakes. Reversibility changes how people <em>feel</em> about using AI: from cautious testing to active trust. In AI products, the ability to reverse is the permission to adopt.</p><p><strong>Reliability accelerates enterprise adoption: </strong>Enterprises don&#8217;t buy &#8220;creativity.&#8221; They buy <em>predictable ROI. </em>A model that is slightly less capable but measurably more stable will always win procurement battles. Reliability reduces perceived implementation risk, which shortens sales cycles. Reliability converts hesitation into contracts. The more visible your guardrails, the faster trust turns into revenue.</p><div><hr></div><h3><strong>Building Trust Is Building Product</strong></h3><p>In traditional software, trust is implicit. Users click a button and assume the system will do what it says. That&#8217;s because software has always lived in a world of <em>determinism</em>: same input, same output, no surprises.</p><p>AI breaks that contract.</p><p>Each new layer of autonomy - every agent that acts, every model that remembers, every chain that calls another API - reopens the negotiation between user and system. &#8220;Do I still know what this thing will do next?&#8221; becomes an everyday question.</p><p>This means that <strong>trust isn&#8217;t an afterthought in AI products; it </strong><em><strong>is</strong></em><strong> the product</strong>.</p><p>In deterministic software, you design <em>features</em>. In AI systems, you design <em>behavior</em>. Every prompt template, fallback rule, and safety margin is part of the product&#8217;s personality. And users form relationships with that personality, not the underlying model weights. When a chatbot calmly admits it doesn&#8217;t know an answer instead of hallucinating, that&#8217;s a product decision. When an image generator previews before committing compute, that&#8217;s a trust decision. When an agent asks for confirmation before executing a purchase, that&#8217;s a product <em>and</em> trust decision. These moments shape the emotional contract between user and system.</p><p>Every log, every preview, every &#8220;undo&#8221; button is part of that architecture. So are things users never see: audit trails, confidence thresholds, temperature settings, moderation layers. Together, they form what might be called a <strong>trust stack</strong> - the invisible scaffolding that lets intelligence operate safely at scale. OpenAI&#8217;s memory control panel, Anthropic&#8217;s constitutional layers, and Apple&#8217;s on-device intelligence boundaries are all versions of this same idea: transparency, containment, reversibility.<br>They differ in implementation, but not in philosophy.</p><p>Trust is a habit you earn through repetition. Every small proof of competence - a correct response, a graceful fallback, a transparent explanation - is a deposit in the trust account. Over time, those deposits accumulate into something powerful: <em>delegation</em>. Once users delegate, they stop questioning the system&#8217;s right to act.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/enhancing-trust-in-ai-systems?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Think this post will be useful for someone you know? Share with your colleagues!</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/enhancing-trust-in-ai-systems?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/enhancing-trust-in-ai-systems?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><h3><strong>Closing Thought</strong></h3><p>The future of AI will be won by the most <em>trusted</em> systems. Intelligence gets you in the door. But trust is what turns pilots into production, skeptics into advocates, and one-time users into people who can&#8217;t imagine working without your system.</p><p>The uncomfortable reality is that we&#8217;ve been optimizing for the wrong thing. We&#8217;ve been in an arms race for capability when the real constraint is confidence. The breakthrough products of the next decade won&#8217;t be the ones that can do more; they&#8217;ll be the ones people are willing to let do more.</p><p>That&#8217;s the unlock trust provides: permission to delegate. Delegation is how AI actually changes work. Everything else is just a demo.</p><p><strong>The systems that admit uncertainty will earn more trust than the ones that fake omniscience. </strong>We&#8217;ve spent two years teaching AI to be confident. The next two years will be about teaching it to be honest about its limits, its uncertainty, its reasoning. Not because transparency is virtuous, but because it&#8217;s the only way to build systems people can actually rely on.</p><p>Intelligence is abundant now. What&#8217;s scarce is the courage to constrain it.</p><p>The builders who understand this, those who design for predictability over performance, who build undo buttons before building autonomy, who earn trust through small proofs rather than big promises, won&#8217;t just build better products.</p><p>They will define how the world learns to work with machines.</p>]]></content:encoded></item><item><title><![CDATA[User Memory as a Product Delighter]]></title><description><![CDATA[Remembering (and Forgetting) What Matters]]></description><link>https://www.builderlab.ai/p/user-memory-as-a-product-delighter</link><guid isPermaLink="false">https://www.builderlab.ai/p/user-memory-as-a-product-delighter</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sat, 04 Oct 2025 20:00:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e613b14b-4735-4c0a-b133-7469ce8d7a39_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Amazon attributes <strong>$100B+ in revenue</strong> to personalization. Spotify users discover <strong>1.9B new tracks every day</strong> through recommendations. TikTok&#8217;s <strong>1B+ active users</strong> swear by their personalized feed.What makes it feel magical? User memory. When memory works, products feel alive, perfectly attuned to you.</p><p>Long before GenAI, this &#8220;memory&#8221; lived in click streams, purchase histories, engagement trails, cancellations, likes, and countless other signals. These data points still drive some of the most powerful algorithms we have. Ranking and recommendation systems remain the beating heart of e-commerce and social, and will be for a while.</p><p>What&#8217;s changed is the arrival of conversational AI. Memory has found a voice. Free-form dialogue has unlocked a new feature factory, producing richer, more nuanced signals about customers than we&#8217;ve ever had before. And for the first time, users themselves can step in. They can say what to remember, what to forget, what to add, remove, or modify.</p><p>That shift makes memory visible. It&#8217;s interactive. It&#8217;s user-driven. And when done well, it feels alive. Stateful, elegant, powerful.</p><p>So the builder&#8217;s question becomes:<em><strong> how do you design memory not just as infrastructure, but as a source of delight?</strong></em><strong><br></strong></p><div><hr></div><h2><strong>Memory isn&#8217;t one thing</strong></h2><p>It&#8217;s tempting to talk about memory as if it were a single bucket: data goes in, user profile comes out. In practice, memory behaves more like layers with purpose, failure modes, and design choices.</p><p><strong>1. Episodic (in-session, short-lived) Memory: </strong>The cart you just filled, the search you ran five minutes ago, the episode you left half-watched, or the conversation you&#8217;re still having. Episodic memory makes the product feel present, like it&#8217;s paying attention right now. If you stretch it too far, it becomes irritating. Think of a travel site still showing you flights long after you&#8217;ve booked them. That&#8217;s episodic memory clinging on when it should have let go.</p><p><strong>2. Medium term (weeks to months, patterns): </strong>Seasonal trips, listening habits, food orders that repeat just often enough to feel like a ritual. Medium-term memory makes personalization feel contextual. Done well, it fades gracefully, giving more weight to what&#8217;s recent while not erasing the past. Done poorly, it becomes stale, trapping users in a loop of yesterday&#8217;s choices.</p><p><strong>3. Long term (identity anchors): </strong>Dietary preferences, languages, loyalty tiers, home cities. They don&#8217;t change often, and when they do, they need to be updated with care. Long-term memory builds trust when it&#8217;s right, and breaks it when it&#8217;s wrong. A system that keeps treating someone as an omnivore after they&#8217;ve changed diets doesn&#8217;t just feel outdated; it feels disrespectful.</p><p>Every layer has its own role. Episodic makes a product feel responsive. Medium-term makes it feel contextual. Long-term makes it feel trustworthy. Together, they form the backbone of personalization. Treat them as one, and you risk breaking the experience. Treat them as layers, and you can design memory that feels alive.</p><blockquote></blockquote><div><hr></div><h2><strong>How to think about memory architectures</strong></h2><p>If memory is layered, the architecture has to reflect it. It&#8217;s not enough to collect signals, you need a model for where they live, who can use them, and when they should move or fade.</p><p><strong>Storage:</strong> Episodic memory belongs in fast, short-lived stores. Medium-term memory sits in feature stores, event streams, or embeddings that decay. Long-term anchors go into profile systems that act like contracts: typed, versioned, auditable.</p><p><strong>Access:</strong> Decide who can read and write each layer. Episodic signals may feed a ranking model but shouldn&#8217;t leak beyond the session. Medium-term patterns can power recommendations or filters. Long-term anchors demand explicit user control and visibility.</p><p><strong>Transfer.</strong> Memory should move with intent. Episodic signals that repeat can become medium-term patterns. Medium-term habits that persist can graduate into long-term anchors. Just as important, old signals must decay or be pruned. Forgetting is part of the design.</p><p><strong>Governance.</strong> Metadata, lineage, and &#8220;why shown&#8221; traces make memory explainable. Consent and retention rules make it trustworthy. Without them, memory may work technically, but it won&#8217;t build confidence.</p><p>Signals flow, fade, or graduate with purpose. The system explains itself simply, and trust is built in from the start.</p><div><hr></div><h2><strong>When to remember, when to forget</strong></h2><p>User&#8217;s preferences, tastes, and wants evolve over time, or sometimes even from session to session. Memory needs to evolve with this change. The design challenge is knowing when to hold on, when to promote, and when to clear the slate.</p><p><strong>Decay</strong>: Medium-term signals should fade over time so fresh behaviour outweighs stale history. Last night&#8217;s pizza order should dominate today&#8217;s suggestions, but it shouldn&#8217;t define them a week from now.</p><p><strong>Conflict resolution</strong>: When facts clash, the system needs deterministic rules. Recency can take precedence, or source authority, or versioning. What matters is consistency: the system should never be uncertain about which fact wins.</p><p><strong>Event-driven forgetting</strong>: Once a purchase is complete, the cart should clear. Once a trip is booked, stop pushing the same flights and shift to hotels or activities. Once a refund is processed, remove the grievance from active memory. These resets prevent memory from becoming noise.</p><p><strong>Explicit control</strong>: Small gestures like &#8220;forget this,&#8221; &#8220;don&#8217;t recommend again,&#8221; &#8220;dismiss,&#8221; or &#8220;clear history&#8221; build trust. Preference management for long-term anchors signal respect for identity.</p><p><strong>Layer transfers</strong>: Repeated episodic signals can be promoted into medium-term patterns. Medium-term habits that prove stable over months can become long-term anchors. Sensitive or one-off events should never be promoted. Cool-off windows can prevent recency noise from hardening into identity too quickly.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoying this breakdown? Join 2,000+ builders who subscribe for essays like this every week.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><blockquote></blockquote><div><hr></div><h2><strong>User interaction with memory (conventional UX + conversational)</strong></h2><p>Memory should be mostly invisible, but offer graceful points of control when users want them.</p><p>Some of those controls are familiar. <strong>Feedback loops</strong> like thumbs up, thumbs down, star ratings, or a quick &#8220;not relevant&#8221; let people steer recommendations without heavy effort. <strong>Dismiss and skip</strong> options like &#8220;not now&#8221; or &#8220;don&#8217;t show again&#8221; give users a lightweight way to prune memory in the flow of use.</p><p>Others are about choice and transparency. <strong>Filters and controls</strong> let people actively shape medium-term memory: genres in streaming, cuisines in food apps, price or dietary tags in e-commerce. <strong>Consent and confirmation</strong> are essential for long-term anchors. An opt-in at the start, and an occasional nudge later: &#8220;Still vegetarian?&#8221; &#8220;Still want us to track direct flights?&#8221;</p><p>Sometimes people want a clean slate. <strong>Reset options</strong> like &#8220;clear history,&#8221; &#8220;reset recommendations,&#8221; or &#8220;start fresh&#8221; may be rarely used, but their presence signals respect and control. Paired with <strong>explainability</strong>, for example with simple reasons like &#8220;Because you watched&#8230;&#8221; or &#8220;Based on your last 3 orders&#8230;&#8221;, they make memory feel less like surveillance and more like a partnership.</p><p>In conversational systems, you can go further. Lightweight <strong>hooks</strong> like &#8220;remember this&#8221; or &#8220;forget that&#8221; let users treat memory as part of the dialogue itself.</p><p>Balance matters. Too much user management, and personalization becomes work. Too little, and memory feels opaque or even manipulative. The sweet spot is memory that mostly stays out of sight, but surfaces just enough for users to shape it when they want.</p><blockquote></blockquote><div><hr></div><h2><strong>Metrics, experiments, and QA</strong></h2><p>If memory is going to be a first-class product lever, it needs to be measured like one. That means moving beyond engagement spikes and looking at how memory shapes both performance and trust.</p><p><strong>The usual product KPIs</strong>: click-through and engagement lift, conversion and repurchase rates, even measures of novelty versus repetition. Just as important are <em>negative signals</em>: complaint rates or drops in usage when personalization feels stale or clingy.</p><p><strong>Trust signals</strong>: How often do people use the memory controls you&#8217;ve provided? Do they clear history, update preferences, click on &#8220;why shown&#8221; explanations? Opt-in and opt-out rates are more than compliance artifacts. They&#8217;re direct indicators of whether users feel safe with your memory design.</p><p><strong>Quality diagnostics</strong>: A repetition index can flag when recommendations loop too tightly. A stale-rec rate shows how often expired signals are resurfacing. Post-completion spam, like pushing flights after booking, should be measured and eliminated. Conflict error rates, where the system holds contradictory facts, can be tracked and brought down.</p><p><strong>Experimentation</strong>: A/B testing decay schedules, promotion thresholds, or conflict resolution policies can reveal how subtle changes affect user experience. Explainability variants - different ways of telling users &#8220;why&#8221; something is shown - are also ripe for experimentation. For long-term anchors, running holdouts help measure accuracy and trust over time.</p><p><strong>Rigorous testing</strong>: Unit tests for promotion and pruning logic. Synthetic edge cases that simulate rapid preference flips. Safety checks to ensure that sensitive attributes never get promoted where they don&#8217;t belong.</p><p>Memory is product behaviour. And like any behaviour, it needs KPIs, experiments, and QA baked into the process.</p><blockquote></blockquote><div><hr></div><h2><strong>Cases in the wild</strong></h2><p>There are products today that show memory done right.</p><p>In <strong>e-commerce</strong>, Amazon&#8217;s &#8220;Frequently bought together&#8221; is a smart use of medium-term memory. It doesn&#8217;t stop at the purchase; it pivots into the next logical step. The best platforms also remember stock status and policy versions so recommendations don&#8217;t feel out of sync.</p><p>In <strong>streaming</strong>, Spotify&#8217;s Discover Weekly is at its best when it balances novelty with anchors, expanding your taste while grounding it in what you&#8217;ve loved before. Netflix&#8217;s &#8220;Because you watched&#8230;&#8221; explanations, when clear, make recommendations feel less like a black box and more like a natural extension of your history.</p><p>In <strong>travel</strong>, loyalty systems are a bright spot. Airlines and OTAs that reliably remember your tier, preferred seat type, or favorite destinations build trust that compounds over years. These long-term anchors are exactly where memory should shine.</p><p>In <strong>productivity</strong>, small touches like Gmail surfacing &#8220;drafts you started&#8221; or Notion letting you pin active projects make episodic memory feel helpful, not intrusive. The best tools make it easy to focus on what&#8217;s current, while still letting you recover past context when you need it.</p><p>The gaps are just as visible. E-commerce platforms still recommend items already purchased, instead of pivoting to complements. Streaming apps often over-personalize, looping users endlessly in a single genre. Travel sites continue to surface the same flights even after booking, when the intent should reset immediately. And productivity apps too often drown you in irrelevant recall, surfacing projects that should have decayed long ago.<br><br></p><blockquote></blockquote><div><hr></div><h2><strong>&#129504; Summary: The Builder&#8217;s Memory Audit</strong></h2><ul><li><p>&#9989; Separate <strong>episodic, medium-term, and long-term</strong> memory.</p></li><li><p>&#9989; Design signals to <strong>decay, reset, or promote</strong> with intent (purchase clears cart, seasonal resets, stable patterns graduate).</p></li><li><p>&#9989; Resolve conflicts with a <strong>deterministic policy</strong> (recency, authority, versioning).</p></li><li><p>&#9989; Give users the ability to <strong>see, edit, reset, or confirm</strong> what&#8217;s remembered.</p></li><li><p>&#9989; Provide simple <strong>explanations</strong> (&#8220;Because you watched&#8230;&#8221;) so recommendations aren&#8217;t a black box.</p></li><li><p>&#9989; Make memory a <strong>delighter</strong>: stateful, fading elegantly, giving agency without extra work.</p></li><li><p>&#9989; Measure the <strong>right KPIs</strong>: novelty vs repetition, staleness, post-completion spam, conflict error rate, trust signals (use of controls, clear-history, opt-ins).</p></li></ul><blockquote></blockquote><div><hr></div><h2><strong>What&#8217;s next</strong></h2><p>My speculation: the next wave will stretch beyond today&#8217;s click streams and conversation logs.</p><p><strong>Multimodal</strong>: Text alone won&#8217;t be enough. Signals from images, voice, and location will be captured and stitched together.</p><p><strong>On-device</strong>: Users will expect faster, more private recall. Local storage, combined with federated updates, will give products the ability to personalize without shipping raw data back to the cloud. That shift will blur the line between device and service.</p><p><strong>Team-level</strong>: In collaborative tools, memory won&#8217;t stop at the individual. It will need to represent shared context like project histories, group decisions, team preferences with clear consent scopes. Who owns the memory becomes as important as what&#8217;s remembered.</p><p><strong>Policy-awareness: </strong>region, age, and sensitivity rules will need to be baked into promotion logic. A one-off crisis, a protected attribute, or data from a restricted geography can&#8217;t be allowed to flow upward into long-term anchors. Governance won&#8217;t sit outside the architecture; it will live inside it.</p><p>All in all, memory will remain an evolving core system across products. It will shape user interactions, and be shaped by them. New paradigms for storage, context, and interaction will continue to augment and evolve our thinking.</p><p><em><strong>I&#8217;m curious, How do you manage memory in what you&#8217;re building? What thoughts does this evoke for you? Does all of this feel like an exciting design opportunity, or does it seem too complex to implement in practice?</strong></em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If this sparked ideas, subscribe to BuilderLab. Every week we unpack one product lever like this: frameworks, cases, and future signals</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Approximating a World Model - The Builder's Guide to Stateful AI]]></title><description><![CDATA[Yann LeCun is pushing the field towards world models.]]></description><link>https://www.builderlab.ai/p/approximating-a-world-model-the-builders</link><guid isPermaLink="false">https://www.builderlab.ai/p/approximating-a-world-model-the-builders</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sat, 27 Sep 2025 06:45:35 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5272c0b1-2201-4702-ad7c-7430042eba10_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Yann LeCun is pushing the field towards world models. And he is right: scaling LLMs alone will not get us to systems that understand, reason, and plan.</p><p>But most builders cannot afford to wait for JEPA v2 papers or Meta&#8217;s research breakthroughs. The question is practical: <strong>how do you approximate a world model today, with the tools already available?</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>What a World Model Really Is</strong></h2><p>Strip away the academic framing, and a world model is not mysterious. It is three ingredients:</p><ol><li><p><strong>State<br></strong> A structured representation of what the world looks like now.</p></li><li><p><strong>Transition<br></strong> A way to update that state when something changes, whether through new observations or actions.</p></li><li><p><strong>Planning<br></strong> The ability to project the state forward and evaluate possible actions against future outcomes.</p></li></ol><p>That is it. A toddler does this instinctively. Builders can do it with databases, rules, and lightweight reasoning loops. You don&#8217;t need a perfect latent predictor to get value. You just need to build the loop.</p><div><hr></div><h2><strong>Where Builders Go Wrong With LLMs</strong></h2><p>The default instinct today is to treat the LLM itself as the &#8220;world model.&#8221; Push more into the prompt, expand the context window, and hope the model &#8220;remembers&#8221; enough to simulate state.</p><p>This breaks in practice:</p><ul><li><p><strong>Memory</strong> becomes brittle. Context stuffing is expensive and inconsistent.</p></li><li><p><strong>State drift</strong> creeps in. The model hallucinates facts that no longer match reality.</p></li><li><p><strong>Actions</strong> lose grounding. The LLM proposes fluent answers that fail in a dynamic environment.</p></li></ul><p>A real-world example: task management agents that forget what tasks are complete, or booking bots that &#8220;assume&#8221; availability without checking the underlying system.</p><p>The fix is to separate <strong>language interface</strong> from <strong>world state</strong>.</p><div><hr></div><h2><strong>Approximating World Models With What You Have</strong></h2><p>Here are four practical ways to approximate a world model today:</p><p>&#128313; <strong>Entity and state memory<br></strong> <em>Keep a structured store of entities, attributes, and goals. It can be a SQL database, a key-value store, or a vector DB with typed embeddings. The key is persistence across sessions.</em></p><ul><li><p><strong>Customer support bot</strong>: Store user identity, open tickets, and previous resolutions. The system recalls a past complaint automatically instead of asking the customer to repeat.</p></li><li><p><strong>Fitness app</strong>: Maintain state for weight, calorie intake, and workout history. When a user logs breakfast, the app updates remaining daily macros without recalculation from scratch.</p></li><li><p><strong>Travel planner</strong>: Remember destinations, dates, and budget goals across sessions. If the user returns a week later, the system still knows they were planning a 7-day trip to Spain under &#8364;2k.</p></li></ul><div><hr></div><p>&#128313; <strong>Simple transition functions<br></strong> <em>You don&#8217;t need deep learning here. Rule-based updates work. If a user confirms a booking, mark their state as &#8220;confirmed.&#8221; If an item disappears from the inventory feed, remove it from the map. These simple transitions bring stability.</em></p><ul><li><p><strong>E-commerce</strong>: When inventory hits zero, the system updates product state to &#8220;out of stock.&#8221; The bot no longer suggests it in recommendations.</p></li><li><p><strong>Banking app</strong>: When a transfer is executed, the balance is reduced immediately. The state reflects reality before the next sync with the core system.</p></li><li><p><strong>Learning app</strong>: When a student completes a module, mark the concept as &#8220;mastered&#8221; and unlock the next lesson.</p></li></ul><div><hr></div><p>&#128313; <strong>LLM-assisted reasoning<br></strong> <em>Use the LLM tactically. It can propose candidate actions, but always ground its reasoning against the state memory. Example: &#8220;Given the current state and constraints, which option satisfies the goal?&#8221;</em></p><ul><li><p><strong>Logistics</strong>: The world model tracks packages in transit. The LLM reasons about re-routing: &#8220;Given a delayed flight and the customer&#8217;s deadline, what delivery route satisfies the goal?&#8221;</p></li><li><p><strong>Healthcare triage</strong>: The state model has symptoms logged. The LLM proposes next questions to narrow possibilities but only within the structured medical knowledge base.</p></li><li><p><strong>Education</strong>: State = student proficiency map. LLM reasons: &#8220;Given weak algebra and strong geometry, which exercise is the best next step?&#8221;</p></li></ul><div><hr></div><p>&#128313; <strong>Simulation by replay<br></strong> <em>Planning does not require generative video models. You can approximate it by rolling forward &#8220;what if&#8221; scenarios with your transition functions. Test possible actions, then choose the one that leads to the best future state.</em></p><ul><li><p><strong>Ride-hailing</strong>: Simulate assigning a driver to passenger A vs passenger B. Roll forward travel times and see which choice minimizes total wait across the system.</p></li><li><p><strong>Personal finance app</strong>: Simulate savings outcomes. If the user invests &#8364;500 a month vs &#8364;1,000, replay the balance forward 10 years and compare results.</p></li><li><p><strong>Warehouse robotics</strong>: Simulate picking orders in sequence A-B-C vs C-B-A. Replay travel distances and pick times, then choose the shortest plan.</p></li></ul><div><hr></div><h2><strong>The Builder&#8217;s Playbook</strong></h2><p>Think of this as a minimum viable world model:</p><p><strong>Step 1. Define your schema<br></strong> Decide what your world consists of. Users, items, locations, states, goals. Make it explicit.</p><p><strong>Step 2. Implement transitions<br></strong> Keep them transparent. Rules that map input to updated state are enough to start.</p><p><strong>Step 3. Insert the LLM as a reasoner<br></strong> Let the LLM interpret ambiguous inputs and propose candidate actions. But never let it rewrite the state directly.</p><p><strong>Step 4. Add a planner<br></strong> Use simple replay. Simulate a few possible next steps, score them, and pick the best.</p><p><strong>Step 5. Log surprises<br></strong> Whenever the real world diverges from the model, log it. Surprises become your training signal for improvement.</p><div><hr></div><h2><strong>Examples Across Domains</strong></h2><p>This approach works across verticals:</p><ul><li><p><strong>Customer support<br></strong> World model = user profile + issue history + open tickets. Transitions = updates from user replies or system events. Planning = suggest next best resolution.</p></li><li><p><strong>E-commerce logistics<br></strong> World model = order states + inventory + shipping nodes. Transitions = order placed, item picked, package scanned. Planning = choose optimal route.</p></li><li><p><strong>Health tracking<br></strong> World model = biometrics + habits + goals. Transitions = new measurements logged. Planning = adjust nutrition or workout plan.</p></li><li><p><strong>Education apps<br></strong> World model = learner progress + mastered concepts + weak areas. Transitions = quiz results, completed lessons. Planning = recommend next module.</p></li></ul><p>In all cases, the pattern is the same: <strong>state, transition, planning.</strong></p><div><hr></div><h2><strong>Why This Works</strong></h2><p>Approximating a world model gives you four practical benefits:</p><ul><li><p><strong>Persistence</strong>: users feel the system remembers them.</p></li><li><p><strong>Grounding</strong>: actions are tied to real states, not hallucinations.</p></li><li><p><strong>Adaptation</strong>: surprises trigger learning loops instead of silent failures.</p></li><li><p><strong>Separation of concerns</strong>: LLMs handle language and reasoning, while structured systems handle reality.</p></li></ul><p>This is enough to turn fragile prototypes into reliable products.</p><div><hr></div><h2><strong>The Strategic Lens</strong></h2><p>LeCun&#8217;s JEPA will eventually give us elegant latent predictors. But builders cannot wait. Approximating a world model with structured memory and simple transitions is enough to start.</p><p>The pattern is not research-lab-only. It is the same loop product teams already know from good system design: <strong>model the state, update the state, plan over the state.</strong></p><p>The opportunity is to <strong>treat this as product architecture,</strong> rather than just prompt engineering.</p><div><hr></div><h2><strong>Closing Thought</strong></h2><p>You do not need Meta&#8217;s stack to build a useful world model. You can build one today with databases, rules, and an LLM stitched in as a reasoner.</p><p>&#128073; What state schema, if it persisted and evolved over time, would make your product instantly smarter?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for more techniques on building with AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why LeCun is Betting on World Models (and Why Builders Should Pay Attention)]]></title><description><![CDATA[At Nvidia&#8217;s GTC 2025, Yann LeCun said something that cut through the noise:]]></description><link>https://www.builderlab.ai/p/why-lecun-is-betting-on-world-models</link><guid isPermaLink="false">https://www.builderlab.ai/p/why-lecun-is-betting-on-world-models</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sat, 20 Sep 2025 14:49:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2de77b06-ba5a-4031-9488-a9581a9dd4a3_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>At Nvidia&#8217;s GTC 2025, Yann LeCun said something that cut through the noise:</p><blockquote><p><em>&#8220;I&#8217;m not so interested in LLMs anymore.&#8221;</em></p></blockquote><p><a href="https://www.youtube.com/watch?v=p1QXZHV4jkM&amp;list=WL&amp;index=8&amp;ab_channel=TheAIGRID">The AI GRID has a strong breakdown of his remarks</a>. The takeaway is clear: LLMs have reached the point of diminishing returns. Scaling them further is useful, but it will not unlock the next stage of intelligence.</p><div><hr></div><h2>The Four Frontiers</h2><p>LeCun outlined four capabilities that define the real challenge ahead:</p><ol><li><p><strong>Grounded understanding of the physical world</strong></p></li><li><p><strong>Persistent memory</strong></p></li><li><p><strong>Reasoning</strong></p></li><li><p><strong>Planning</strong></p></li></ol><p>These are the exact gaps builders encounter when trying to turn LLM demos into production-grade systems.</p><div><hr></div><h2>Where Tokens Break Down</h2><p>Tokens work well for language, but they do not map neatly to the continuous and unpredictable nature of the real world.</p><ul><li><p><strong>Discretization hides the richness of physical dynamics</strong><br>Breaking the world into fixed tokens oversimplifies reality. Physical processes are continuous and fluid, not neatly bucketed.<br><em>Example: A robot arm pouring liquid cannot be reduced to a finite set of &#8220;token&#8221; movements. The flow and slosh of water are continuous, and discretization misses those dynamics.</em></p></li><li><p><strong>Predicting pixels wastes resources on details that are unknowable</strong><br>Models spend capacity trying to guess noise, like the exact shape of someone&#8217;s face in the next video frame. The important structure gets lost in the effort.<br><em>Example: If a model tries to predict every pixel of a crowded street scene, it wastes energy guessing who will turn their head left or right, instead of focusing on the physics of cars moving through traffic.</em></p></li><li><p><strong>Next-token prediction does not translate into reliable next-action prediction</strong><br>Language fluency does not guarantee grounded decisions. Predicting the &#8220;right word&#8221; is not the same as predicting the &#8220;right move&#8221; in a real environment.<br><em>Example: An LLM can generate a perfect recipe description but cannot guarantee the cake will rise. The transition from text to action requires grounding in physical cause and effect.</em></p></li></ul><p>LeCun&#8217;s answer is <strong>Joint Embedding Predictive Architectures (JEPA / V-JEPA)</strong>. These architectures predict in <strong>latent space</strong>: abstract representations that capture what matters, ignore what does not, and reveal when the world violates learned patterns.</p><div><hr></div><h2>System 1 and System 2</h2><p>Psychologists describe two modes of thought:</p><ul><li><p><strong>System 1</strong>: fast, reactive, policy-driven</p></li><li><p><strong>System 2</strong>: slower, deliberate, planning-oriented</p></li></ul><p>LLMs approximate System 1. System 2 requires models that can manipulate abstract representations of the world and plan within them. That is the gap LeCun is targeting.</p><div><hr></div><h2>The Data Imbalance</h2><p>Consider the scale of human sensory input versus text:</p><ul><li><p>LLMs train on ~30 trillion tokens (~10^14 bytes)</p></li><li><p>A four-year-old child sees ~10^14 bytes through vision alone in just four years</p></li></ul><p>The conclusion is direct: text-only training cannot match the richness of multimodal world experience.</p><div><hr></div><h2>Implications for Builders</h2><p>This research agenda has immediate consequences for how we design products today:</p><p>&#129504; <strong>Memory as a product surface</strong><br>Create explicit, structured memory for entities, states, and goals.</p><p>&#9881;&#65039; <strong>Two operating loops</strong><br>Planning in latent space, then distilled policies for fast execution.</p><p>&#128202; <strong>Rethink evaluation</strong><br>Track latent prediction accuracy, surprise recovery, and policy efficiency instead of surface-level fluency.</p><p>&#128172; <strong>Tactical use of LLMs</strong><br>Use them where they excel: language, orchestration, and code generation. Connect them to models that represent real-world state and dynamics.</p><div><hr></div><h2>A 90-Day Blueprint</h2><p>You do not need a research lab to explore this direction.</p><p><strong>Weeks 1&#8211;2</strong><br>Select a task with real dynamics. Define a compact state representation.</p><p><strong>Weeks 3&#8211;6</strong><br>Train a self-supervised encoder and predictor to model transitions. Measure plausibility rather than pixel fidelity.</p><p><strong>Weeks 7&#8211;10</strong><br>Layer in a planner to act over the latent space. Track when and how the system recovers from surprises.</p><p><strong>Weeks 11&#8211;12</strong><br>Distill the planner into a lightweight policy. Add a &#8220;surprise guard&#8221; that triggers re-planning when prediction errors spike.</p><div><hr></div><h2>The Bigger Picture</h2><p>LeCun&#8217;s shift away from LLMs signals where the field is heading. Language interfaces will remain essential. Progress depends on systems that can model the world, remember, reason, and plan.</p><p>The open challenge for builders:<br>&#128073; Where in your product could latent-space planning outperform a larger prompt?</p><h2>How to Spot If Your Product Is Hitting LLM Limits</h2><p>If you are building with LLMs today, here are clear signs you are running into structural limits rather than prompt tuning problems:</p><ol><li><p><strong>Forgetting across sessions</strong><br>Users expect the system to remember entities, states, and goals. Your product relies on hacks like re-prompting or long context windows.</p></li><li><p><strong>Slow or expensive reasoning</strong><br>Tasks require multi-step reasoning chains, but the model burns cost and latency on re-generating context every time.</p></li><li><p><strong>Unreliable action prediction</strong><br>The model produces fluent text but fails to reliably trigger the correct action in dynamic or stateful environments.</p></li><li><p><strong>Evaluation mismatch</strong><br>Your metrics focus on fluency or token accuracy, but users care about planning quality, adaptation speed, and success rates.</p></li><li><p><strong>Scaling without leverage</strong><br>Bigger prompts and bigger models give marginal gains. What you need is structured state, not more tokens.</p></li></ol><div><hr></div><p>If you see these patterns, it is time to rethink your architecture. Keep the LLM as the language interface, but pair it with a <strong>world model</strong> that can represent state, reason in latent space, and plan actions. That is the step change LeCun is pointing towards.</p><p><strong>In my next post,</strong> I will go deeper into practical patterns: concrete examples of how to approximate world models today using four simple building blocks, for example, entity and state memory, transition functions, LLM-assisted reasoning, and simulation by replay. Each with real use cases you can start testing immediately.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Testing Something New: A Lightning Lesson on AI Strategy]]></title><description><![CDATA[A while ago I wrote that I&#8217;d be building a course on AI Strategy for Product Leaders.]]></description><link>https://www.builderlab.ai/p/testing-something-new-a-lightning</link><guid isPermaLink="false">https://www.builderlab.ai/p/testing-something-new-a-lightning</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Wed, 03 Sep 2025 06:55:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BoiJ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0e09a97-71ce-40b8-85cd-9976fe218d89_268x268.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A while ago I wrote that I&#8217;d be building a course on AI Strategy for Product Leaders. Before launching the full program, I&#8217;m starting small.</p><p>I&#8217;ve put together a short 30 minute lightning lesson to test the waters. It&#8217;s focused on a skill I think every product leader will need: evaluating which AI opportunities are actually worth pursuing.</p><p>We&#8217;ll go through the PROVE-IT framework &#8212; a structured way to score ideas on pain, reach, owned data, verifiability, execution loops, integration time, and cost. The goal is to help you cut through noise and focus on bets that can actually create impact.</p><p>If you&#8217;d like to join, you can sign up <a href="https://maven.com/p/6cd972/spot-the-right-ai-opportunities?utm_campaign=NTk4MjU3&amp;utm_medium=ll_share_link&amp;utm_source=instructorhttps://maven.com/p/6cd972/spot-the-right-ai-opportunities?utm_campaign=NTk4MjU3&amp;utm_medium=ll_share_link&amp;utm_source=instructor">here</a>.</p><p>As always, I&#8217;ll share learnings from the session afterward too.</p><p>Best,</p><p>Pranav</p><p>PS: If you know a PM who&#8217;s wrestling with AI opportunities right now, feel free to forward this to them &#8212; they might find it useful</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/testing-something-new-a-lightning?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/testing-something-new-a-lightning?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[How to Spot the 10% of AI Projects That Work]]></title><description><![CDATA[You&#8217;ve probably lived this d&#233;j&#224; vu: someone pitches a shiny new feature, the slides are filled with screenshots of chat interfaces or &#8220;copilots,&#8221; and there&#8217;s a passing reference to OpenAI or Anthropic.]]></description><link>https://www.builderlab.ai/p/how-to-spot-the-10-of-ai-projects</link><guid isPermaLink="false">https://www.builderlab.ai/p/how-to-spot-the-10-of-ai-projects</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sat, 30 Aug 2025 21:23:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9918724d-6c8d-48af-874a-19aa05fcc122_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>You&#8217;ve probably lived this d&#233;j&#224; vu: someone pitches a shiny new feature, the slides are filled with screenshots of chat interfaces or &#8220;copilots,&#8221; and there&#8217;s a passing reference to OpenAI or Anthropic. Everyone nods politely. A pilot is greenlit.</p><p>Three months later - nothing.</p><p>No metric moved. No one adopted it. The project limps along because nobody wants to admit it failed.</p><p>This isn&#8217;t the exception, it&#8217;s the rule. <strong>90% of AI projects don&#8217;t actually work.</strong> Not because the models fail to generate output. They work in that narrow sense. They fail in the more important sense: they don&#8217;t change a KPI, they don&#8217;t create adoption, and they don&#8217;t survive the first quarter once the demo excitement fades.</p><p>The 10% that succeed stick. They scale. They generate incremental revenue or reduce cost in ways you can see in the P&amp;L. I&#8217;ve seen those rare wins add hundreds of millions to the bottom line.</p><p>The obvious question: <strong>how do you spot the 10% and not waste time on the 90%?</strong></p><div><hr></div><h2><strong>Why most AI projects die</strong></h2><p>The biggest killer isn&#8217;t AI performance. It&#8217;s novelty bias.</p><p>A founder sees ChatGPT draft an essay and imagines an AI chatbot for their product. An exec hears a competitor talk about &#8220;copilots&#8221; on an earnings call and suddenly every team is tasked with building one.</p><p>But novelty is not impact, and if you&#8217;ve ever built anything, you know this. <strong>Impact comes from solving a problem that already hurts, in a place users already live, with data you actually own, tied to a metric the business already cares about.</strong></p><p>When you flip the lens from &#8220;what&#8217;s new?&#8221; to &#8220;what moves numbers?&#8221;, the 10% stand out clearly.</p><div><hr></div><h3>The PROVE-IT Filter: 7 Rules That Predict Which AI Projects Will Fail (and Which Will Stick)</h3><p>Over the last few years, I&#8217;ve reviewed hundreds of AI product ideas, from scrappy prototypes to global roadmaps. The same pattern always shows up: the successful ones share the same DNA.</p><p>That DNA lives in a filter I call <strong>PROVE-IT (</strong>pretty consistent with my first response to every AI pitch I hear). My filter turns hand-wavy excitement into a concrete go/no-go decision.</p><ul><li><p><strong>Pain</strong>: The problem has to be hair-on-fire. If it doesn&#8217;t connect directly to a revenue leak, an operational bottleneck, or a customer frustration that already has budget and urgency, adoption odds drop sharply.</p></li><li><p><strong>Reach</strong>: The solution must live inside the user&#8217;s existing workflow or journey. If it demands a new app, tab, or process change, expect abandonment. Impact only compounds when the AI is embedded where work already happens.</p></li><li><p><strong>Owned Data</strong>: Proprietary signals are the moat. Logs, transactions, outcomes, and labeled history that only you have access to, and that can&#8217;t be trivially replicated by a competitor or model provider, make the system defensible.</p></li><li><p><strong>Verifiability</strong>: Every strong AI project has a single KPI that can be measured weekly. That might be conversion, SLA adherence, average order value, containment, or refund rate. If success can&#8217;t be expressed in one crisp metric, it usually won&#8217;t last.</p></li><li><p><strong>Execution Loop</strong>: Good systems learn. They take user actions -- clicks, edits, corrections, returns -- and feed them back into the product so it improves week after week. Static prompts without feedback loops stall quickly.</p></li><li><p><strong>Integration Time</strong>: A project that takes six months before seeing real data rarely survives. Fast integration de-risks impact and builds confidence.</p></li><li><p><strong>Total Cost of Ownership</strong>: Unit economics need to hold as usage scales. Inference costs, retrieval layers, human review, and compliance overhead all matter. If cost grows faster than value, the project collapses under its own weight.</p></li></ul><p>Each project gets a score from <strong>0&#8211;2</strong> on every dimension. If something lands below a <strong>9</strong>, it&#8217;s weak. Anything <strong>11 or higher</strong> is worth testing in an <strong>8&#8211;12 week sprint</strong> with clear metrics.</p><h2>The PROVE&#8209;IT Field Guide</h2><p><strong>Scoring:</strong> Each line item gets <strong>0&#8211;2</strong>. Below <strong>9</strong> is weak. <strong>11+</strong> is worthy of the next 12 weeks.<br><strong>Scale:</strong> 0 = missing or hand&#8209;wavy, 1 = partial or unproven, 2 = strong, evidenced.</p><h3>P: Pain: &#8220;Hair&#8209;on&#8209;fire or nice&#8209;to&#8209;have?&#8221;</h3><ul><li><p><strong>What it means:</strong> Ties directly to a leak (cost), bottleneck (throughput), or urgent frustration (NPS).</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> Vague &#8220;delight&#8221; pitch, no money or minutes at stake.</p></li><li><p><strong>1:</strong> Qualitative pain without quantified stakes.</p></li><li><p><strong>2:</strong> Quantified leak or delay with owner and budget.</p></li></ul></li><li><p><strong>How to test this week:</strong> Pull last 90 days of tickets, refunds, or cycle times. Put euro or minute signs next to each.</p></li></ul><h3>R: Reach: &#8220;Inside the user&#8217;s existing path&#8221;</h3><ul><li><p><strong>What it means:</strong> Lives where work already happens, ideally auto&#8209;invoked.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> New app or tab with context switching.</p></li><li><p><strong>1:</strong> Deep&#8209;link from existing tool.</p></li><li><p><strong>2:</strong> Inline in the flow, triggered by the event that creates the need.</p></li></ul></li><li><p><strong>Proof in the wild:</strong> Gmail&#8217;s <strong>Smart Reply</strong> and <strong>Smart Compose</strong> sit <em>in</em> the inbox and auto&#8209;suggest replies; at launch Smart Reply generated ~12% of mobile replies in Inbox, a classic &#8220;in&#8209;flow&#8221; win. <a href="https://workspaceupdates.googleblog.com/2017/05/saving-you-time-with-smart-reply-in-gmail.html?utm_source=chatgpt.com">Workspace Updates Blog</a></p></li></ul><h3>O: Owned Data: &#8220;Do we have a moat?&#8221;</h3><ul><li><p><strong>What it means:</strong> Proprietary signals or history others can&#8217;t copy: clickstreams, outcomes, refunds, SLAs, process docs, device or call audio.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> Public web + generic LLM.</p></li><li><p><strong>1:</strong> Limited first&#8209;party logs without outcomes.</p></li><li><p><strong>2:</strong> Rich first&#8209;party data with labels/outcomes and compliance clarity.</p></li></ul></li><li><p><strong>How to test this week:</strong> List top 10 tables or stores you control. For each, mark: freshness, coverage, and whether outcomes are labeled.</p></li></ul><h3>V: Verifiability: &#8220;One weekly KPI&#8221;</h3><ul><li><p><strong>What it means:</strong> A single metric the team can move and defend.</p></li><li><p><strong>Examples:</strong> AHT, CSAT, repeat&#8209;contact rate, conversion, Average Order Value, refund rate, SLA%, developer cycle time.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> Vanity metrics or subjective &#8220;wow&#8221;.</p></li><li><p><strong>1:</strong> Proxy metric without business linkage.</p></li><li><p><strong>2:</strong> Primary KPI with baseline, target, owner, and dashboard.</p></li></ul></li><li><p><strong>Proof in the wild:</strong> For Microsoft 365 Copilot, controlled studies and TEI analyses track <strong>time saved by task</strong> (e.g., ~20% on email writing, ~30% on search, ~34% on content creation). Whatever you think of ROI studies, the KPI is explicit and measurable weekly. <a href="https://tei.forrester.com/go/microsoft/M365Copilot/docs/TheTEIOfMicrosoft365Copilot.pdf?utm_source=chatgpt.com">Forrester</a></p></li></ul><h3>E: Execution Loop: &#8220;Gets better with feedback&#8221;</h3><ul><li><p><strong>What it means:</strong> Productized learning from clicks, edits, thumbs, returns, QA outcomes.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> Static system.</p></li><li><p><strong>1:</strong> Logging exists, but no updates.</p></li><li><p><strong>2:</strong> Regular retrains or policy updates; in&#8209;product feedback visibly improves results.</p></li></ul></li><li><p><strong>How to test this week:</strong> Ship explicit feedback affordances (accept, edit, reject). Close the loop with a weekly model or heuristic update.</p></li></ul><h3>I: Integration Time: &#8220;Shadow mode in weeks&#8221;</h3><ul><li><p><strong>What it means:</strong> Thin slice ships into <strong>shadow</strong> or <strong>assist</strong> mode fast, with human&#8209;in&#8209;the&#8209;loop.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> Big&#8209;bang build.</p></li><li><p><strong>1:</strong> Pilot in a new tool or subset of users.</p></li><li><p><strong>2:</strong> Shadowing real traffic in the primary surface within weeks.</p></li></ul></li><li><p><strong>Proof in the wild:</strong> Klarna launched an AI support agent that immediately shouldered real workload in&#8209;flow; shipping quickly in the primary channel mattered. <a href="https://www.klarna.com/international/press/klarna-ai-assistant-handles-two-thirds-of-customer-service-chats-in-its-first-month/?utm_source=chatgpt.com">Klarna Italia</a></p></li></ul><h3>T: Total Cost of Ownership: &#8220;Unit economics survive scale&#8221;</h3><ul><li><p><strong>What it means:</strong> Costs scale sub&#8209;linearly vs. value. Model, retrieval, inference, human review, compliance, and re&#8209;training all counted.</p></li><li><p><strong>Score rubric:</strong></p><ul><li><p><strong>0:</strong> &#8220;We&#8217;ll figure out cost later.&#8221;</p></li><li><p><strong>1:</strong> Rough per&#8209;interaction math without adoption curve.</p></li><li><p><strong>2:</strong> Full P&amp;L view with sensitivity to token prices, guardrail costs, and human fallback.</p></li></ul></li></ul><div><hr></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>The PROVE&#8209;IT scorecard</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2LaG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2LaG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 424w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 848w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 1272w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2LaG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png" width="1456" height="775" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:775,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148385,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/172360157?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2LaG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 424w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 848w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 1272w, https://substackcdn.com/image/fetch/$s_!2LaG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35014d5b-4183-46a6-a39b-ef7de05834d0_1628x866.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>Real&#8209;world snapshots mapped to PROVE&#8209;IT</h2><h3>1) Microsoft 365 Copilot in knowledge work</h3><ul><li><p><strong>What was built and why</strong>: Knowledge workers spend countless hours on repetitive tasks like email, document search, and summarization, creating a clear pain point. The solution integrates directly into Word, Outlook, and Teams, eliminating context switching and placing AI where the work already happens. Its impact is verifiable through a simple KPI  [time saved per task] with studies showing consistent 20&#8211;30% efficiency gains. And because the system improves through tenant feedback and policy updates, accuracy and guardrails get stronger over time, creating a product that delivers compounding value the longer it runs.</p></li><li><p><strong>Evidence:</strong> Forrester TEI reports average time savings like <strong>~20%</strong> on email writing and <strong>~30%</strong> on search; UK public&#8209;sector trials documented meaningful daily minutes saved across thousands of employees. <a href="https://tei.forrester.com/go/microsoft/M365Copilot/docs/TheTEIOfMicrosoft365Copilot.pdf?utm_source=chatgpt.com">Forrester</a><a href="https://assets.publishing.service.gov.uk/media/683db42bd23a62e5d32680d0/M365_Copilot_Experiment_Findings_Report.pdf?utm_source=chatgpt.com">GOV.UK</a></p></li><li><p><strong>Example score:</strong> P2 R2 O2 V2 E1 I1 T1 = <strong>11</strong>.</p></li></ul><blockquote><p>Note on evidence quality: Microsoft&#8209; and partner&#8209;run studies skew positive; third&#8209;party reporting occasionally questions magnitude, so insist on your own baseline and measurement. </p></blockquote><h3>2) UPS ORION route optimization</h3><ul><li><p><strong>What was built and why:</strong> Delivery routes were a classic &#8220;hair-on-fire&#8221; problem [fuel waste, driver inefficiency, and delayed packages all eroded margins]. UPS developed ORION (On-Road Integrated Optimization and Navigation), an AI optimization engine embedded into driver workflows. With proprietary GPS, package, and driver data across millions of routes, UPS had an unmatched data moat. Impact was tracked through clear KPIs: miles saved, fuel reduction, CO&#8322; emissions avoided, and on-time delivery performance.</p></li><li><p><strong>Evidence:</strong> UPS reports <strong>~100M miles</strong> and <strong>~10M gallons</strong> saved annually after deployment -- hundreds of millions in operational savings. <a href="https://www.globenewswire.com/news-release/2020/01/29/1977072/0/en/UPS-To-Enhance-Orion-With-Continuous-Delivery-Route-Optimization.html?utm_source=chatgpt.com">GlobeNewswire</a><a href="https://www.supplychaindive.com/news/last-mile-sustainability-strategies-ups-amazon-disclosure/546005/?utm_source=chatgpt.com">supplychaindive.com</a></p></li><li><p><strong>Example score:</strong> P2 R2 O2 V2 E2 I0 T1 = <strong>11</strong>.</p></li></ul><h3>3) Wendy&#8217;s FreshAI for drive&#8209;thru ordering</h3><ul><li><p><strong>What was built and why:</strong> Peak-time labor shortages and slow order throughput created a visible bottleneck at Wendy&#8217;s drive-thrus. The company introduced FreshAI, a voice-ordering assistant integrated directly into the ordering lane, where customers were already speaking to place orders. The key metrics are service time, order accuracy, and upsell effectiveness. While the system benefited from fast integration into physical drive-thrus, feedback loops and consistency across accents, background noise, and menu complexity remain challenges.</p></li><li><p><strong>Evidence:</strong> Expansion plan to <strong>500&#8211;600</strong> locations with reported ~<strong>22 seconds</strong> faster service and near&#8209;<strong>99%</strong> order accuracy at pilots; industry peers show mixed outcomes, so instrumentation matters. <br><a href="https://www.businessinsider.com/wendys-expanding-ai-ordering-hundreds-more-drive-thrus-2025-2?utm_source=chatgpt.com">Business Insider</a><a href="https://www.restaurantdive.com/news/wendys-deploy-digital-menu-boards-drive-thru-ai-500-restaurants-2025/746977/?utm_source=chatgpt.com">restaurantdive.com</a></p></li><li><p><strong>Example score:</strong> P2 R2 O1 V1 E1 I1 T1 = <strong>9</strong>.</p></li></ul><blockquote><p>Counter&#8209;example, same pattern: DoorDash ended its voice&#8209;ordering pilot in 2025 -- a reminder to validate accuracy, edge cases, and unit economics in shadow before scaling. <a href="https://www.bloomberg.com/news/articles/2025-05-20/doordash-ends-ai-voice-ordering-product-for-restaurants?utm_source=chatgpt.com">Bloomberg</a><a href="https://restaurantbusinessonline.com/technology/doordash-scraps-its-ai-voice-ordering-business?utm_source=chatgpt.com">Restaurant Business Online</a></p></blockquote><h3>4) Stripe Radar fraud prevention</h3><ul><li><p><strong>What was built and why:</strong> Online payments are constantly exposed to fraud risk, creating both direct financial loss (chargebacks) and indirect cost (false positives blocking legitimate customers). Stripe built Radar directly into its checkout and payments flow, embedding AI where the transactions already happen. Its moat comes from proprietary cross-network fraud data, enriched by outcomes and labels at global scale. The KPI is sharp: reduction in fraud rate with controlled false positives. Radar continuously retrains on outcomes, feeding fraud decisions back into the model to improve accuracy at scale.</p></li><li><p><strong>Evidence:</strong> Stripe reports users see <strong>~42% SEPA</strong> and <strong>~20% ACH</strong> fraud reduction on average with Radar&#8217;s new models. <a href="https://stripe.com/newsroom/news/radar-for-ach-sepa?utm_source=chatgpt.com">Stripe</a></p></li><li><p><strong>Example score:</strong> P2 R2 O2 V2 E2 I2 T2 = <strong>14</strong>.</p></li></ul><h3>5) Gmail Smart Reply / Compose</h3><ul><li><p><strong>What was built and why:</strong> Email is a universal time sink, with knowledge workers spending hours drafting routine responses. Google built Smart Reply and later Smart Compose directly into Gmail, meeting users inside their existing workflow with zero friction. The moat came from proprietary training data [billions of anonymized email threads] with real response outcomes. Success was tracked through a simple KPI: adoption and usage of suggested replies. The system improved through user edits and accept/reject signals, forming a natural feedback loop that sharpened predictions over time.</p><p><strong>Evidence:</strong> At launch, Smart Reply generated ~12% of all replies in Google&#8217;s Inbox app, scaling to billions of characters saved each week once rolled into Gmail. Over time, Google has emphasized efficiency gains and adoption at massive scale across its user base.</p><p><strong>Example score:</strong> P1 R2 O1 V2 E2 I2 T2 = <strong>12</strong> (slightly lower on Pain and Owned Data, since time saved per email is useful but less acute than cost or revenue drivers, and training relied heavily on scale rather than uniquely owned business-critical data).<br><a href="https://workspaceupdates.googleblog.com/2017/05/saving-you-time-with-smart-reply-in-gmail.html?utm_source=chatgpt.com">Workspace Updates Blog</a><a href="https://ucalgary.ca/news/making-email-more-efficient-means-answering-more-emails-even-faster?utm_source=chatgpt.com">University of Calgary in Alberta</a></p></li></ul><div><hr></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/how-to-spot-the-10-of-ai-projects?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/how-to-spot-the-10-of-ai-projects?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>How to run PROVE&#8209;IT in your org</h2><h3>1) Week 0&#8211;1: Baseline and thin slice</h3><ul><li><p><strong>Define the one KPI</strong> and baseline it for 4 weeks of historical data.</p></li><li><p><strong>Pick the surface you already own</strong> (e.g., search box, ticket console, compose box).</p></li><li><p><strong>Ship a thin slice in shadow</strong>: log model suggestions, human decisions, cost per suggestion, and &#8220;edit distance.&#8221;</p></li><li><p><strong>Events to instrument:</strong> <code>suggest_shown</code>, <code>suggest_accepted</code>, <code>edit_distance</code>, <code>task_time</code>, <code>fallback_triggered</code>, <code>token_cost_eur</code>.</p></li></ul><h3>2) Week 2&#8211;4: Close the loop</h3><ul><li><p><strong>Feedback UX:</strong> thumbs, edit&#8209;before&#8209;send, &#8220;good vs safe to ship&#8221;.</p></li><li><p><strong>Weekly updates:</strong> retrain heuristics or models; publish &#8220;what improved&#8221; changelog.</p></li><li><p><strong>Guardrails:</strong> auto&#8209;fallback rules and blocked intents.</p></li></ul><h3>3) Week 5&#8211;8: Controlled exposure</h3><ul><li><p><strong>Holdout design:</strong> user&#8209; or session&#8209;level A/B.</p></li><li><p><strong>Human&#8209;in&#8209;the&#8209;loop:</strong> queue low&#8209;confidence cases to reviewers; learn from corrections.</p></li><li><p><strong>Cost controls:</strong> cache, truncation, and retrieval; measure <strong>effective cost per accepted outcome</strong>.</p></li></ul><h3>4) Week 9&#8211;12: Scale or stop</h3><ul><li><p><strong>Unit economics review:</strong></p><ul><li><p><strong>Value per action:</strong> minutes saved &#215; loaded hourly rate, or revenue delta &#215; margin.</p></li><li><p><strong>Cost per action:</strong> model + retrieval + orchestration + review overhead.</p></li><li><p><strong>Adoption:</strong> % of eligible actions where AI is accepted.</p></li><li><p><strong>Decision:</strong> Kill, iterate, or scale with hard gates.</p></li></ul></li></ul><div><hr></div><h2><strong>Common failure modes and how PROVE&#8209;IT catches them</strong></h2><ul><li><p><strong>Chatbot in a new tab</strong> fails <strong>Reach</strong>. Expect abandonment even if the model is &#8220;good.&#8221;</p></li><li><p><strong>Cool demo with public data</strong> fails <strong>Owned Data</strong>. Anyone can copy it.</p></li><li><p><strong>Laundry list of metrics</strong> fails <strong>Verifiability</strong>. Pick one KPI and live by it.</p></li><li><p><strong>Static prompts</strong> fail <strong>Execution Loop</strong>. Improvements stall after Week 2.</p></li><li><p><strong>Quarter&#8209;long integration</strong> fails <strong>Integration Time</strong>. Push for shadow mode in the primary surface fast.</p></li><li><p><strong>Ignoring review labor</strong> fails <strong>TCO</strong>. Human&#8209;in&#8209;the&#8209;loop costs can swamp token savings.</p></li></ul><div><hr></div><h2><strong>A few playbooks you can borrow</strong></h2><ul><li><p><strong>Support assistant</strong>: Shadow&#8209;mode suggestions + containment rate + edit distance; promote to agent&#8209;assist then to partial autonomy by intent. Klarna&#8217;s trajectory shows why containment and AHT are the north stars.<a href="https://www.klarna.com/international/press/klarna-ai-assistant-handles-two-thirds-of-customer-service-chats-in-its-first-month/?utm_source=chatgpt.com"> Klarna Italia</a><a href="https://www.customerexperiencedive.com/news/klarna-reinvests-human-talent-customer-service-AI-chatbot/747586/?utm_source=chatgpt.com">customerexperiencedive.com</a></p></li><li><p><strong>Knowledge worker copilot</strong>: Task&#8209;level time trials per artifact type; Microsoft&#8217;s task&#8209;specific minutes saved is the right granularity.<a href="https://tei.forrester.com/go/microsoft/M365Copilot/docs/TheTEIOfMicrosoft365Copilot.pdf?utm_source=chatgpt.com"> Forrester</a></p></li><li><p><strong>Ops optimization</strong>: Start with deterministic heuristics, then learn residuals; measure miles or minutes saved like ORION.<a href="https://www.globenewswire.com/news-release/2020/01/29/1977072/0/en/UPS-To-Enhance-Orion-With-Continuous-Delivery-Route-Optimization.html?utm_source=chatgpt.com"> GlobeNewswire</a></p></li><li><p><strong>Voice ordering</strong>: Gate by confidence and fallback fast; scale only when service time and accuracy match human benchmarks. The DoorDash reversal is your cautionary tale.<a href="https://www.bloomberg.com/news/articles/2025-05-20/doordash-ends-ai-voice-ordering-product-for-restaurants?utm_source=chatgpt.com"> Bloomberg</a><a href="https://restaurantbusinessonline.com/technology/doordash-scraps-its-ai-voice-ordering-business?utm_source=chatgpt.com">Restaurant Business Online</a></p></li><li><p><strong>Fraud</strong>: Optimize both containment and false positives; Stripe&#8217;s cross&#8209;network signals showcase the power of Owned Data at scale.<a href="https://stripe.com/newsroom/news/radar-for-ach-sepa?utm_source=chatgpt.com">Stripe</a></p></li></ul><div><hr></div><h2><strong>Startups vs scaleups: different lenses</strong></h2><p>The same filter works in both contexts, but what you weight changes.</p><ul><li><p><strong>Startups</strong> should bias toward <em>reach and speed.</em> If you can&#8217;t land in a user flow in &lt;1 month, you&#8217;ve lost. An e-commerce startup testing AI-generated bundle suggestions at checkout can run the experiment in two weeks and see uplift or not. That&#8217;s the advantage.</p></li><li><p><strong>Scaleups</strong> must bias toward <em>verifiability and TCO.</em> Distribution is your edge, but bad economics will crush you. A large retailer once piloted AI-generated imagery for product listings. The demo wowed. But inference + QA costs meant negative margin at scale. It was dead on arrival.</p></li></ul><div><hr></div><h2><strong>Strong bets vs weak bets: patterns across industries</strong></h2><p>Patterns repeat.</p><p><strong>Strong bets:</strong></p><ul><li><p>B2B: triage and routing, summarization into action, structured extraction.</p></li><li><p>E-commerce: search rerankers, AI-driven product bundles, fraud detection (flagging high-risk orders with proprietary behavioral data), returns prevention (AI-powered sizing and fit recommendations).</p></li></ul><p><strong>Weak bets:</strong></p><ul><li><p>&#8220;AI chat for our store.&#8221;</p></li><li><p>Copilots without a KPI.</p></li><li><p>Anything that depends on users changing their primary tool.</p></li></ul><p>Why do returns prevention and fraud detection score high? Because they tie to pain (margin leakage), live in the flow (checkout, post-purchase), use owned data (transactions, return history), and are verifiable (chargeback rate, return rate). They are boring. But they matter.</p><div><hr></div><h2><strong>The builder&#8217;s takeaway</strong></h2><p>At the end of the day, spotting the 10% is about discipline.</p><ul><li><p>Prove the pain with numbers.</p></li><li><p>Live where the user already is.</p></li><li><p>Write the ROI on a napkin.</p></li><li><p>Define the fallback.</p></li></ul><p>If you can&#8217;t, kill it. If you can, run the 12-week play.</p><p>The temptation in AI is to believe every idea is a moonshot. The reality is that the projects that work are plumbing, not poetry. They quietly move a KPI. They integrate seamlessly. They survive past the demo stage.</p><p><strong>Think big. Start small. Iterate to greatness.</strong> That&#8217;s how you avoid the 90% trap, and build the 10% that actually stick.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><a href="https://stripe.com/newsroom/news/radar-for-ach-sepa?utm_source=chatgpt.com"><br></a></p>]]></content:encoded></item><item><title><![CDATA[Why I’m Building a Course on AI Strategy for Product Leaders]]></title><description><![CDATA[A few months ago, Gagan Biyani (co-founder of Maven) reached out and nudged me to consider creating a course.]]></description><link>https://www.builderlab.ai/p/why-im-building-a-course-on-ai-strategy</link><guid isPermaLink="false">https://www.builderlab.ai/p/why-im-building-a-course-on-ai-strategy</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sat, 17 May 2025 07:39:36 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bfef213e-200c-4c4d-b04e-841c748ae9ee_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few months ago, Gagan Biyani (co-founder of Maven) reached out and nudged me to consider creating a course. I wasn&#8217;t looking to launch anything at the time, but the idea lingered. It kept resurfacing in conversations with other product leaders, in my mentoring sessions, and during internal strategy reviews.</p><p>And here&#8217;s what I&#8217;ve come to realize:</p><blockquote><p><strong>Most product leaders today are flying blind when it comes to AI.</strong></p></blockquote><p>Not because they lack intelligence or initiative. But because the AI landscape, especially with ML and GenAI, is moving too fast, and product orgs haven&#8217;t caught up. There&#8217;s a growing gap between what teams <em>could</em> do with AI, and what they <em>are</em> doing.</p><div><hr></div><h3>The Problem: We&#8217;re Drowning in AI Hype, But Starving for Strategy</h3><p>Over the past year, I&#8217;ve worked with ML teams and execs across Meta, Booking.com, Flipkart, Goldman Sachs, and several startups. Some consistent themes keep emerging:</p><ul><li><p>ML products often feel fragmented.</p></li><li><p>GenAI experiments get spun up without real alignment to user problems.</p></li><li><p>Tech debt builds fast. Strategy&#8230; not so much.</p></li></ul><p>In short, <strong>product leaders are overwhelmed, under-supported, and missing the strategic frameworks needed to build AI products that actually move the needle.</strong></p><p>And I get it. I&#8217;ve been there.</p><p>At Flipkart, I was one of very few ML PMs in the company, building during crunch-time. At Booking.com, I&#8217;ve helped lead one of the most scaled AI transformations in travel tech, spanning classical ML, ranking systems, and GenAI platforms. At Meta, I&#8217;ve built ML products at a scale so vast that the resources at my disposal seemed infinite. I&#8217;ve seen what works, <em>and</em> what falls apart.</p><div><hr></div><h3>The Vision: A Course for Impact-Driven Product Leadership in AI</h3><p>So here&#8217;s what I&#8217;m building:</p><p>A <strong>high-intensity course for product leaders</strong> who want to go beyond the buzzwords and actually drive business impact with AI.</p><p>We&#8217;ll cover things like:</p><ul><li><p><strong>Identifying high-leverage ML/GenAI opportunities</strong> in your org</p></li><li><p><strong>Designing product strategy for intelligent systems</strong>, not just interfaces</p></li><li><p><strong>Structuring AI bets across Horizon 1, 2, and 3</strong></p></li><li><p><strong>Balancing GenAI and classical ML</strong> for practical business value</p></li><li><p>And yes, the hard stuff too: fallback design, ML evaluation, experimentation, and compliance</p></li></ul><p>It&#8217;s built for people leading product at startups, scale-ups, and large tech companies, who want to push their thinking forward and build with confidence.</p><div><hr></div><h3>Help Me Shape It</h3><p>I&#8217;m still refining the curriculum, and <em>I&#8217;d love your input</em>.</p><p><strong>If you're a PM, product lead, or product-minded founder working on AI</strong>, take 2 minutes to share your thoughts in this short survey:<br>&#128073; <a href="https://maven.com/forms/341112">https://maven.com/forms/341112</a></p><p>If you leave your email I&#8217;ll also send early access invites + priority seats before we open public enrollment.</p><p>Let&#8217;s build the next generation of AI<em>-</em>native product leaders together.</p><div><hr></div>]]></content:encoded></item><item><title><![CDATA[Prompt Frameworks for Unlocking ChatGPT's Deep Research Capabilities]]></title><description><![CDATA[And Exactly How to Use Them]]></description><link>https://www.builderlab.ai/p/prompt-frameworks-for-unlocking-chatgpts</link><guid isPermaLink="false">https://www.builderlab.ai/p/prompt-frameworks-for-unlocking-chatgpts</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Sun, 13 Apr 2025 10:20:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a3be3c1e-1385-4a6e-baf4-fc20e2be6704_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I love deep research&#8212;done right, it transforms surface-level questions into strategic clarity. After some structured experimentation, I've distilled my personal approach into 10 reusable prompt frameworks, explicitly designed to leverage ChatGPT&#8217;s deep research and search capabilities.</p><p>The frameworks reflect my own prompting style:</p><ul><li><p><strong>Goal-Oriented and Context-Rich:<br></strong>I clearly set the strategic context and intended outcome upfront&#8212;prompts become precise tools rather than generic queries.</p></li><li><p><strong>Builder&#8217;s Intent:<br></strong>Every prompt is deliberately outcome-driven, optimized to directly inform products, refine strategies, or enhance decision-making workflows.</p></li><li><p><strong>Specificity Balanced with Flexibility:<br></strong>I provide clear constraints while leaving room for nuanced, creative exploration&#8212;essential for deep research.</p></li><li><p><strong>System-Aware Prompting:<br></strong>Frameworks explicitly leverage how ChatGPT works best, maximizing deep research accuracy and insight quality.</p></li><li><p><strong>Structured for Workflow Integration:<br></strong>These prompts aren't standalone queries; they're strategically sequenced tools, designed for deliberate use at different workflow stages to progressively deepen strategic insights.</p></li></ul><p>A note before we jump in &#8212; my prompt writing style is VERY iterative. I rarely write one detailed prompt and let it do the heavy lifting. I like to see the intermediate steps and build from there. I often interrupt the LLM, making adjustments and changing the requirements on the fly. While this is a bit less efficient, I&#8217;ve found that it brings clarity to my thought process while giving me the outputs I want from the LLM.</p><h3><strong>Recommended Integration into Your Workflow:</strong></h3><ul><li><p><strong>Early-Stage Exploration: </strong>Use frameworks 1, 2, and 7 (<strong>First-Principles</strong>, <strong>Socratic Inquiry</strong>, <strong>Rapid Evidence Review</strong>) with deep research and search to quickly build foundational knowledge.</p></li><li><p><strong>Mid-Stage Refinement: </strong>Deploy frameworks 4, 5, and 6 (<strong>Scenario Futurecasting</strong>, <strong>Six Hats</strong>, <strong>Analogical Synthesis</strong>) alongside your own documents and search for rich, well-rounded strategic exploration.</p></li><li><p><strong>Late-Stage Decision &amp; Communication: </strong>Apply frameworks 3, 8, 9, and 10 (<strong>Inversion</strong>, <strong>Pros-Cons-Remedies</strong>, <strong>Role-Play Simulation</strong>, <strong>Strategy Canvas</strong>) in conjunction with your own documents, deep research, and visual canvas tools for sharp clarity, risk mitigation, and executive-ready communication.</p></li></ul><p>With that, let&#8217;s dive straight into the frameworks, starting with one of my foundational favourites&#8212;the <strong>First-Principles Breakdown</strong>.</p><div><hr></div><h2><strong>1. First-Principles Breakdown</strong></h2><p>The First-Principles Breakdown approach systematically deconstructs complex problems, strategies, or technologies into their fundamental truths or building blocks. It deliberately challenges existing assumptions, conventions, and perceived constraints to isolate core elements. By doing so, it empowers clear thinking, reduces cognitive biases, and reveals innovative paths that traditional, analogy-based thinking might miss. This foundational clarity then allows you to logically reconstruct robust, defensible strategies or solutions from the ground up.</p><p><strong>At What Point: </strong>Use this one for early-stage exploration or when confronting stuck or biased thinking.</p><p><strong>Where: </strong>Clarifying product ideas, complex strategies, or new technologies.</p><p><strong>How: </strong>Break down concepts to their fundamental building blocks, removing assumptions and biases, then rebuild solutions logically.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Break down <strong>[complex problem, idea, technology, or strategy]</strong> into its fundamental principles. Use insights from deep research and the provided supporting documents (<strong>[e.g., technical specs, market research, user feedback]</strong>) to clearly identify hidden assumptions, perceived constraints, and foundational truths. Then, logically reconstruct this into a structured, unbiased analysis, explicitly highlighting unexplored opportunities."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Break down the business model of subscription-based SaaS products into their fundamental principles. Identify hidden assumptions about customer behavior and pricing strategies, then rebuild a logical model highlighting unexplored opportunities."</p></li><li><p>"Deconstruct the key technological barriers to autonomous driving using first principles. Clearly list the fundamental components and assumptions, then logically identify areas where significant innovation is still needed."</p></li><li><p>"Explain the core challenges of scalability in blockchain technologies from first principles. Highlight fundamental constraints, and propose logical pathways for overcoming these obstacles."</p></li></ul><p><strong>Use in combination with: </strong>Your own supporting documents (technical specs, market research) - add these to the prompt - and &#8216;deep research&#8217; to validate foundational truths.</p><div><hr></div><h2><strong>2. Socratic Inquiry Chain</strong></h2><p>The Socratic Inquiry Chain is an iterative approach inspired by classical Socratic dialogue. It leverages ChatGPT to progressively deepen understanding by prompting the model to continually question and refine its own explanations. Each subsequent question reveals new nuances, surfaces overlooked subtleties, and challenges superficial assumptions. This method helps quickly build thorough expertise in complex or unfamiliar subjects by ensuring no stone remains unturned in the analysis.</p><p><strong>At What Point: </strong>After initial baseline research, before diving into deeper expert-level understanding.</p><p><strong>Where: </strong>Quickly mastering unfamiliar or complex scientific/technical topics.</p><p><strong>How: </strong>Iteratively prompt ChatGPT to question its own initial explanations, progressively uncovering deeper insights.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Briefly explain <strong>[complex or unfamiliar topic]</strong>. Then, iteratively deepen this explanation by critically questioning your initial response through <strong>[number of iterative cycles, e.g., 3&#8211;5]</strong> follow-up questions and answers. In each iteration, explicitly surface hidden nuances, clarify common misconceptions, and incorporate authoritative insights verified through deep research."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Briefly describe CRISPR gene editing. Now, deepen your explanation by self-questioning and answering through five iterative cycles, explicitly addressing risks, ethical issues, and technical complexities."</p></li><li><p>"Summarize the main concepts behind Large Language Models. Follow up by iteratively challenging your summary with detailed questions that address potential misunderstandings, limitations, and future directions."</p></li></ul><p><strong>Use in combination with:<br></strong>Search and deep research to verify accuracy and add authoritative context.</p><div><hr></div><h2><strong>3. Inversion (Devil&#8217;s Advocate)</strong></h2><p>Inversion, or playing the Devil&#8217;s Advocate, deliberately engages ChatGPT in skeptical, contrarian thinking to rigorously test strategic initiatives and decisions. This approach compels the model to systematically identify weaknesses, overlooked assumptions, and potential points of failure by imagining worst-case scenarios. In doing so, it helps decision-makers proactively mitigate risks, strengthen strategies, and guard against overly optimistic or biased planning.</p><p><strong>At What Point: </strong>Late-stage decision-making, immediately prior to final commitment.</p><p><strong>Where: </strong>Evaluating strategic initiatives, product launches, and significant business decisions.</p><p><strong>How: </strong>Prompt the model to adopt a skeptical viewpoint and identify critical points of failure proactively.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Imagine that <strong>[your strategic decision, product launch, or initiative]</strong> results in a significant failure (<strong>[specify worst-case outcome]</strong>). Using insights from deep research and the provided supporting documents (<strong>[e.g., strategic plans, market analysis, user feedback]</strong>), clearly outline the exact reasons behind this failure. Explicitly identify overlooked risks, hidden assumptions, and faulty logic. Finally, recommend proactive strategies or mitigations to prevent these points of failure."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Imagine our new AI-based recommendation system completely fails within the first six months after launch. Clearly outline why this failure occurred, detailing specific blind spots, faulty assumptions, and overlooked risks."</p></li><li><p>"Consider our upcoming market entry strategy into Asia and assume it leads to significant financial loss. Analyze the exact reasons behind this outcome, emphasizing hidden assumptions or underestimated challenges."</p></li><li><p>"Assume the decision to switch our company entirely to remote work severely damages productivity and culture. Identify the critical assumptions we made incorrectly, and suggest ways we could mitigate these risks in advance."</p></li></ul><p><strong>Use in combination with: </strong>Your own supporting strategic planning documents and deep research for thorough risk assessment.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Finding this helpful? Subscribe to save and find more content like this.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>4. Scenario Futurecasting</strong></h2><p>Scenario Futurecasting involves generating multiple, diverse, and plausible visions of future developments related to a specific topic, technology, or market. It explicitly prompts the AI to explore optimistic, pessimistic, and moderate trajectories, enabling decision-makers to anticipate potential outcomes and their implications strategically. By visualizing and examining varied future possibilities, leaders can better prepare for uncertainties, allocate resources wisely, and create flexible strategic plans.</p><p><strong>At What Point: </strong>Mid-to-late strategic planning stages, especially when uncertainty is high.</p><p><strong>Where: </strong>Innovation foresight, long-term strategic planning, exploring market uncertainties.</p><p><strong>How: </strong>Prompt for multiple plausible future scenarios, explicitly detailing strategic implications.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Develop <strong>three distinct future scenarios</strong> for <strong>[specific technology, market, or trend]</strong> over <strong>[specific time horizon]</strong>:</em></p><ul><li><p><em><strong>Optimistic scenario:</strong> characterized by rapid positive developments.</em></p></li><li><p><em><strong>Pessimistic scenario:</strong> involving significant setbacks or adverse conditions.</em></p></li><li><p><em><strong>Moderate scenario:</strong> depicting incremental, steady progress.</em></p></li></ul><p><em>Clearly outline each scenario using authoritative insights from deep research and provided industry analyses (<strong>[e.g., market data, trend reports, competitive benchmarks]</strong>). For each scenario, explicitly detail the strategic implications and recommended actions our organization should consider."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Create three detailed scenarios for the evolution of generative AI adoption in the enterprise over the next five years: an optimistic scenario with rapid adoption and integration, a pessimistic scenario characterized by major setbacks and regulatory hurdles, and a moderate scenario of steady, incremental growth. Clearly outline strategic implications and recommended actions for each."</p></li><li><p>"Envision three possible futures for electric vehicle (EV) market penetration globally by 2030&#8212;high adoption, low adoption, and moderate adoption&#8212;and explain the strategic consequences for traditional automakers."</p></li><li><p>"Develop three distinct future scenarios for blockchain's role in supply chain management by 2028, covering optimistic widespread adoption, pessimistic regulatory pushbacks, and moderate incremental progress. Outline clear strategies our organization should consider for each scenario."</p></li></ul><p><strong>Use in combination with: </strong>Deep research and your own industry analyses to ensure scenario realism. Actually, ALWAYS use this one with deep research!</p><div><hr></div><h2><strong>5. Six Thinking Hats Analysis</strong></h2><p>The Six Thinking Hats Analysis method systematically guides ChatGPT through six distinct analytical lenses&#8212;facts (white hat), emotions and intuition (red hat), risks and potential negatives (black hat), optimistic benefits (yellow hat), creative alternatives (green hat), and structured overview or process (blue hat). This structured perspective-taking ensures a holistic, balanced exploration of decisions, providing clear, nuanced, and actionable insights for complex issues.</p><p><strong>At What Point: </strong>Before making critical decisions, ensuring balanced consideration of all perspectives.</p><p><strong>Where: </strong>Comprehensive decision-making, product reviews, or market-entry assessments.</p><p><strong>How: </strong>Explicitly guide ChatGPT through sequential analytical perspectives: factual, emotional, risk-oriented, optimistic, creative, and procedural.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Conduct a comprehensive <strong>Six Thinking Hats analysis</strong> for <strong>[decision, strategic initiative, or idea]</strong>, systematically addressing each perspective:</em></p><ul><li><p><em><strong>White Hat (Facts):</strong> Clearly outline relevant data and factual information.</em></p></li><li><p><em><strong>Red Hat (Emotions):</strong> Summarize emotional reactions and intuitive responses of key stakeholders.</em></p></li><li><p><em><strong>Black Hat (Risks):</strong> Identify specific potential negatives, risks, and critical obstacles.</em></p></li><li><p><em><strong>Yellow Hat (Benefits):</strong> Highlight explicit, optimistic benefits and opportunities.</em></p></li><li><p><em><strong>Green Hat (Creative Alternatives):</strong> Suggest creative or innovative alternatives to improve or enhance outcomes.</em></p></li><li><p><em><strong>Blue Hat (Structured Recommendation):</strong> Provide a clear, structured summary and actionable recommendation.</em></p></li></ul><p><em>Incorporate authoritative insights from deep research and provided supporting documents (<strong>[e.g., user surveys, market data, internal reports]</strong>) to substantiate each analytical perspective."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Using the Six Thinking Hats method, evaluate our company's proposal to shift entirely to subscription-based pricing. Include clear insights covering factual data (market benchmarks, pricing models), emotional reactions (customer perceptions), risks (potential churn), benefits (stable revenue streams), creative alternatives (tiered models), and conclude with a structured final recommendation."</p></li><li><p>"Conduct a Six Thinking Hats analysis of our planned expansion into emerging markets. Clearly structure your response by assessing the facts (market size, competitors), emotional reactions (internal team morale), risks (political instability, economic volatility), benefits (growth potential), creative strategies (joint ventures, localized branding), and finally provide a concise recommendation."</p></li><li><p>"Evaluate the introduction of AI-driven customer service bots through the Six Thinking Hats framework. Methodically cover factual analysis (expected efficiency gains), emotional insights (customer satisfaction and frustrations), risks (loss of personal touch, technical issues), benefits (24/7 availability), creative alternatives (hybrid solutions), and deliver a structured strategic recommendation."</p></li></ul><p><strong>Use in combination with: </strong>Your own supporting documents (user surveys, market data) and optionally a canvas to visualise outcomes clearly.</p><div><hr></div><h2><strong>6. Analogical Synthesis</strong></h2><p>Analogical Synthesis is a creative reasoning technique prompting ChatGPT to explicitly draw comparisons between your current strategic challenge and analogous situations from entirely unrelated fields or historical contexts. This method unlocks novel perspectives and innovative solutions by borrowing insights from outside conventional industry or domain boundaries, stimulating creativity when traditional problem-solving methods fall short.</p><p><strong>At What Point: </strong>When conventional thinking has reached a limit and fresh perspectives are needed.</p><p><strong>Where: </strong>Creative problem-solving, innovation ideation, tough strategic problems.</p><p><strong>How: </strong>Prompt for explicit cross-domain analogies to uncover novel insights and solutions.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Identify <strong>two clear analogies</strong> from entirely unrelated fields, industries, or historical contexts that parallel our current challenge of <strong>[describe specific strategic challenge or issue]</strong>. For each analogy, explicitly:</em></p><ul><li><p><em>Explain the parallel clearly and concretely.</em></p></li><li><p><em>Highlight key insights, lessons, or innovations from the analogous situation.</em></p></li><li><p><em>Recommend specific, actionable strategies or ideas inspired by these analogies.</em></p></li></ul><p><em>Incorporate authoritative insights verified through deep research and historical or cross-industry examples to ensure accuracy and applicability."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Identify two clear analogies from historical innovations or disruptions in completely different industries that parallel our current challenge of increasing customer retention in SaaS products. Clearly explain each analogy and suggest actionable strategies inspired by these parallels.</p></li><li><p>"Our organization faces significant communication and coordination issues due to remote work. Provide two analogies from entirely unrelated fields (e.g., orchestras, space exploration, or military logistics) and suggest practical solutions we could adopt based on these analogies."</p></li><li><p>"Draw clear parallels between managing misinformation on social media platforms and controlling the spread of infectious diseases in public health. Outline specific strategies from epidemiology that we could effectively apply to manage misinformation."</p></li></ul><p><strong>Use in combination with: </strong>Search for historical or cross-industry examples, enhanced by deep research for accuracy.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Finding this helpful? Subscribe to save and find more content like this.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>7. Rapid Evidence Review</strong></h2><p>The Rapid Evidence Review approach directs ChatGPT to quickly and systematically summarize complex topics by highlighting key findings, central debates, research gaps, and actionable implications. It mimics a concise yet thorough literature or market review, offering rapid comprehension and actionable insights, ideal for informing strategic decisions or briefing stakeholders efficiently.</p><p><strong>At What Point: </strong>Early research phase or when needing rapid, concise insights for stakeholder briefings.</p><p><strong>Where: </strong>Quick summaries of scientific research, technological trends, or industry knowledge.</p><p><strong>How: </strong>Prompt structured summaries explicitly addressing findings, debates, gaps, and implications.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Rapidly summarize the current state of knowledge on <strong>[specific scientific topic, technology, trend, or issue]</strong>. Explicitly structure your response to address:</em></p><ul><li><p><em><strong>Key Findings:</strong> Clearly highlight the major established insights or breakthroughs.</em></p></li><li><p><em><strong>Ongoing Debates:</strong> Identify central discussions or controversies currently shaping the field.</em></p></li><li><p><em><strong>Research Gaps:</strong> Explicitly note significant unresolved questions or missing areas of knowledge.</em></p></li><li><p><em><strong>Strategic Implications:</strong> Clearly outline practical implications or actionable opportunities for <strong>[relevant stakeholder or business context]</strong>.</em></p></li></ul><p><em>Incorporate authoritative insights verified through deep research and recent sources to ensure accuracy and relevance."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Provide a structured overview of recent developments in renewable energy storage technologies, explicitly addressing key innovations, technical and economic debates, knowledge gaps, and potential business opportunities."</p></li><li><p>"Rapidly summarize the latest research on generative AI ethics and regulation. Clearly identify key research findings, highlight ongoing industry and academic debates, note significant knowledge gaps, and outline strategic implications for technology firms."</p></li></ul><p><strong>Use in combination with: </strong>Deep research and search for verifying accuracy and currency of insights.</p><div><hr></div><h2><strong>8. Pros-Cons-Remedies List</strong></h2><p>The Pros-Cons-Remedies framework systematically guides ChatGPT to clearly outline advantages (pros), disadvantages (cons), and explicit practical solutions or mitigations (remedies) for identified disadvantages. This structured evaluation promotes balanced and rigorous decision-making by proactively addressing potential challenges alongside their solutions, ensuring well-informed and resilient strategic choices.</p><p><strong>At What Point: </strong>Mid-stage decision-making, after shortlisting options, before final commitment.</p><p><strong>Where: </strong>Evaluating strategic choices involving clear trade-offs, product features, or tech investments.</p><p><strong>How: </strong>Prompt balanced, actionable assessments detailing pros, cons, and mitigations explicitly.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Provide a structured <strong>Pros-Cons-Remedies</strong> analysis for <strong>[specific strategic decision, product feature, or technology investment]</strong>. Clearly outline:</em></p><ul><li><p><em><strong>Pros:</strong> Explicit advantages, opportunities, and potential positive outcomes.</em></p></li><li><p><em><strong>Cons:</strong> Specific disadvantages, risks, and potential challenges.</em></p></li><li><p><em><strong>Remedies:</strong> Practical, actionable solutions or mitigations for each identified disadvantage or risk.</em></p></li></ul><p><em>Incorporate authoritative insights and verify accuracy through deep research and the provided supporting documents (<strong>[e.g., internal reports, market data, technical assessments]</strong>)."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"List the pros and cons of integrating GPT-based chatbots into our customer support systems. Clearly provide realistic mitigation strategies for each identified disadvantage or risk."</p></li><li><p>"Evaluate adopting remote-first working permanently. Identify clear advantages, potential drawbacks, and explicitly outline effective remedies or strategies to manage each identified con."</p></li><li><p>"Analyze switching our technology infrastructure entirely to cloud-based solutions. Present a balanced list of benefits and risks, providing detailed practical solutions or mitigations for each of the risks identified."</p></li></ul><p><strong>Use in combination with: </strong>Your own supporting documentation and deep research to enhance the rigor of identified risks and remedies.</p><div><hr></div><h2><strong>9. Role-Play Simulation</strong></h2><p>Role-Play Simulation prompts ChatGPT to realistically embody and simulate interactions between multiple stakeholders, capturing their distinct perspectives, reactions, and concerns. This approach uncovers interpersonal dynamics, identifies friction points, and predicts emotional and rational reactions to new decisions or changes. It's particularly effective for preempting stakeholder objections, aligning internal teams, and refining communication strategies before implementation.</p><p><strong>At What Point: </strong>Late in planning stages, before rollout or communication, ensuring stakeholder readiness and minimizing pushback.</p><p><strong>Where: </strong>Anticipating stakeholder reactions, user experiences, organizational dynamics around new decisions.</p><p><strong>How: </strong>Simulate realistic dialogues or stakeholder interactions explicitly surfacing friction points.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Simulate a realistic dialogue involving <strong>[list specific stakeholders involved]</strong> discussing <strong>[specific decision, change, or issue]</strong>. Explicitly:</em></p><ul><li><p><em>Embody each stakeholder's distinct perspective, emotional reactions, practical concerns, and rational objections.</em></p></li><li><p><em>Clearly highlight friction points, misunderstandings, or areas of disagreement.</em></p></li><li><p><em>Surface key unresolved questions or objections requiring attention before rollout or communication.</em></p></li></ul><p><em>Ground the dialogue in authoritative insights from provided stakeholder research, customer feedback, and deep research to ensure authenticity and accuracy."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Simulate a detailed conversation involving a product manager, a senior engineer, and a skeptical long-term customer discussing our decision to sunset an old but popular feature. Clearly highlight their concerns, objections, and questions."</p></li><li><p>"Imagine a dialogue between HR, team leaders, and front-line employees discussing the upcoming shift to hybrid remote work. Explicitly surface each stakeholder&#8217;s emotional reactions, practical objections, and unresolved concerns."</p></li><li><p>"Create a realistic interaction involving a privacy-conscious customer, a data scientist, and a compliance officer discussing the introduction of advanced user-data-driven personalization in our product. Identify their differing viewpoints, ethical concerns, and objections."</p></li></ul><p><strong>Use in combination with: </strong>Your own stakeholder research or customer feedback, optionally visualized using a canvas.</p><div><hr></div><h2><strong>10. Strategy Synthesis Canvas</strong></h2><p>The Strategy Synthesis Canvas guides ChatGPT to distill extensive research and analysis into a concise, structured strategic memo or executive summary. This approach clearly outlines strategic objectives, relevant market context, key insights, strategic options (each with explicit pros and cons), and concludes with a strongly justified recommendation and immediate next steps. It ensures clarity, coherence, and actionable guidance for executive-level decision-making and broader stakeholder alignment.</p><p><strong>At What Point: </strong>After comprehensive research, immediately before executive-level communication or broader organizational dissemination.</p><p><strong>Where: </strong>Concise strategic summaries, executive memos, translating extensive analysis into actionable strategy.</p><p><strong>How: </strong>Prompt explicitly structured strategic summaries including goals, context, insights, strategic options, and clear recommendations.</p><p><strong>Prompt Template:</strong></p><blockquote><p><em>"Convert the extensive analysis conducted on <strong>[specific strategic topic, decision, or initiative]</strong> into a concise, structured <strong>executive strategy memo</strong>. Explicitly structure your response into the following sections:</em></p><ul><li><p><em><strong>Strategic Objectives:</strong> Clearly define our core goals and intended outcomes.</em></p></li><li><p><em><strong>Relevant Context:</strong> Summarize essential market, competitive, or industry context based on deep research and supporting documentation (<strong>[specify documents]</strong>).</em></p></li><li><p><em><strong>Key Insights:</strong> Highlight the three most critical insights derived from the analysis.</em></p></li><li><p><em><strong>Strategic Options:</strong> Clearly present viable strategic alternatives, explicitly listing pros and cons for each.</em></p></li><li><p><em><strong>Recommended Strategy:</strong> Provide a strongly justified recommendation.</em></p></li><li><p><em><strong>Immediate Next Steps:</strong> Clearly outline actionable next actions and responsibilities.</em></p></li></ul><p><em>Ensure accuracy, clarity, and actionable recommendations by grounding your synthesis in the provided extensive supporting documentation, deep research, and visual summaries if available."</em></p></blockquote><p><strong>Example Prompts:</strong></p><ul><li><p>"Summarize our extensive analysis on entering the Asian market into a concise executive strategy memo. Clearly structure it around our key strategic objective, market context, the three most critical insights discovered, strategic options we considered (each with clearly listed pros and cons), the recommended strategy with rationale, and immediate next steps for implementation."</p></li><li><p>"Convert the detailed research conducted on integrating generative AI into our existing product ecosystem into a structured strategy memo suitable for executive review. Include clear sections on objectives, industry and competitive context, the top insights derived, strategic alternatives evaluated with explicit pros and cons, a robust recommendation, and clearly outlined next actions."</p></li><li><p>"Create a clear and concise strategic synthesis summarizing our recent customer research initiative. Clearly articulate our goals, customer context, key findings, strategic paths forward (detailing pros and cons for each), our recommended course of action with justification, and immediate steps for the product and marketing teams."</p></li></ul><p><strong>Use in combination with: </strong>Your own extensive supporting documents and analysis, summarized visually or structured clearly via a canvas.</p><div><hr></div><p>Structured prompting can greatly enhance your strategic thought process. Combine these prompt frameworks strategically with deep research, search, canvas visualisation, and your own rich internal documentation for exceptional outcomes.</p><p>Which combination resonates most with your style or current project? Would love your perspective.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/prompt-frameworks-for-unlocking-chatgpts?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/prompt-frameworks-for-unlocking-chatgpts?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><br></p>]]></content:encoded></item><item><title><![CDATA[Demystifying the Model Context Protocol]]></title><description><![CDATA[LLMs are no longer just text predictors &#8212; they&#8217;re turning into reasoning engines.]]></description><link>https://www.builderlab.ai/p/demystifying-the-model-context-protocol</link><guid isPermaLink="false">https://www.builderlab.ai/p/demystifying-the-model-context-protocol</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Fri, 04 Apr 2025 10:54:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/fb90778a-8c8b-48ca-9906-e9e8cb843f2f_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>LLMs are no longer just text predictors &#8212; they&#8217;re turning into reasoning engines. But while model performance improves, <strong>our ability to manage their context has lagged behind</strong>.</p><p>As we build richer GenAI systems &#8212; personalized shopping assistants, agentic workflows, retrieval-augmented copilots &#8212; we&#8217;re increasingly duct-taping memory, tools, user state, and business logic into brittle prompt strings. It works, until it doesn't.</p><p><strong>Model Context Protocol (MCP)</strong> is a response to this architectural gap. It offers a way to <strong>define, structure, and operationalize context</strong> so that models reason more intelligently &#8212; and systems scale more reliably.</p><p>This isn&#8217;t just about prompt templates. It&#8217;s about a design philosophy that treats context as data, not prose.</p><p>In this article, I&#8217;ll break down:</p><ul><li><p>What MCP is and what it includes</p></li><li><p>Why it matters for production-grade LLM systems</p></li><li><p>A step-by-step example of MCP in a GenAI e-commerce assistant</p></li><li><p>Ownership and design implications across PM, eng, and infra teams</p></li><li><p>When and how to introduce MCP in your own stack</p></li></ul><p>Let&#8217;s get into it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">New to BuilderLab.ai? Subscribe for free to receive new posts on AI, curated for product builders and pioneers.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3>What is Model Context Protocol (MCP)?</h3><p><strong>Model Context Protocol (MCP)</strong> is a proposed <strong>standard way to define, share, and manage the context</strong> that LLMs use during inference.</p><p>It&#8217;s meant to solve a growing problem: As we build <strong>complex GenAI systems</strong> &#8212; think multi-agent workflows, RAG pipelines, or tools interacting with LLMs &#8212; we need a reliable and interpretable way to <strong>pass &#8220;context&#8221; into models</strong>. </p><p>If you&#8217;re new to all this terminology (or to get the most out of this write-up), be sure to read through some other deep dives on <a href="https://www.builderlab.ai/p/understanding-llm-agents">LLM Agents</a>, <a href="https://www.builderlab.ai/p/what-llm-technique-is-right-for-you">RAG systems, and general LLM techniques.</a></p><h4>Why is this important?</h4><p>LLMs don&#8217;t work in isolation anymore. Modern AI products involve:</p><ul><li><p>Tool use (e.g., APIs, search, plugins)</p></li><li><p>Memory (long-term, short-term)</p></li><li><p>Retrieval-augmented generation (RAG)</p></li><li><p>User state, preferences, history</p></li><li><p>Chain-of-thought reasoning</p></li><li><p>Multi-step agents</p></li></ul><p>Each component needs <strong>clear, structured context</strong> to work well. But today, most developers just&#8230; shove it all into a prompt. MCP aims to bring <strong>order and clarity</strong>.</p><h4>What counts as "context"?</h4><p>In MCP, <strong>context</strong> is any information that guides the model&#8217;s behaviour, such as:</p><ul><li><p><strong>System instructions</strong></p></li><li><p><strong>Examples / few-shot prompts</strong></p></li><li><p><strong>Documents retrieved via RAG</strong></p></li><li><p><strong>Tools available to the model</strong></p></li><li><p><strong>User history or preferences</strong></p></li><li><p><strong>Session memory</strong></p></li></ul><h4>What does MCP actually define?</h4><p>It defines a <strong>structured format</strong> (usually JSON or similar) to organize these components. Think of it as a schema or protocol &#8212; like how HTTP structures web communication, MCP structures LLM interaction.</p><p>Here&#8217;s a simplified example:</p><pre><code>{
  "system_instruction": "You are a helpful personal shopping assistant.",
  "user_goal": "Find waterproof sneakers under &#8364;150 in a minimalist style.",
  "retrieved_documents": [...],
  "tools": ["price_tracker", "style_matcher"],
  "chat_history": [...],
  "memory": {
    "name": "Pranav",
    "shoe_size": "EU 43",
    "style": ["Minimalist", "Neutral colors"]
  }
}</code></pre><p>Then, an <strong>MCP interpreter</strong> or <strong>router</strong> translates this context into a full prompt the LLM understands.</p><h4>Who&#8217;s behind MCP?</h4><p>It's being explored and developed by folks in the <strong>open-source AI infra community</strong> &#8212; particularly people behind projects like LangChain, LlamaIndex, and agent frameworks like AutoGen or CrewAI. It&#8217;s not yet a universal standard, but momentum is growing.</p><h4>Why it matters for product builders (like you):</h4><ul><li><p>You can <strong>decouple logic from prompts</strong> &#8212; better modularity and reuse</p></li><li><p>Build <strong>multi-agent or tool-using systems</strong> more cleanly</p></li><li><p>Easier to debug and track what context influenced output</p></li><li><p>Better support for <strong>personalization</strong>, memory, and long-term interactions</p></li></ul><p>Let&#8217;s walk through an <strong>end-to-end Model Context Protocol (MCP)</strong>-style flow in a <strong>GenAI-powered e-commerce assistant</strong>. Think of it as a personal shopping concierge that can:</p><ul><li><p>Understand the customer&#8217;s preferences</p></li><li><p>Retrieve relevant product info</p></li><li><p>Use tools like price comparison or size guides</p></li><li><p>Maintain long-term memory for better recommendations over time</p></li></ul><p>As with everything, examples best illustrate the concepts I&#8217;ve mentioned before. So let&#8217;s drill down with an e-commerce scenario.</p><div><hr></div><h2>Scenario: GenAI Shopping Assistant</h2><p><strong>User goal:</strong> &#8220;Find me stylish, waterproof sneakers under &#8364;150 for city use. I prefer neutral colors and minimalist design.&#8221;</p><p>Let&#8217;s break this into an <strong>MCP-style flow</strong> with structured context, tools, memory, and model behavior.</p><h4>Step 1: Define the Context (MCP JSON)</h4><pre><code>{
  "system_instruction": "You are a stylish personal shopping assistant. Always recommend relevant, high-quality, fashion-forward products.",
  "user_goal": "Find stylish, waterproof sneakers under &#8364;150 for city use in neutral colors and minimalist design.",
  "user_profile": {
    "name": "Pranav",
    "style": ["minimalist", "neutral", "luxury-but-practical"],
    "shoe_size": "EU 43",
    "past_purchases": ["Common Projects", "On Cloudnova", "Nike Air Max 1"],
    "preferred_price_range": "100-200 EUR"
  },
  "retrieved_documents": [
    {
      "source": "ProductCatalog",
      "query": "waterproof sneakers neutral minimalist under 150",
      "results": [
        {
          "name": "Vessi Cityscape",
          "price": "&#8364;135",
          "color": "Slate Grey",
          "features": ["Waterproof", "Lightweight", "Minimalist"],
          "image_url": "...",
          "product_link": "..."
        },
        {
          "name": "Allbirds Mizzle",
          "price": "&#8364;145",
          "color": "Natural White",
          "features": ["Water-resistant", "Wool blend", "Sustainable"],
          "image_url": "...",
          "product_link": "..."
        }
      ]
    }
  ],
  "tools": [
    {
      "name": "PriceTracker",
      "description": "Checks if the product is on sale",
      "args": ["product_link"]
    },
    {
      "name": "StyleComparer",
      "description": "Compares visual style similarity to past purchases",
      "args": ["image_url", "past_purchases"]
    }
  ],
  "chat_history": [
    {
      "role": "user",
      "content": "I'm looking for shoes I can wear in Amsterdam &#8212; rainproof but still stylish."
    },
    {
      "role": "assistant",
      "content": "Got it! I&#8217;ll prioritize waterproof features with a clean aesthetic. Any preferred brands?"
    }
  ],
  "memory": {
    "long_term": {
      "Pranav_shoe_style": "modern minimalist, neutral colors",
      "purchase_behavior": "Willing to pay premium for quality"
    }
  }
}
</code></pre><h4>Step 2: MCP Interpreter Builds Prompt</h4><p>The interpreter compiles all this into a prompt like:</p><blockquote><p><strong>System:</strong> You are a fashion-savvy personal shopping assistant helping Pranav find stylish waterproof sneakers.</p><p><strong>User Goal:</strong> Pranav is looking for waterproof, minimalist sneakers in neutral colors under &#8364;150.</p><p><strong>User Profile:</strong> Pranav wears EU 43, prefers neutral colors, and has previously bought Common Projects and On Cloudnova. He likes quality over hype.</p><p><strong>Retrieved Products:</strong></p><ul><li><p>Vessi Cityscape: Waterproof, &#8364;135, Slate Grey, minimalist</p></li><li><p>Allbirds Mizzle: Water-resistant, &#8364;145, sustainable, Natural White</p></li></ul><p><strong>Use Tools:</strong> PriceTracker, StyleComparer</p><p>Based on the above, suggest the best options. Justify your picks with reasoning and style compatibility.</p></blockquote><h4>Step 3: Tools Are Called</h4><p>Model suggests using tools:</p><ul><li><p>Call <code>StyleComparer</code> to compare product images to Pranav's past taste</p></li><li><p>Use <code>PriceTracker</code> to flag any products on sale</p></li></ul><p>Example outputs from tools:</p><ul><li><p>Vessi: 92% style match, price stable</p></li><li><p>Allbirds: 85% style match, currently 10% off</p></li></ul><h4>Step 4: Assistant Response (LLM Output)</h4><blockquote><p>&#8220;Based on your style and past purchases, I&#8217;d recommend the <strong>Vessi Cityscape</strong> (&#8364;135). It&#8217;s highly waterproof, clean in design, and very close to your previous aesthetic.</p><p>The <strong>Allbirds Mizzle</strong> is another solid choice &#8212; a bit more casual and sustainable, and currently <strong>10% off</strong>. Want me to check sizing availability or delivery times?&#8221;</p></blockquote><h4>Step 5: Memory Updated</h4><p>Long-term memory can be updated with:</p><ul><li><p>"Pranav liked Vessi design."</p></li><li><p>"Responded positively to sale info."</p></li><li><p>"Might be open to sustainable materials."</p></li></ul><h4>Why this matters (and why MCP helps)</h4><ul><li><p><strong>Structured, modular context</strong> &#8594; Better model reasoning</p></li><li><p><strong>Tool-use integration</strong> &#8594; Richer assistant behavior</p></li><li><p><strong>Memory-aware</strong> &#8594; More personalized long-term experiences</p></li><li><p><strong>Easy to debug + maintain</strong> &#8594; Each part of the context is traceable</p></li></ul><div><hr></div><h3>Actors and Actions in the MCP context</h3><p>So MCP can be quite useful to log, keep track of, and manage large LLM systems. But who&#8217;s building what? Who has input in what part of the system? Let&#8217;s break down the <strong>ownership of each part of the flow</strong>: who defines what in an <strong>MCP-driven GenAI system</strong>, especially in an e-commerce context like your shopping assistant.</p><p>We'll break this into the main <strong>phases</strong>, then for each component, map out <strong>who is responsible</strong>:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/demystifying-the-model-context-protocol?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/demystifying-the-model-context-protocol?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kYZu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kYZu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 424w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 848w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 1272w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kYZu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png" width="1456" height="1144" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1144,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:211222,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/160567703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kYZu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 424w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 848w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 1272w, https://substackcdn.com/image/fetch/$s_!kYZu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44bd2777-2816-47b6-8626-249f861c7177_1540x1210.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>MCP itself doesn&#8217;t define the values</strong> &#8212; it defines the <strong>structure</strong> and <strong>interfaces</strong> to handle these cleanly.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!da--!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!da--!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 424w, https://substackcdn.com/image/fetch/$s_!da--!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 848w, https://substackcdn.com/image/fetch/$s_!da--!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 1272w, https://substackcdn.com/image/fetch/$s_!da--!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!da--!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png" width="1456" height="725" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:725,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:130315,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/160567703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!da--!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 424w, https://substackcdn.com/image/fetch/$s_!da--!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 848w, https://substackcdn.com/image/fetch/$s_!da--!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 1272w, https://substackcdn.com/image/fetch/$s_!da--!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5913e5a6-0c8b-4651-86aa-06724036c6e3_1510x752.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vX8N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vX8N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 424w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 848w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 1272w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vX8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png" width="1456" height="698" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:698,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:129857,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/160567703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vX8N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 424w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 848w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 1272w, https://substackcdn.com/image/fetch/$s_!vX8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe2a169d-8ac5-4837-a3c4-b69a4eda349d_1514x726.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zwlM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zwlM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 424w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 848w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 1272w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zwlM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png" width="1456" height="678" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:678,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:115521,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/160567703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zwlM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 424w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 848w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 1272w, https://substackcdn.com/image/fetch/$s_!zwlM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77340a20-f7c6-494f-916b-ea942f085b6c_1520x708.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_aSj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_aSj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 424w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 848w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 1272w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_aSj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png" width="1456" height="696" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:696,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:121687,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.builderlab.ai/i/160567703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_aSj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 424w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 848w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 1272w, https://substackcdn.com/image/fetch/$s_!_aSj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ca02532-cca6-41d8-8b4e-58ccb1817497_1502x718.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h3><strong>Who is MCP useful for, why, and when</strong></h3><p><strong>MCP</strong> is especially useful for <strong>teams building complex LLM-powered systems</strong> &#8212; whether you're working on:</p><ul><li><p>Personalized shopping assistants</p></li><li><p>Multi-step agent workflows</p></li><li><p>Retrieval-Augmented Generation (RAG) systems</p></li><li><p>Memory-aware conversational bots</p></li><li><p>LLM-powered tools and internal copilots</p></li></ul><h4><strong>Why use MCP</strong></h4><p>MCP brings <strong>clarity, structure, and scalability</strong> to LLM interactions by separating out and formalizing all the "hidden" context behind model behavior &#8212; such as goals, memory, tools, history, and retrieved documents.</p><p>Instead of shoving everything into one monolithic prompt, MCP turns context into a <strong>modular, inspectable object</strong>. That means:</p><ul><li><p>Easier debugging and observability</p></li><li><p>Better reuse across different agents and tasks</p></li><li><p>Cleaner personalization and memory updates</p></li><li><p>More consistent and maintainable prompt engineering</p></li></ul><h4><strong>When to use MCP</strong></h4><p>You should reach for MCP when:</p><ul><li><p>Your LLM use case is <strong>growing in complexity</strong> (e.g. agents, RAG, tools, long-term memory)</p></li><li><p>You're working in a <strong>production or multi-user environment</strong></p></li><li><p>You want <strong>reliability and clarity</strong> in how context flows into your LLM</p></li><li><p>You're building a <strong>platform</strong> or shared infra where multiple apps or agents interact with models</p></li></ul><div><hr></div><h3>What&#8217;s Next?</h3><p>Model Context Protocol is still in its early days, but it&#8217;s quickly becoming essential infrastructure for building scalable, modular, and intelligent GenAI systems.</p><p>If you're working on:</p><ul><li><p>A personalized shopping assistant</p></li><li><p>A multi-agent research tool</p></li><li><p>An LLM-powered internal copilot</p></li><li><p>Or anything that combines RAG, tools, and memory...</p></li></ul><p>&#8230;consider adopting MCP principles now &#8212; even informally &#8212; to future-proof your architecture.</p><p>In upcoming posts, I&#8217;ll dive deeper into:</p><ul><li><p>MCP templates and schema design</p></li><li><p>Live examples across e-commerce, HR tech, and SaaS copilots</p></li></ul><p>Want a reusable MCP starter kit or LangChain-compatible boilerplate? Drop a comment or DM &#8212; I&#8217;m putting something together soon.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/demystifying-the-model-context-protocol?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/demystifying-the-model-context-protocol?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[What LLM technique is right for you?]]></title><description><![CDATA[Should you use RAG? Agents? Fine tuning? Or stick to system prompts?]]></description><link>https://www.builderlab.ai/p/what-llm-technique-is-right-for-you</link><guid isPermaLink="false">https://www.builderlab.ai/p/what-llm-technique-is-right-for-you</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Wed, 22 Jan 2025 21:44:30 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1a76634f-9693-430a-8015-1bde74ec40aa_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Wuth a growing suite of techniques available to product builders&#8212;from system and user prompts to fine-tuning, Retrieval-Augmented Generation (RAG), tools, and agents&#8212;it can be confusing to decide which approach to adopt for your use case &#8212; and the most performant option isn&#8217;t always the most expensive one! Each method comes with its own trade-offs, costs, and expertise requirements. Lets break down these techniques, understand when to use each, and whether to combine them for an effective outcome.</p><div><hr></div><h2>TL;DR: Here&#8217;s a handy decision matrix:</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EyQc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EyQc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 424w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 848w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 1272w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EyQc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png" width="1456" height="654" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:654,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:149070,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EyQc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 424w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 848w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 1272w, https://substackcdn.com/image/fetch/$s_!EyQc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14a1deb8-2b53-47e1-8bd6-de8e5a524e2f_1510x678.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>The Foundations: System and User Prompts</strong></h3><p>The best place to get started, and often the most useful, is the simple user prompt / system prompt setup. This is pretty much just using prompts to do your bidding. When you write a prompt, split it by a system behavior definition and a user input, and that&#8217;s it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>What They Are:</strong> System prompts define the behavior of the LLM (e.g., &#8220;You are a helpful assistant&#8221;), while user prompts encapsulate the input provided by the end-user. Together, they form the simplest way to interact with an LLM.</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Rapid Prototyping and quick first launches:</strong> Perfect for proving immediate value from LLMs, this approach is especially useful when the LLM&#8217;s world knowledge is sufficient to interact with users. Some additional context can be added into system prompts, but it&#8217;s usually limited.</p></li><li><p><strong>Limited Complexity:</strong> Best used when user needs are straightforward (e.g., answering FAQs or generating content).</p></li><li><p><strong>Minimal Budget:</strong> Ideal for use cases where cost is a major constraint, as prompts alone require no additional infrastructure or model modifications.</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> A chatbot answering common travel queries like &#8220;What&#8217;s the best time to visit Paris?&#8221; or &#8220;Can I carry a power bank on a flight?&#8221; - you would typically rely on the LLM&#8217;s world knowledge for this.</p></li><li><p><strong>Ecommerce:</strong> Generating product descriptions or providing instant answers to &#8220;What are the dimensions of this sofa?&#8221; - you would typically pass on the product context (the catalog page) along with the user prompt.</p></li><li><p><strong>Customer Service:</strong> Addressing FAQs such as &#8220;What is your return policy?&#8221; or &#8220;How can I track my order?&#8221; - you would typically pass topic specific customer support workflows along with the user prompt.</p></li></ul><p><strong>Business Outlook:</strong> This approach offers a quick and cost-effective way to enhance customer experience without extensive development or infrastructure investment. However, it lacks depth for handling complex or highly personalized interactions.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Low Cost:</strong> No need for additional computational resources.</p></li><li><p><strong>Simplicity:</strong> Easy to set up and iterate on.</p></li><li><p><strong>Flexibility:</strong> Quickly adaptable to new tasks.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>No Context Retention:</strong> Each prompt is treated as a standalone interaction. If you want to have conversation continuity, you need to pass the previous user and LLM messages along with the user prompt.</p></li><li><p><strong>Lack of Customization:</strong> Limited ability to tailor responses for domain-specific needs. Since all info is contained within the prompt, it&#8217;s difficult to answer from a large knowledge base, or build recommendations based on inventory etc.</p></li></ul><p><strong>Cost:</strong> Lowest among all methods; typically only API call costs.</p><p><strong>Expertise Needed:</strong> Basic understanding of prompt design.</p><div><hr></div><h3><strong>Retrieval-Augmented Generation (RAG)</strong></h3><p><strong>What It Is:</strong> RAG integrates external data sources into LLM responses by retrieving relevant documents or data and appending it to the prompt. For example, user manuals, support manuals, and even supply data can be used alongside the LLM. The basic premise is, given a user prompt, search through a dataset and retrieve the most relevant user context. Append it to the prompt, and pass this to the LLM to respond.</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Dynamic Data Needs:</strong> When the LLM must access real-time or frequently updated information (e.g., current news, weather updates).</p></li><li><p><strong>Domain-Specific Knowledge:</strong> For proprietary or highly specific datasets that the model wasn&#8217;t trained on, or if the corpus of information is exceptionally large.</p></li><li><p><strong>Regulatory Compliance:</strong> When responses need to be based on verified or auditable information.</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> Providing real-time flight status updates or suggesting activities based on current weather conditions.</p></li><li><p><strong>Ecommerce:</strong> Recommending products based on a user&#8217;s purchase history or inventory data.</p></li><li><p><strong>Customer Service:</strong> Answering questions like &#8220;What are the latest promotions?&#8221; or &#8220;Is my order eligible for free shipping?&#8221;</p></li></ul><p><strong>Business Outlook:</strong> RAG enables businesses to bridge the gap between static LLM knowledge and dynamic, real-world data. It&#8217;s particularly useful for enhancing personalization and ensuring responses are accurate and up-to-date, making it a strong choice for competitive differentiation.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Enhanced Relevance:</strong> Incorporates the latest or domain-specific data.</p></li><li><p><strong>Scalability:</strong> Handles large datasets effectively.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>Latency:</strong> Retrieval can introduce delays in response times by adding one more step before generating a response.</p></li><li><p><strong>Complexity:</strong> Requires a retrieval system (e.g., Pinecone, Weaviate) and integration with the LLM.</p></li><li><p><strong>Incomplete Data Issues:</strong> If the retrieval system fails to fetch relevant data, the LLM&#8217;s response might lack accuracy.</p></li></ul><p><strong>Cost:</strong> Medium; involves additional infrastructure for indexing and retrieval.</p><p><strong>Expertise Needed:</strong> Moderate; knowledge of retrieval systems, embeddings, and prompt tuning.</p><div><hr></div><h3><strong>Fine-Tuning</strong></h3><p><strong>What It Is:</strong> Fine-tuning involves training an LLM on a domain-specific dataset to specialize it for particular tasks or industries.</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Highly Specialized Tasks:</strong> When the LLM needs deep expertise in a narrow domain (e.g., legal, medical, or scientific contexts), or when the outputs need to be structured and predictable (eg JSON formatted outputs).</p></li><li><p><strong>Brand Consistency:</strong> To ensure outputs align with a specific tone or style.</p></li><li><p><strong>Improving Base Model Limitations:</strong> When the base model struggles with task performance despite prompt optimization.</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> Tailoring an assistant to provide expert advice on niche destinations or luxury travel planning.</p></li><li><p><strong>Ecommerce:</strong> Ensuring product descriptions align with brand guidelines and customer preferences.</p></li><li><p><strong>Customer Service:</strong> Creating a specialized bot for handling complex policy inquiries or multilingual support.</p></li></ul><p><strong>Business Outlook:</strong> Fine-tuning offers precision and long-term ROI for businesses with specialized needs, though the upfront investment can be significant. It&#8217;s ideal for companies that require high-quality, domain-specific outputs consistently.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Precision:</strong> Produces highly accurate and tailored responses.</p></li><li><p><strong>Efficiency:</strong> Reduces the need for overly verbose prompts.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>Upfront Cost:</strong> Fine-tuning requires significant computational resources.</p></li><li><p><strong>Maintenance:</strong> Updated versions of the base model require re-tuning.</p></li></ul><p><strong>Cost:</strong> High; includes computational costs and potential licensing fees for access to base models.</p><p><strong>Expertise Needed:</strong> High; requires expertise in machine learning, data preparation, and evaluation.</p><div><hr></div><h3><strong>Tools and Function Calling</strong></h3><p><strong>What They Are:</strong> Tools extend LLM capabilities by allowing them to invoke external functions, APIs, or workflows dynamically (Read more about tools and agents here). For example, an application that needs the LLM to do simple math operations may give it access to a calculator. The LLM, when it encounters a math problem, passes the relevant variables to the calculator, and then uses the output in its response.</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Task Automation:</strong> For applications that require executing specific actions, like booking tickets or checking inventory.</p></li><li><p><strong>Contextual Responses:</strong> When the LLM needs to gather information dynamically (e.g., fetching user details from a database).</p></li><li><p><strong>Interactive Workflows:</strong> Ideal for use cases requiring multiple steps (e.g., assembling reports, running calculations).</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> Booking flights or hotels based on user preferences and available options.</p></li><li><p><strong>Ecommerce:</strong> Checking product availability and placing orders directly from a chatbot interface.</p></li><li><p><strong>Customer Service:</strong> Initiating refunds, modifying orders, or providing account-specific details.</p></li></ul><p><strong>Business Outlook:</strong> The ability to automate actions increases operational efficiency and reduces friction in user interactions. While initial setup costs can be high, tools unlock significant value in automating repetitive or complex workflows.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Dynamic Capabilities:</strong> Extends the LLM&#8217;s usefulness beyond static text generation.</p></li><li><p><strong>Modularity:</strong> New functions can be added without retraining the model.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>Rigidity:</strong> Limited to predefined functions.</p></li><li><p><strong>Development Overhead:</strong> Requires backend integration and robust validation.</p></li></ul><p><strong>Cost:</strong> Medium to High; depends on the number of API calls and backend requirements.</p><p><strong>Expertise Needed:</strong> Moderate; familiarity with API integration and LLM prompt design.</p><div><hr></div><h3><strong>Agents</strong></h3><p><strong>What They Are:</strong> Agents are advanced LLMs that use dynamic reasoning to plan and execute multi-step tasks, leveraging multiple tools or APIs autonomously (read more about agents here).</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Complex Problem Solving:</strong> When workflows involve multiple interdependent steps.</p></li><li><p><strong>Adaptable Systems:</strong> For tasks that require reasoning and decision-making on-the-fly.</p></li><li><p><strong>End-to-End Solutions:</strong> When the LLM needs to autonomously achieve goals with minimal human input.</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> Creating personalized itineraries by combining weather forecasts, user preferences, and activity options.</p></li><li><p><strong>Ecommerce:</strong> Assisting users with end-to-end purchase decisions, from product discovery to delivery tracking.</p></li><li><p><strong>Customer Service:</strong> Handling escalations by dynamically pulling data, invoking tools, and proposing solutions autonomously.</p></li></ul><p><strong>Business Outlook:</strong> Agents offer transformative potential by automating complex tasks with minimal human oversight. However, output unpredictability and high operational costs mean they&#8217;re best suited for high-value or experimental use cases.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Autonomy:</strong> Reduces human oversight for repetitive or complex workflows.</p></li><li><p><strong>Context Retention:</strong> Maintains continuity across multi-step interactions.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>Cost:</strong> Agents may make multiple API calls, significantly increasing costs.</p></li><li><p><strong>Unpredictability:</strong> Requires rigorous error handling to avoid unexpected behaviors.</p></li></ul><p><strong>Cost:</strong> Very High; due to complexity, API usage, and monitoring requirements.</p><p><strong>Expertise Needed:</strong> High; includes prompt engineering, tool integration, and error handling.</p><div><hr></div><h3><strong>Combining Techniques</strong></h3><p><strong>What It Is:</strong> Using multiple techniques together&#8212;e.g., combining RAG with tools or fine-tuning with agents&#8212;to maximize capabilities.</p><p><strong>When to Use:</strong></p><ol><li><p><strong>Hybrid Needs:</strong> When no single technique fully addresses your requirements.</p></li><li><p><strong>Scale and Specialization:</strong> To handle diverse user needs (e.g., personalized recommendations + real-time data retrieval).</p></li><li><p><strong>Future-Proofing:</strong> For systems that need to adapt as LLM capabilities evolve.</p></li></ol><p><strong>Use Case Examples:</strong></p><ul><li><p><strong>Travel:</strong> Combining RAG for real-time flight data with agents for personalized travel planning.</p></li><li><p><strong>Ecommerce:</strong> Using fine-tuned models for tailored product descriptions alongside tools for inventory management.</p></li><li><p><strong>Customer Service:</strong> Integrating RAG for policy retrieval with agents for handling multi-step issue resolution.</p></li></ul><p><strong>Business Outlook:</strong> Hybrid approaches deliver comprehensive solutions but come with increased complexity and costs. They&#8217;re best for organizations seeking to build robust, scalable, and future-proof systems.</p><p><strong>Advantages:</strong></p><ul><li><p><strong>Comprehensive Solutions:</strong> Addresses multiple challenges simultaneously.</p></li><li><p><strong>Resilience:</strong> Reduces reliance on a single point of failure.</p></li></ul><p><strong>Limitations:</strong></p><ul><li><p><strong>Complexity:</strong> Increases development and maintenance efforts.</p></li><li><p><strong>Cost:</strong> Combines the costs of all involved techniques.</p></li></ul><p><strong>Cost:</strong> Highest; involves layered infrastructure and operational expenses.</p><p><strong>Expertise Needed:</strong> Very High; requires deep technical understanding across multiple domains.</p><div><hr></div><h3><strong>Decision Matrix</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VdNR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VdNR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 424w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 848w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 1272w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VdNR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png" width="1456" height="654" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:654,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:149070,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VdNR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 424w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 848w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 1272w, https://substackcdn.com/image/fetch/$s_!VdNR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5b7d621-17df-4c53-af08-feb2eebe1de8_1510x678.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Choosing the right technique&#8212;or combination of techniques&#8212;is critical to building impactful and cost-effective systems. Start with the simplest approach that meets your needs, and scale up as your requirements become more sophisticated. By understanding the trade-offs of each method, you can optimize for both business value and technical feasibility.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Understanding LLM Agents]]></title><description><![CDATA[A product builder's perspective]]></description><link>https://www.builderlab.ai/p/understanding-llm-agents</link><guid isPermaLink="false">https://www.builderlab.ai/p/understanding-llm-agents</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Mon, 06 Jan 2025 04:45:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c938bdbb-834e-4ffe-ae87-9a2c89b6b084_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I&#8217;m assuming basic familiarity with large language models (LLMs) as I write this one. If you&#8217;d like a refresher, check out &#8216;<a href="https://foundit.substack.com/p/a-pms-introduction-to-large-language">A product builder&#8217;s introduction to large language models</a>&#8217;.</em></p><p>We&#8217;ve come quite far with using LLMs in their raw capabilities to build some very cool products. Better content personalisation, search, question answering, etc have proven great customer value. What we haven&#8217;t seen much of in B2C production yet is LLM agents (and this post will start with the basics and cover what agents are, and how to implement them!) - and for good reason! </p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/understanding-llm-agents?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! This post is public so feel free to share it!</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/understanding-llm-agents?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/understanding-llm-agents?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p>The Agent landscape is about to change though, and I predict that in the next year or so we will all be very comfortable using LLM agents in our production systems. Before we jump in, let&#8217;s dive a bit deeper into how we&#8217;ve been using LLMs in production so far.</p><p>We&#8217;ve seen so far 3 distinct &#8216;waves&#8217; of LLM usage evolution, each increasing in complexity and usefulness.</p><h1>Wave 1: Basic text generation</h1><p>The first wave, starting early 2023, focused on basic text generation by leveraging the capabilities of LLMs to:</p><ol><li><p><strong>Understand text intent</strong> (what is the topic being discussed in this text, what is the sentiment, &#8230;)</p></li><li><p><strong>Summarise text</strong> (summarise this very long article into a TL;DR &#8211; pretty handy for the article you are reading now ;) )</p></li><li><p><strong>Generate new text</strong> (describe &#8216;Paris&#8217; to a young family wanting to go for the first time)</p></li><li><p><strong>Answer questions</strong> (are there long queues at the Eiffel tower?)</p></li><li><p>Through a combination of these, <strong>have chat dialogues</strong> with users (plan a trip to Paris!).</p></li></ol><p>From a B2C industry perspective, this led to some pretty useful stuff, including enhancements to search, personalised content generation, question answer capabilities, and level 1 frontline customer support for chatbots. These capabilities remain useful today, and some of the largest realised impact of LLMs throughout the industry builds on these capabilities.</p><h2>Why we need more</h2><p>One limitation these capabilities present is the LLM&#8217;s lack of real-time world knowledge - for example, LLMs by themselves cannot answer &#8220;what&#8217;s the weather in Paris today?&#8221;. Because LLMs are trained on data snapshots, real-timeness is not a possibility. </p><p>The other limitation is, the LLM has world knowledge - but not business specific knowledge that&#8217;s proprietary to you and your business - for example, what&#8217;s your refund policy in case of a cancellation?</p><p>Many different approaches have been tried to work around the problem - including augmenting LLMs with pre-retrived, real time data through putting it in prompts (retrieve the weather for Paris from a weather API and make it available in the prompt), or retrieval augmentation systems to provide more bounded context (give the LLM access to your specific customer service procedures via data retrieved from your customer service manuals and added to prompts).</p><p>These approaches work fine for a number of use cases, but as the size of data increases, they begin to hit limitations. Imagine if you want to respond to that weather query for any location across the globe, you need to retrieve the weather all over the globe at every point in time and pass it on to the prompt - pretty unmanageable. Plus, you can&#8217;t really &#8216;do&#8217; much with this approach in terms of taking action. For our customer service use case, we can retrieve what the correct procedure is (let&#8217;s say it&#8217;s &#8216;initiate a refund&#8217;), but we can&#8217;t actually act on it.</p><p>The second wave of using LLMs, relying on function and tool calling, is a response to these limitations.</p><h1>Wave 2: Function calling</h1><p>We&#8217;ve defined in the last wave a few &#8216;functions&#8217; or tools the LLM needs to be able to use dynamically in order to get over its limitations. For our weather use case, it needs to be able to dynamically access a weather API. For our customer service use case, the LLM needs to be able to dynamically access product flows to initiate cancellations and refunds. </p><p>With function calling, we give the LLM access to these capabilities. We rely on the LLM&#8217;s discretion to &#8216;call&#8217; these functions when necessary, by analysing what our users want, and what the right function to call is that might generate this response.</p><p>Ok, that sounds like a beautiful system - why haven&#8217;t we been doing this extensively for the last couple of years? Because of this: <strong>We rely on the LLM&#8217;s discretion to &#8216;call&#8217; these functions when necessary. </strong>At the dawn of the ChatGPT era, LLMs were simply not good enough at using their discretion and following structure / prompts effectively. My early attempts (as well as many other&#8217;s) at function calling, way back in 2023, did not yield good enough outputs. That landscape has since drastically changed with better performing LLMs.</p><p>So how does this work in practice? In wave 1, our LLM prompts had two components - a &#8216;system prompt&#8217; that defined the system behaviour for the LLM (&#8220;you are a helpful customer service agent&#8221;), and a &#8216;user prompt&#8217;, which was the user&#8217;s dialog / context (&#8220;can I cancel my order?&#8221;). Function calling introduces a third input - a list of functions available to the LLM. Here&#8217;s an example:</p><pre><code>response = openai.ChatCompletion.create(
    model="gpt-3.5-turbo",
    messages=[

        {"role": "system", "content": "You are a customer service assistant..."},
        {"role": "user", "content": "I want to cancel order #12345"}
    ],


    functions=[
        {
            "name": "check_order_status",
            "description": "Validates order exists and gets current status",
            "parameters": {
                "type": "object",
                "properties": {
                    "order_id": {"type": "string"}
                },
                "required": ["order_id"]
            }
        },
        # ... other function definitions ...
    ]
)</code></pre><p>"role": "system", "content": is our system prompt.</p><p>"role": "user", "content": is our user prompt</p><p>functions = is our list of functions available to the LLM, notably in our example, the &#8216;check order status&#8217; function. Conceptually, just as simple as normal prompt writing.</p><p>The beauty of this setup? The LLM knows exactly what tools it has (from the function definitions) and how to use them (from the prompt), but can still maintain a natural conversation flow with users. Your backend then implements these functions to actually perform the requested actions.</p><p>This approach gives you the best of both worlds: structured, predictable function calls for your backend, and natural, helpful interactions for your users.</p><p>The key to good function calling prompts is being specific about the tools available and crystal clear about how they should be used. It's like giving someone a toolbox and an instruction manual - they need to know both what tools they have and how to use them properly.</p><h2>Limitations of function calling</h2><p>The main limitation of function calling is that the LLM can only work with the functions you've defined. If a user asks "what's the best time to visit Paris" but you haven't defined a function for seasonal tourism data, the LLM is stuck with generic knowledge from its training data.</p><p>This is why function calling, while powerful, can feel rigid. Every possible action needs to be anticipated and defined upfront. Want your assistant to help with hotel bookings too? That's a whole new set of functions to define. Want it to check local events? More functions. The system doesn't figure out new capabilities on its own - it can only use the tools you've explicitly given it.</p><p>The beauty of LLMs so far has been their non-deterministic, natural, ask-me-anything kind of outputs. Function calling, while making LLMs more useful, takes away from this non-deterministic beauty and brings back the rigidity of deterministic coding. </p><p>LLM Agents are an effective response to this limitation.</p><div><hr></div><h1>Wave 3: The era of LLM Agents</h1><p>The idea behind agents is simple. In the last few examples in Wave 2, we&#8217;ve defined a list of functions and given our LLM access to these tools. The LLM could, in our previous example, determine which tool to use. </p><p>To make the limitations of function calling clearer, let&#8217;s say we give our LLM access to two tools - a weather API (&#8216;what&#8217;s the weather like in Paris on a given day), and a booking API (a list of hotel bookings made by our user). In our function calling setup, we can&#8217;t answer the question &#8220;should I carry an umbrella today?&#8221;, because our weather function needs a definite day, and our booking API has no weather information.</p><p>An agent, however, is smarter. It will first check if the user is already in Paris today, by calling the booking API. If it finds that the user is in Paris today, it will check the weather for Paris today. If the weather indicates rain, it will respond with &#8216;yes, it&#8217;s wise to carry an umbrella today!&#8217;.</p><p>So, Agent LLMs go beyond simple function calling by being able to:</p><ol><li><p>Plan steps to achieve a goal</p></li><li><p>Choose their own tools and approach</p></li><li><p>Learn from and adapt to results</p></li><li><p>Maintain context and memory</p></li><li><p>Make decisions about what to do next</p></li></ol><h2>Setting up an Agent</h2><p>The setup is quite uncomplicated, at least conceptually. We have the same 3 components - the system prompt, the user prompt, and the list of tools. </p><p>Agent prompts are different though - they focus on goals and reasoning rather than specific steps. Let&#8217;s look at this through an e-commerce example, here is what an agent prompt would likely look like:</p><pre><code>You are a shopping assistant helping customers find and purchase products.

Your goal is to help customers find what they want and complete purchases.

For each customer interaction:
1. Understand their needs and preferences
2. Think through the best approach to help them
3. Use available tools as needed
4. Adapt your approach based on results
5. Keep track of important details they share

Explain your thinking process and what you're doing to help them.</code></pre><p>This prompt gives the LLM a lot more agency to reason and figure out what combination of tools can help.</p><p>Agents use the same fundamental structure as function calls. The prompts make them markedly different, by enabling the agent to perform:</p><ol><li><p><strong>Dynamic Planning</strong></p><ul><li><p>Function calling: Follows preset workflows</p></li><li><p>Agents: Create their own plans based on goals</p></li></ul></li><li><p><strong>Tool Usage</strong></p><ul><li><p>Function calling: Uses tools in predefined ways</p></li><li><p>Agents: Creatively combines tools to solve problems</p></li></ul></li><li><p><strong>Memory and Context</strong></p><ul><li><p>Function calling: Handles one interaction at a time</p></li><li><p>Agents: Maintains context across conversations</p></li></ul></li><li><p><strong>Problem Solving</strong></p><ul><li><p>Function calling: Fixed decision trees</p></li><li><p>Agents: Adaptive problem solving</p></li></ul></li></ol><p>Here's how an agent handles a complex request:</p><pre><code>User: "I need gift ideas for my mom who loves gardening"

Agent Thought Process:
1. Gather more information about preferences
2. Search multiple categories
3. Consider seasonality
4. Check reviews and ratings
5. Verify availability

[Asks clarifying questions about budget and specific interests]
[Searches across multiple categories]
[Checks inventory for top options]
[Creates curated list with reasoning]
[Adapts suggestions based on user feedback]</code></pre><h2>A practical example</h2><p>Lets build that e-commerce agent a little bit fuller.</p><p>Here is how I would <strong>define the tools</strong>:</p><pre><code>{
  "tools": [
    {
      "name": "search_products",
      "description": "Search product catalog, returns matching items",
      "parameters": {...}
    },
    {
      "name": "check_inventory",
      "description": "Check current stock levels for a product",
      "parameters": {...}
    }
  ]
}</code></pre><p><strong>Here&#8217;s the prompt:</strong></p><pre><code>You are a shopping assistant that can help customers find products, check availability, and place orders.

You can:
- Search the product catalog
- Check inventory levels
- Place orders
- Calculate shipping times

You cannot:
- Modify existing orders
- Apply custom discounts
- Access customer payment details</code></pre><p><strong>Some error handling</strong></p><pre><code>{
  "error": "NOT_AUTHORIZED",
  "message": "Agent cannot access customer payment information",
  "allowed_actions": ["redirect_to_payment_page"]
}</code></pre><p>And that&#8217;s it! Here&#8217;s how the agent would behave in practice:</p><p><strong>The user prompt:</strong></p><p>User: "I need running shoes under $100 that can be delivered by next week"</p><p>The agent's thinking process might look like:</p><pre><code>Thought: I need to find suitable shoes and check delivery options. Let me break this down:
1. Search for running shoes with price filter
2. Check inventory for quick delivery
3. Verify shipping times
4. Present options to user

Action: Let me search first
[calls search_products({
    "query": "running shoes",
    "filters": {
        "price_max": 100,
        "category": "athletic/running"
    }
})]

Thought: Found several options. Need to check stock and delivery for each
[calls check_inventory({
    "product_id": "shoe123",
    "location": "nearest_warehouse"
})]

Thought: This option has stock and can be delivered in time. Let me show it to the user and ask about size preferences...</code></pre><p>As you can see in the example above, the power of agents comes from their ability to:</p><ol><li><p>Break down complex problems</p></li><li><p>Choose appropriate tools</p></li><li><p>Adapt their approach based on results</p></li><li><p>Maintain context throughout the interaction</p></li><li><p>Make decisions about what information they need</p></li></ol><p>This flexibility makes agents particularly good for tasks where the path to the solution isn't straightforward or needs to adapt based on what's discovered along the way.</p><h2>Why don&#8217;t we have agents everywhere yet?</h2><p>Despite being so radically helpful, we haven't seen agents take over yet, despite their apparent potential.</p><p>The main challenges fall into three buckets:</p><ol><li><p><strong>Technical Reliability</strong></p><ul><li><p>Agents are unpredictable - they might take unexpected actions or get stuck in loops. We&#8217;ve essentially got only the prompts guiding the agents, and we rely on the reasoning capabilities of the underlying models to &#8216;figure out&#8217; the problem.</p></li><li><p>Error handling is complex because agents can attempt things in ways you didn't anticipate - given the amount of freedom and agency, you have far less control on the outcome.</p></li><li><p>Cost management is tricky, since agents might make multiple API calls to figure things out - if they&#8217;re stuck in a loop or if the prompt is suboptimal, the cost will really spiral out of control.</p></li><li><p>Each tool integration needs robust validation, rate limiting, and error states - adding further complexity and need for trial-and-error, domain expertise, etc.</p></li></ul></li><li><p><strong>Product Complexity</strong></p><ul><li><p>Building good agent UX is hard - users get frustrated when agents take too long to think, and can abandon workflows or even churn away from the product. A few workarounds for the UX do exist, including providing the user greater visibility into &#8216;what&#8217; the LLM is doing.</p></li><li><p>Setting the right expectations is crucial - users either expect too much or too little with agents. A well made agent is often underutilised because the user is not primed to explore all of its capabilities. An agent that is good at one task can often appear to overpromise because the UX / copy is not well done.</p></li><li><p>Recovery flows are complex - what happens when an agent fails halfway through a task? Users can get frustrated when these flows are not well managed, as they may need to start all over again without guarantees of success.</p></li><li><p>You need extensive logging and monitoring to understand what your agents are actually doing, to make prompt adjustments and debugging even remotely possible.</p></li></ul></li><li><p><strong>Business Risk</strong></p><ul><li><p>Agents can make costly mistakes if not properly constrained - imagine if your agent determines that the best possible result for every customer&#8217;s cancelled order is a full refund.</p></li><li><p>The development overhead is significant compared to simple function calling - as I mentioned, you need some degree of experience handling LLMs before venturing into agents.</p></li><li><p>Many use cases work fine with simpler approaches - while we have discussed quite a few examples using agents here, there are more pragmatic ways to start - for improving user experience as well as for gaining efficiencies.</p></li><li><p>The ROI isn't clear yet for many applications, and the costs are high. Industry estimates currently forecast that it will take an agent upto $20 to accomplish a task at human quality, that a human can do for $6. The unit economics will get better, but are not there yet!</p></li></ul></li></ol><p>That said, I think we're at an interesting inflection point. As the tooling improves and we figure out the right patterns for building with agents, we'll likely see more adoption in specific use cases where the benefits outweigh the complexity.</p><p>For the next writeup, I&#8217;ll focus a bit on some examples of agents in the wild, as well as when and for what problems you should consider using agents - and also when not to use them. We will go into the details of the product building flows, the metrics and setups to use, etc, as I continue writing this series.</p><p>Meanwhile, if this was helpful, </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/subscribe?"><span>Subscribe now</span></a></p><p>If you would like to start with the basics of LLMs, check out this writeup:</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:124622697,&quot;url&quot;:&quot;https://foundit.substack.com/p/a-pms-introduction-to-large-language&quot;,&quot;publication_id&quot;:275725,&quot;publication_name&quot;:&quot;BuilderLab.ai&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;title&quot;:&quot;A Product Builder's introduction to Large Language Models&quot;,&quot;truncated_body_text&quot;:&quot;This is the first in a series of posts covering Generative AI for product builders.&quot;,&quot;date&quot;:&quot;2023-05-31T21:33:21.373Z&quot;,&quot;like_count&quot;:25,&quot;comment_count&quot;:1,&quot;bylines&quot;:[{&quot;id&quot;:26602548,&quot;name&quot;:&quot;Pranav Pathak&quot;,&quot;handle&quot;:&quot;foundit&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/05125a0b-c0c2-4717-8e60-3e36028add07_690x690.png&quot;,&quot;bio&quot;:&quot;Product Leader @Booking.com. I write about my learnings in Product Management and the startup scene in Europe here: https://foundit.substack.com/ @FoundItBlog&quot;,&quot;profile_set_up_at&quot;:&quot;2022-12-01T18:43:20.651Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:257571,&quot;user_id&quot;:26602548,&quot;publication_id&quot;:275725,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:275725,&quot;name&quot;:&quot;BuilderLab.ai&quot;,&quot;subdomain&quot;:&quot;foundit&quot;,&quot;custom_domain&quot;:&quot;builderlab.ai&quot;,&quot;custom_domain_optional&quot;:true,&quot;hero_text&quot;:&quot;The AI Builder&#8217;s Toolkit &#8211; the definitive guide to ideate, design, scale, and optimize groundbreaking products with AI at their core.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;author_id&quot;:26602548,&quot;theme_var_background_pop&quot;:&quot;#FF0000&quot;,&quot;created_at&quot;:&quot;2021-02-01T08:07:28.823Z&quot;,&quot;rss_website_url&quot;:null,&quot;email_from_name&quot;:&quot;Pranav Pathak from BuilderLab.ai&quot;,&quot;copyright&quot;:&quot;Pranav Pathak&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;twitter_screen_name&quot;:&quot;pranavVersed&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://foundit.substack.com/p/a-pms-introduction-to-large-language?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!Z74f!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png" loading="lazy"><span class="embedded-post-publication-name">BuilderLab.ai</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">A Product Builder's introduction to Large Language Models</div></div><div class="embedded-post-body">This is the first in a series of posts covering Generative AI for product builders&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">3 years ago &#183; 25 likes &#183; 1 comment &#183; Pranav Pathak</div></a></div><p>If you would like to read about writing good prompts, check out this one:</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:153460596,&quot;url&quot;:&quot;https://foundit.substack.com/p/the-elusive-art-of-prompt-writing&quot;,&quot;publication_id&quot;:275725,&quot;publication_name&quot;:&quot;BuilderLab.ai&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;title&quot;:&quot;The Elusive Art of Prompt Writing&quot;,&quot;truncated_body_text&quot;:&quot;The quality of your prompts determines the success of your interactions with AI systems. Whether you're a product manager leveraging AI for ideation or a developer using it to build tools, writing effective prompts is both an art and a science. This guide will walk you through the principles of crafting high-quality prompts, with real-world examples and&#8230;&quot;,&quot;date&quot;:&quot;2024-12-21T18:37:06.017Z&quot;,&quot;like_count&quot;:11,&quot;comment_count&quot;:1,&quot;bylines&quot;:[{&quot;id&quot;:26602548,&quot;name&quot;:&quot;Pranav Pathak&quot;,&quot;handle&quot;:&quot;foundit&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/05125a0b-c0c2-4717-8e60-3e36028add07_690x690.png&quot;,&quot;bio&quot;:&quot;Product Leader @Booking.com. I write about my learnings in Product Management and the startup scene in Europe here: https://foundit.substack.com/ @FoundItBlog&quot;,&quot;profile_set_up_at&quot;:&quot;2022-12-01T18:43:20.651Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:257571,&quot;user_id&quot;:26602548,&quot;publication_id&quot;:275725,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:275725,&quot;name&quot;:&quot;BuilderLab.ai&quot;,&quot;subdomain&quot;:&quot;foundit&quot;,&quot;custom_domain&quot;:&quot;builderlab.ai&quot;,&quot;custom_domain_optional&quot;:true,&quot;hero_text&quot;:&quot;The AI Builder&#8217;s Toolkit &#8211; the definitive guide to ideate, design, scale, and optimize groundbreaking products with AI at their core.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;author_id&quot;:26602548,&quot;theme_var_background_pop&quot;:&quot;#FF0000&quot;,&quot;created_at&quot;:&quot;2021-02-01T08:07:28.823Z&quot;,&quot;rss_website_url&quot;:null,&quot;email_from_name&quot;:&quot;Pranav Pathak from BuilderLab.ai&quot;,&quot;copyright&quot;:&quot;Pranav Pathak&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;twitter_screen_name&quot;:&quot;pranavVersed&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://foundit.substack.com/p/the-elusive-art-of-prompt-writing?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!Z74f!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png" loading="lazy"><span class="embedded-post-publication-name">BuilderLab.ai</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The Elusive Art of Prompt Writing</div></div><div class="embedded-post-body">The quality of your prompts determines the success of your interactions with AI systems. Whether you're a product manager leveraging AI for ideation or a developer using it to build tools, writing effective prompts is both an art and a science. This guide will walk you through the principles of crafting high-quality prompts, with real-world examples and&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">a year ago &#183; 11 likes &#183; 1 comment &#183; Pranav Pathak</div></a></div><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BuilderLab.ai! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Think Big, Start Small]]></title><description><![CDATA[&#128161;Ideation for AI Products]]></description><link>https://www.builderlab.ai/p/think-big-start-small</link><guid isPermaLink="false">https://www.builderlab.ai/p/think-big-start-small</guid><dc:creator><![CDATA[Pranav Pathak]]></dc:creator><pubDate>Mon, 30 Dec 2024 02:30:44 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2e194959-d863-4c8c-9a50-408b6d093602_944x944.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>LLMs are reshaping the way we work, but turning big ideas into actionable solutions can feel overwhelming. As builders, we can sometimes get stuck between ambition and execution. This post will guide you through identifying impactful areas to apply generative AI and provide a detailed exercise to spark actionable ideas for integration.</p><div><hr></div><h3><strong>&#129488; Why start small?</strong></h3><p>Simply put, the possibilities with Generative AI are overwhelming. If you&#8217;ve got an idea or are passionate about a problem, there likely exists a genAI tool that can help you build your solution. From no code to LLM driven coding interfaces to entire flows managing and orchestrating agentic LLMs, a tool exists to facilitate building your product. If you&#8217;ve never coded before, and this was your barrier to building, well, now you don&#8217;t have to, as long as you understand the fundamentals. If you&#8217;re decent at software development, used right, LLMs can bring down your coding time to as low as 20%. From ideation to production, several new products entirely built by LLMs have now hit our app stores. If you&#8217;re leading an existing product or service, everyone including your customers, your employers, and your employees, expects you to use AI &#8211; have you seen the newest line of AI powered vaccum cleaners (!)?</p><p>If building excites you, all this can be overwhelming. If you&#8217;re anything like me, this feels like a time of endless possibility, but also a time of rapid paradigm shifts. Things are changing so fast it&#8217;s hard to keep up. There are so many ideas, so many experiments you want to try, that you&#8217;re starting things just to abandon them halfway before moving on to the next big idea. If that&#8217;s been your mental state, I&#8217;d urge you to breathe, and focus on starting (and finishing) small.</p><p>Focusing on, and iterating on, a small problem to solve, is usually a great way to develop the toolkit required to build a larger solution. It&#8217;s tempting to aim for the moon when exploring GenAI. But revolutionary change often begins with evolutionary steps. By identifying a small, impactful area to implement generative AI, you reduce the risks of overreach, speed up iteration, and build confidence in your ability to execute.</p><p>For example, an AI-driven customer support chatbot might start with FAQ generation. Once the feature demonstrates value, it could expand to handling complex customer interactions, learning from live agent interventions, and even having access to tools over time, creating incremental value at each step.</p><p>The key is to let the technology serve the problem, not the other way around. Instead of building a flashy, overcomplicated solution, focus on solving a single, meaningful pain point. Iterate, gather feedback, and expand incrementally.</p><div><hr></div><h3><strong>&#127919; Identifying Impactful Areas for Generative AI</strong></h3><p>Before diving into brainstorming, it&#8217;s crucial to understand where generative AI can make the most impact. Start by asking:</p><ol><li><p><strong>What are the repetitive tasks in this product?</strong></p><ul><li><p><em>Example:</em> In a customer support platform, generating responses to common queries can automatically reduce the workload on agents and improve response time. LLMs could generate comprehensive templates for troubleshooting, or dynamically create email drafts tailored to customer concerns based on past interactions. Beyond this, they might analyze call logs and automatically create scripts for escalation scenarios, or even invoke tools to directly guide users to finish their tasks, instead of requiring a human agent.</p></li></ul></li><li><p><strong>Where does personalization matter most?</strong></p><ul><li><p><em>Example:</em> In an e-commerce app, using generative AI to create dynamic product descriptions based on individual user preferences or past purchases can elevate the customer experience. For instance, LLMs could rewrite descriptions to highlight eco-friendly attributes for sustainability-conscious buyers, or suggest complementary products for repeat customers.</p></li></ul></li><li><p><strong>What unmet needs exist for better content creation?</strong></p><ul><li><p><em>Example:</em> In a marketing tool, LLMs can enable drafting social media posts or ad copy tailored to specific audience segments. The could help test variations of campaign content in real-time, learning what resonates most with each audience segment. This could bt taken further by integrating performance data from past campaigns to adjust tone and style dynamically, offering iterative improvement over time.</p></li></ul></li><li><p><strong>Where could insights lead to better decisions?</strong></p><ul><li><p>In a financial analysis platform, LLMs (augmented with existing data systems and reports) can be used to summarize trends from large datasets and generate reports with actionable recommendations. In the product discovery phases, they can synthesise user research documents to drive insights and provide meaningful recommendations for what to build next.</p></li></ul></li></ol><p>By focusing on these areas, you&#8217;ll uncover opportunities where GenAI&#8217;s capabilities align with real-world needs.</p><div><hr></div><h3><strong>&#129504; A Comprehensive Brainstorming Exercise</strong></h3><p>Let&#8217;s take a more in-depth approach to ideation. This extended exercise will help you explore opportunities for generative AI integration through a small step, iterative approach.</p><h4><strong>Step 1: Pick a Product You&#8217;ve Thought About Deeply</strong></h4><p>Choose a product you&#8217;ve worked on, use frequently, have studied, or really want to build. It could be anything&#8212;a project management tool, an e-commerce platform, or even your favorite fitness app. If you&#8217;re unsure where to start, pick the product you interact with daily, as familiarity often leads to deeper insights.</p><h4><strong>Step 2: Map Out Core User Journeys and Supporting Workflows</strong></h4><p>Document the main workflows accomplished by the product. For example:</p><ul><li><p>For a project management tool: creating tasks, assigning resources, generating progress reports, integrating with time-tracking apps.</p></li><li><p>For an e-commerce platform: browsing items, comparing products, completing a purchase, post-purchase support, and feedback collection.</p></li><li><p>For a fitness app: tracking workouts, planning meals, analyzing progress over time, connecting with fitness communities, and sharing achievements.</p></li></ul><p>As much as possible, break these workflows into micro-tasks to reveal nuanced inefficiencies AI could address. Additionally, consider how workflows intersect&#8212;for instance, how the browse and save behavior on an ecommerce catalog can affect future purchase behaviors (spoiler &#8211; more saves = more purchases).</p><h4><strong>Step 3: Identify Pain Points or Inefficiencies</strong></h4><p>For each task, ask:</p><ul><li><p>Where do users spend unnecessary time?</p></li><li><p>Where do errors or bottlenecks occur?</p></li><li><p>What&#8217;s currently &#8220;good enough&#8221; but could be exceptional with GenAI?</p></li><li><p>How does scaling impact these workflows?</p></li></ul><p>For example, in a fitness app, LLMs could streamline planning by generating meal suggestions that align with dietary restrictions, nutritional goals, and user preferences&#8212;all while learning from logged meals to improve future suggestions. Scaling this across millions of users could offer hyper-personalized plans without additional manual effort. Additionally, the system could proactively recommend substitutions based on seasonal ingredient availability.</p><h4><strong>Step 4: Map GenAI Capabilities to Each Pain Point</strong></h4><p>Using the questions from earlier, brainstorm how generative AI could enhance or replace parts of these workflows. Here are some examples:</p><ul><li><p><strong>Content Creation:</strong> Automatically generate templates for recurring tasks. <em>Example:</em> GenAI could create interactive project timelines that adapt dynamically based on real-time input from multiple collaborators.</p></li><li><p><strong>Summarization:</strong> Summarize meeting notes or user reviews. <em>Example:</em> In a conferencing app, LLMs could identify patterns across multiple meetings and provide strategic insights, like recurring blockers or themes.</p></li><li><p><strong>Personalization:</strong> Suggest tailored next steps based on user behavior. <em>Example:</em> In a fitness app, AI could evolve into a virtual coach, generating not just workout plans but also motivational prompts based on detected patterns in user engagement.</p></li><li><p><strong>Prototyping:</strong> Quickly draft mockups or design ideas for review. <em>Example:</em> In a design tool, GenAI could produce end-to-end workflows based on wireframes, including user onboarding sequences.</p></li></ul><h4><strong>Step 5: Prioritize and Refine</strong></h4><p>Evaluate:</p><ul><li><p>Scalability of the GenAI solution.</p></li><li><p>Expected ROI based on user feedback and market trends.</p></li><li><p>Technical complexity and data requirements for implementation.</p></li></ul><p>Consider creating a feature backlog where each proposed AI enhancement is tagged by difficulty, data dependency, and potential user value. Use this to map out short-term experiments and long-term strategic goals.</p><div><hr></div><h3><strong>&#128640; From Ideation to Execution</strong></h3><p>Once you&#8217;ve brainstormed and prioritized ideas, it&#8217;s time to bring them to life. Here are some possible next steps:</p><ol><li><p><strong>Prototype and Test:</strong> Use existing tools or APIs to create a minimum viable version of your idea. Leverage APIs and low code tools to generate functional but fast prototypes that can be immediately validated with real users.</p></li><li><p><strong>Learn and Iterate:</strong> Analyze feedback to refine your solution. Think of where your prototype users got &#8216;stuck&#8217;, what challenges did they face in usage, and what they found easy enough. Define a &#8216;viable&#8217; product that solves for pain points and stumbling blocks.</p></li><li><p><strong>Scale Strategically:</strong> Once validated, scale your features thoughtfully. Consider localization, compliance with regulations, and infrastructure optimizations.</p></li><li><p><strong>Monitor and Maintain:</strong> Build ongoing monitoring pipelines for your AI features to ensure quality over time. Use dashboards to track user engagement, AI-generated errors, and performance metrics.</p></li></ol><p>The transformative potential here isn&#8217;t about creating something that wows people on paper. It&#8217;s about delivering tools and experiences that feel indispensable in practice. Start small, but keep a clear vision of how your solution could scale and evolve into a broader ecosystem of AI-driven innovation.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/p/think-big-start-small?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.builderlab.ai/p/think-big-start-small?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>&#127937; Ready to Start?</strong></h3><p>Generative AI offers a playground of possibilities for builders. By thinking big but starting small, you can tackle meaningful problems, unlock new value, and set the stage for innovation that scales. Remember that GenAI is more than a tool&#8212;it&#8217;s a mindset shift. The intersection of user insights, AI capabilities, and strategic execution can lead to breakthroughs that redefine industries.</p><p>So, how would you integrate GenAI into a product you know well? The answer might be the seed of your next big breakthrough.</p><p>Read these to dive deeper into building &#8212;</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:34158031,&quot;url&quot;:&quot;https://foundit.substack.com/p/my-framework-for-machine-learning-products-part-1-of-3-2354a0da63d2&quot;,&quot;publication_id&quot;:275725,&quot;publication_name&quot;:&quot;BuilderLab.ai&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;title&quot;:&quot;The Product Builder's Guide to Machine Learning&#8202;&#8212;&#8202;Part 1 of 3&quot;,&quot;truncated_body_text&quot;:&quot;If you haven&#8217;t already, consider subscribing to FoundIt for more product building content!&quot;,&quot;date&quot;:&quot;2018-03-05T22:30:45.684Z&quot;,&quot;like_count&quot;:6,&quot;comment_count&quot;:0,&quot;bylines&quot;:[{&quot;id&quot;:26602548,&quot;name&quot;:&quot;Pranav Pathak&quot;,&quot;handle&quot;:&quot;foundit&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/05125a0b-c0c2-4717-8e60-3e36028add07_690x690.png&quot;,&quot;bio&quot;:&quot;Product Leader @Booking.com. I write about my learnings in Product Management and the startup scene in Europe here: https://foundit.substack.com/ @FoundItBlog&quot;,&quot;profile_set_up_at&quot;:&quot;2022-12-01T18:43:20.651Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:257571,&quot;user_id&quot;:26602548,&quot;publication_id&quot;:275725,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:275725,&quot;name&quot;:&quot;FoundIt&quot;,&quot;subdomain&quot;:&quot;foundit&quot;,&quot;custom_domain&quot;:&quot;builderlab.ai&quot;,&quot;custom_domain_optional&quot;:true,&quot;hero_text&quot;:&quot;The AI Builder&#8217;s Toolkit &#8211; the definitive guide to ideate, design, scale, and optimize groundbreaking products with AI at their core.&quot;,&quot;logo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/3b4def76-5112-4941-9903-47c0744ccedd_414x414.png&quot;,&quot;author_id&quot;:26602548,&quot;theme_var_background_pop&quot;:&quot;#FF0000&quot;,&quot;created_at&quot;:&quot;2021-02-01T08:07:28.823Z&quot;,&quot;rss_website_url&quot;:null,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Pranav Pathak&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;twitter_screen_name&quot;:&quot;pranavVersed&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://foundit.substack.com/p/my-framework-for-machine-learning-products-part-1-of-3-2354a0da63d2?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!Z74f!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png" loading="lazy"><span class="embedded-post-publication-name">BuilderLab.ai</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The Product Builder's Guide to Machine Learning&#8202;&#8212;&#8202;Part 1 of 3</div></div><div class="embedded-post-body">If you haven&#8217;t already, consider subscribing to FoundIt for more product building content&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">8 years ago &#183; 6 likes &#183; Pranav Pathak</div></a></div><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:32111485,&quot;url&quot;:&quot;https://foundit.substack.com/p/building-metrics-that-matter&quot;,&quot;publication_id&quot;:275725,&quot;publication_name&quot;:&quot;BuilderLab.ai&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png&quot;,&quot;title&quot;:&quot;Building Metrics that Matter&quot;,&quot;truncated_body_text&quot;:&quot;Metrics are my thing. I love thinking about, talking about, and working with metrics. I believe metrics can make or break products. When done right, they can supercharge a product and create a great user experience, but when done wrong, they can create a recipe for disaster.&quot;,&quot;date&quot;:&quot;2021-02-02T08:17:37.586Z&quot;,&quot;like_count&quot;:16,&quot;comment_count&quot;:8,&quot;bylines&quot;:[{&quot;id&quot;:26602548,&quot;name&quot;:&quot;Pranav Pathak&quot;,&quot;handle&quot;:&quot;foundit&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/05125a0b-c0c2-4717-8e60-3e36028add07_690x690.png&quot;,&quot;bio&quot;:&quot;Product Leader @Booking.com. I write about my learnings in Product Management and the startup scene in Europe here: https://foundit.substack.com/ @FoundItBlog&quot;,&quot;profile_set_up_at&quot;:&quot;2022-12-01T18:43:20.651Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:257571,&quot;user_id&quot;:26602548,&quot;publication_id&quot;:275725,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:275725,&quot;name&quot;:&quot;FoundIt&quot;,&quot;subdomain&quot;:&quot;foundit&quot;,&quot;custom_domain&quot;:&quot;builderlab.ai&quot;,&quot;custom_domain_optional&quot;:true,&quot;hero_text&quot;:&quot;The AI Builder&#8217;s Toolkit &#8211; the definitive guide to ideate, design, scale, and optimize groundbreaking products with AI at their core.&quot;,&quot;logo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/3b4def76-5112-4941-9903-47c0744ccedd_414x414.png&quot;,&quot;author_id&quot;:26602548,&quot;theme_var_background_pop&quot;:&quot;#FF0000&quot;,&quot;created_at&quot;:&quot;2021-02-01T08:07:28.823Z&quot;,&quot;rss_website_url&quot;:null,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Pranav Pathak&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;twitter_screen_name&quot;:&quot;pranavVersed&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://foundit.substack.com/p/building-metrics-that-matter?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!Z74f!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79671461-6560-4243-9304-f48b9cd32557_1024x1024.png" loading="lazy"><span class="embedded-post-publication-name">BuilderLab.ai</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">Building Metrics that Matter</div></div><div class="embedded-post-body">Metrics are my thing. I love thinking about, talking about, and working with metrics. I believe metrics can make or break products. When done right, they can supercharge a product and create a great user experience, but when done wrong, they can create a recipe for disaster&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">5 years ago &#183; 16 likes &#183; 8 comments &#183; Pranav Pathak</div></a></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.builderlab.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Or subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item></channel></rss>