<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Forward Deployed]]></title><description><![CDATA[At the intersection of AI, software development, and the enterprise.]]></description><link>https://www.forwarddeployed.com</link><image><url>https://substackcdn.com/image/fetch/$s_!xznZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2fc53ef-d888-4b36-9a31-95961074ba5e_178x178.png</url><title>Forward Deployed</title><link>https://www.forwarddeployed.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 13 May 2026 11:48:38 GMT</lastBuildDate><atom:link href="https://www.forwarddeployed.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Alephic, LLC]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[fwddeployed@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[fwddeployed@substack.com]]></itunes:email><itunes:name><![CDATA[Noah Brier]]></itunes:name></itunes:owner><itunes:author><![CDATA[Noah Brier]]></itunes:author><googleplay:owner><![CDATA[fwddeployed@substack.com]]></googleplay:owner><googleplay:email><![CDATA[fwddeployed@substack.com]]></googleplay:email><googleplay:author><![CDATA[Noah Brier]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Forward Deployed, Episode 5: Aligning Agents]]></title><description><![CDATA[Taylor Pearson joins Noah to explore what organization theory, complexity science, military doctrine, and markets can teach us about building agentic systems.]]></description><link>https://www.forwarddeployed.com/p/forward-deployed-episode-5-aligning</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/forward-deployed-episode-5-aligning</guid><dc:creator><![CDATA[Alephic HQ]]></dc:creator><pubDate>Thu, 07 May 2026 13:37:11 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/196681783/b59c041e231c431f02bf24e5518ca24d.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome to episode five of Forward Deployed. Noah sits down with <a href="https://taylorpearson.me/">Taylor Pearson</a> to continue the conversation about how we align agents, and why the best models may come from outside traditional software engineering.</p><p>Taylor brings a background that cuts across history, internet businesses, <em><a href="https://www.amazon.com/End-Jobs-Meaning-9-5-ebook/dp/B010L8SYRG">The End of Jobs</a></em>, risk parity investing, complexity science, and recent deep work with <a href="https://www.alephic.com/writing/the-magic-of-claude-code">Claude Code</a>. The conversation moves from firms and transaction costs to Toyota, military doctrine, memory, skills, and the problem of getting agents to work toward the right goal.</p><h2>Key Topics Covered</h2><ul><li><p><strong>Aligning agents:</strong> Why the episode starts with the question of how to align agents and what engineering can borrow from organization design, markets, and systems theory</p></li><li><p><strong>Taylor&#8217;s path:</strong> From history, SEO, ecommerce, and <em>The End of Jobs</em> to finance, risk parity, complexity science, and AI work</p></li><li><p><strong>Claude Code as a turning point:</strong> Why agentic command-line systems felt more transformative for Taylor than chatbot workflows</p></li><li><p><strong>Historical analogies for AI:</strong> Electricity, factory design, Toyota, and the need for a new pattern language for agentic work</p></li><li><p><strong>Companies as agentic systems:</strong> Why firms may be a more useful model than deterministic software systems for coordinating agents</p></li><li><p><strong>Junior employees and agents:</strong> The familiar failure mode where the work is done well but aimed at the wrong problem</p></li><li><p><strong>Bottlenecks and specs:</strong> Why fixing pull requests is not the whole game, and why the bottleneck may move upstream into briefs, specs, and coordination</p></li><li><p><strong>Pace layers and skills:</strong> Best practices, project architecture, plans, code, and how different layers of a system move at different speeds</p></li><li><p><strong>Memory and context:</strong> How skills, files, and externalized memory help agents carry useful context between sessions and systems</p></li><li><p><strong>Jobs and firm boundaries:</strong> How AI changes the calculus around what belongs inside a company, what gets outsourced, and which roles collapse together</p></li><li><p><strong>Writing, investing, and complex systems:</strong> Why some domains resist fixed best practices because everyone adapts to the same patterns</p></li></ul><h2>Timestamps</h2><p><em>Note: timestamps are approximate</em></p><ul><li><p><strong>00:00</strong> - Introduction: Taylor Pearson, complexity theory, organizations, and aligning agents</p></li><li><p><strong>01:15</strong> - Taylor&#8217;s path from history to SEO, ecommerce, <em>The End of Jobs</em>, finance, and AI</p></li><li><p><strong>03:30</strong> - The GFC, <em>The Black Swan</em>, markets, systems thinking, and transaction costs</p></li><li><p><strong>07:00</strong> - Claude Code, Codex, and the agentic workflow shift</p></li><li><p><strong>13:45</strong> - Historical analogies for AI, electricity, factories, and pattern languages</p></li><li><p><strong>18:00</strong> - Why companies may be the better model for agentic systems than software systems</p></li><li><p><strong>20:45</strong> - Organization metaphors, the Toyota Production System, and agents as junior employees</p></li><li><p><strong>24:00</strong> - Bottlenecks, specs, briefs, and the coordination work before code</p></li><li><p><strong>27:05</strong> - Mission alignment, pace layers, skills, best practices, and architecture</p></li><li><p><strong>32:20</strong> - Cynefin, best practices, good practices, and emergent practices</p></li><li><p><strong>34:50</strong> - Boyd&#8217;s OODA loop, Schwerpunkt, and shared objectives</p></li><li><p><strong>39:05</strong> - Memory, role boundaries, and what changes inside and outside the firm</p></li><li><p><strong>43:25</strong> - Externalized memory, skills, and context across systems</p></li><li><p><strong>46:25</strong> - Generalization, AI writing, de-slopping, and verifiable rewards</p></li><li><p><strong>49:15</strong> - Writing, investing, complex systems, and the Maginot Line problem</p></li><li><p><strong>51:20</strong> - Closing thoughts</p></li></ul><h2>Links &amp; References</h2><h3>Core References</h3><ul><li><p><a href="https://taylorpearson.me/">Taylor Pearson</a> - Guest on this episode</p></li><li><p><em><a href="https://www.amazon.com/End-Jobs-Meaning-9-5/dp/1619613352">The End of Jobs</a></em> - Taylor&#8217;s book on work, technology, and entrepreneurship</p></li><li><p><em><a href="https://www.penguinrandomhouse.com/books/541852/the-black-swan-second-edition-by-nassim-nicholas-taleb/">The Black Swan</a></em><a href="https://www.penguinrandomhouse.com/books/541852/the-black-swan-second-edition-by-nassim-nicholas-taleb/"> by Nassim Nicholas Taleb</a> - The GFC-era entry point Taylor mentions</p></li><li><p><em><a href="https://www.amazon.com/Thinking-Systems-Donella-H-Meadows/dp/1603580557">Thinking in Systems</a></em><a href="https://www.amazon.com/Thinking-Systems-Donella-H-Meadows/dp/1603580557"> by Donella Meadows</a> - The systems thinking reference in the conversation</p></li><li><p><em><a href="https://www.amazon.com/Images-Organization-Gareth-Morgan/dp/1412939798">Images of Organization</a></em><a href="https://www.amazon.com/Images-Organization-Gareth-Morgan/dp/1412939798"> by Gareth Morgan</a> - The organization metaphor book Taylor recommends</p></li><li><p><em><a href="https://www.amazon.com/Goal-Process-Ongoing-Improvement/dp/0884271951">The Goal</a></em><a href="https://www.amazon.com/Goal-Process-Ongoing-Improvement/dp/0884271951"> by Eliyahu Goldratt</a> - The bottleneck and theory-of-constraints reference Noah returns to</p></li><li><p><em><a href="https://itrevolution.com/product/the-phoenix-project/">The Phoenix Project</a></em> - The IT follow-up to <em>The Goal</em></p></li></ul><h3>Concepts &amp; Frameworks</h3><ul><li><p><a href="https://en.wikipedia.org/wiki/Ronald_Coase">Ronald Coase</a> and <a href="https://en.wikipedia.org/wiki/Transaction_cost">transaction cost economics</a> - Why work happens inside or outside firms</p></li><li><p><a href="https://en.wikipedia.org/wiki/Toyota_Production_System">Toyota Production System</a> - The people-and-machines operating system discussed in the episode</p></li><li><p><em><a href="https://en.wikipedia.org/wiki/A_Pattern_Language">A Pattern Language</a></em> - Christopher Alexander&#8217;s pattern-language idea applied to agentic work</p></li><li><p><a href="https://jods.mitpress.mit.edu/pub/issue3-brand/release/2">Stuart Brand&#8217;s pace layers</a> - The model Noah uses for best practices, architecture, plans, and code</p></li><li><p><a href="https://thecynefin.co/about-us/about-cynefin-framework/">Cynefin framework</a> - Simple, complicated, complex, and chaotic work domains</p></li><li><p><a href="https://en.wikipedia.org/wiki/OODA_loop">John Boyd&#8217;s OODA loop</a> - Decision-making and shared objective reference</p></li><li><p><a href="https://en.wikipedia.org/wiki/Schwerpunkt">Schwerpunkt</a> - The point-of-effort concept Taylor connects to agent alignment</p></li><li><p><em><a href="https://en.wikipedia.org/wiki/Systemantics">Systemantics</a></em><a href="https://en.wikipedia.org/wiki/Systemantics"> by John Gall</a> - The Maginot Line and previous-war problem discussed near the end</p></li></ul><h3>Tools &amp; Platforms</h3><ul><li><p><a href="https://github.com/heyitsnoah/claudesidian">Claudesidian</a> - Noah&#8217;s Obsidian/Claude framework</p></li><li><p><a href="https://openai.com/codex/">OpenAI Codex</a> - The competing agentic coding interface discussed in the episode</p></li><li><p><a href="https://obsidian.md/">Obsidian</a> - Note-taking and memory system context for agent workflows</p></li><li><p><a href="https://www.alephic.com/">Alephic</a> - Noah&#8217;s AI consulting company</p></li></ul><h3>Previous Episodes</h3><ul><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-1-the-bitter">Episode 1: The Bitter Lesson</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-2-claude">Episode 2: Claude Code Skills and the Progressive Disclosure Problem</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-3-context">Episode 3: Context Engineering</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-4-the-special">Episode 4: The Special Forces Model</a></p></li></ul><h2>About the Hosts</h2><p><strong>Noah Brier</strong> is co-founder of <a href="https://www.alephic.com/">Alephic</a>, an AI consulting company helping brands and enterprises build custom AI systems.</p><p><strong>Taylor Pearson</strong> is an author and investor whose work spans entrepreneurship, markets, complexity theory, and AI. He is the author of <em><a href="https://www.amazon.com/End-Jobs-Meaning-9-5-ebook/dp/B010L8SYRG">The End of Jobs</a></em>.</p><h2>Connect with the Hosts</h2><ul><li><p>Noah Brier: <a href="https://www.linkedin.com/in/noahbrier/">LinkedIn</a> | <a href="https://twitter.com/heyitsnoah">X/Twitter</a></p></li><li><p>Taylor Pearson: <a href="https://taylorpearson.me/">Website</a> | <a href="http://twitter.com/TaylorPearsonMe">X/Twitter</a></p></li><li><p>Alephic: <a href="https://www.alephic.com/">alephic.com</a></p></li></ul><div><hr></div><p>Subscribe for weekly episodes exploring how AI is actually being deployed in the real world.</p><p>Newsletter: Sign up for updates at <a href="https://forwarddeployed.com/">forwarddeployed.com</a></p>]]></content:encoded></item><item><title><![CDATA[A Timeline of AI Progress]]></title><description><![CDATA[Thinking about what matters and what doesn't in AI.]]></description><link>https://www.forwarddeployed.com/p/a-timeline-of-ai-progress</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/a-timeline-of-ai-progress</guid><dc:creator><![CDATA[Noah Brier]]></dc:creator><pubDate>Tue, 05 May 2026 16:46:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xSVC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I originally put this slide together for the September 2025 edition of <a href="http://brxnd.ai">BRXND</a>. It is a timeline of what I would consider the major milestones in AI over the last five years. The orange ones are the things I think are worth paying attention to. It was designed to scale as best I could with my limited Figma skills.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xSVC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xSVC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 424w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 848w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 1272w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xSVC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xSVC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 424w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 848w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 1272w, https://substackcdn.com/image/fetch/$s_!xSVC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9090cd7-f407-4fb8-8340-5536f0d239ca_2048x1152.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There are two things that should stand out:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.forwarddeployed.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Forward Deployed! Subscribe for new posts and episodes at the intersection of AI, engineering, and the enterprise.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><ol><li><p>There are fewer moments than you would assume over the last five years. That&#8217;s because I purposely left out every small model update and focused on the bigger events.</p></li><li><p>There are only five highlighted moments. Despite arguments to the contrary and the never-ending parade of model releases, I genuinely believe AI has undergone only a few major changes over the last five years.</p></li></ol><p>My five moments with their defenses:</p><ol><li><p><strong>November, 2021&#8212;GPT-3 API goes GA:</strong> This one is easy for me. We could draw the line in the sand almost anywhere after the Attention is All You Need paper, but to me, putting these things in the general public&#8217;s hands (albeit in API form) is a good demarcation point. There are still the &#8220;I&#8217;ve been doing AI for years&#8221; people complaining that ML is AI, but what&#8217;s possible before and after the GPT-3 API is very different.</p></li><li><p><strong>June, 2022&#8212;GitHub Copilot goes GA:</strong> it&#8217;s easy to take for granted now, particularly with the conversations happening around uptime and GitHub, and the eventual direction Microsoft took the Copilot brand, but that initial autocomplete product in VS Code felt like magic and was the first real productization of AI.</p></li><li><p><strong>November, 2022&#8212;ChatGPT Launch:</strong> This is the easiest one, and everyone likes to talk about it. Lots of us were building chatbots on our own, and this felt like such an obvious next step. Funny enough, this also immediately shifted Copilot use because while autocomplete was cool, chat turned out to be the killer app for coding (a fact Copilot seemed to miss and Cursor eventually ate their lunch on).</p></li><li><p><strong>September, 2024&#8211;o1 Reasoning:</strong> The Google chain of thought paper came out in early 2022 and showed that adding reasoning to prompts significantly improved the models&#8217; ability to reason, but it took two and a half years for OpenAI to build it into a model. o1 was big and slow, but set the groundwork for everything that followed. Though we didn&#8217;t realize it at the time, the real magic of these reasoning models wasn&#8217;t just that they were much better thought partners. Function calling was introduced to models in June 2023, but before reasoning models, it was incredibly inconsistent&#8212;better for demos than for real work. Reasoning changed that, allowing the model to plan its function calls before executing them.</p></li><li><p><strong>May, 2025&#8211;Claude Code:</strong> finally, we have Claude Code, which celebrates its first birthday this month. If o1 made function calling work, Claude Code wrapped it in an agent loop: inspect the repo, decide what to do next, call a tool, read the result, and keep going. The tools themselves were simple UNIX commands, deeply covered in the pre-training data thanks to half a century of UNIX documentation. While lots of people believe that Opus 4.7 unlocked the harness, I think the causality goes the other way.</p></li></ol><p>Despite how easy (and fun) it can be to get lost in each new model release or tool update, the reality is that most of it is just gradual improvement. That&#8217;s not to say it&#8217;s not tremendously valuable&#8212;GPT 4o unlocked ChatGPT in new ways&#8212;but those weren&#8217;t step changes in what is possible, they were the kind of slow progress that we always find hard to measure.</p><p>To that end, I have a few observations and thoughts from this list.</p><p>First off, each of these step changes is still very much building on the previous one. ChatGPT doesn&#8217;t happen without GPT-3, just like Claude Code doesn&#8217;t happen without o3. Reasoning models themselves are obviously evolved from prompting techniques. That can make AI progress feel Darwinian, but W. Brian Arthur&#8217;s point in <em><a href="https://sites.santafe.edu/~wbarthur/thenatureoftechnology.htm">The Nature of Technology</a></em> is more specific: technologies evolve by combination. New things are assembled out of existing components and the phenomena they harness; sometimes better variations win, and sometimes a new principle arrives that makes the old family tree less useful. The jet engine isn&#8217;t just a better propeller. His focus is mostly hardware, not software, but it still makes me wonder if we should expect more radical changes, or maybe we&#8217;re just settled into this architecture, and that&#8217;s how we should expect innovation to happen. <a href="https://importai.substack.com/p/import-ai-455-automating-ai-research?r=2hql&amp;utm_medium=ios&amp;triedRedirect=true">Here&#8217;s how Anthropic co-founder Jack Clark put it recently</a>:</p><blockquote><p>As a field, AI moves forward on the basis of doing ever larger experiments that utilize more and more inputs (e.g, data and compute). Every so often, humans come up with some paradigm-shifting idea which can make it dramatically more resource efficient to do things - a good example here is the transformer architecture and another is the idea of mixture-of-expert models. But mostly the field of AI moves forward through humans methodically going through some loop of taking a well performing system, scaling up some aspect of it (e.g, the amount of data and compute it is trained on), seeing what breaks when you scale it up, figuring out the engineering fix to allow it to scale, then scaling it again.</p></blockquote><p>Another interesting observation is that image and video models are mostly missing on this timeline. I don&#8217;t quite know what to make of these models, and that&#8217;s part of the reason I&#8217;m writing this post. Every time a new step change in capability comes along (like Nano Banana or, more recently, GPT Image 2.0), we see people say everything is going to change, and then mostly not a lot changes. As practitioners of AI and marketing, we are making good use of the current crop of models with a variety of customers for a variety of use cases, but it feels like the Copilot/ChatGPT/Claude Code moment just hasn&#8217;t arrived yet. Obviously, diffusion as an approach is amazing and probably deserves a place on the list, but the leaps feel much more gradual and make it hard to draw conclusions. With all that said, the release of GPT Image in March 2025 caused an explosion of usage in ChatGPT.</p><p>To that point, maybe the most fundamental takeaway from the timeline is that the real leaps in usage come from a kind of weird alchemy between model, market, and harness. The model has to be good enough to do the job, the harness has to expose that capability in a form people can actually use, and the market has to be in the right place to see the full value. Each one of these episodes caused a step change in token demand. Anthropic&#8217;s own <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">annualized run-rate numbers</a> make the point: the company went from approximately $9 billion at the end of 2025 to more than $30 billion by early April 2026.</p><p>Sholto Douglas, an Anthropic researcher focused on scaling reinforcement learning, made a version of this point on <a href="https://www.dwarkesh.com/p/sholto-trenton-2">a 2025 episode of the Dwarkesh Podcast</a>: Cursor had been around for a while, but with Claude 3.5 Sonnet, &#8220;the model was finally good enough that the vision they had of how people would program, hit.&#8221;</p><p>That&#8217;s why the causation is so hard to pin down. Lots of people are sure that Claude Code exploded in December because of the release of Opus 4.7. As a relatively early adopter of CC, I&#8217;m a lot more skeptical. <a href="https://every.to/podcast/how-to-use-claude-code-as-a-thinking-partner">Many of us were getting tremendous value out of the harness in June and July</a>. I think it&#8217;s much more likely that the real thing that happened in December is that a bunch of people just had more time to give the tool an honest go, and when they did, they were blown away.</p><p><a href="https://www.alephic.com/writing/the-magic-of-claude-code">Claude Code creator Boris Cherny calls this</a> &#8220;product overhang.&#8221; The model&#8217;s capabilities often remain untapped until the right code comes along to unlock them. What I think this timeline adds is that there is a market overhang, too. The model can be good enough, and the harness can exist, but the market also has to be ready to notice, care, and change behavior. The highlighted moments are when model, harness, and market meet.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XjcZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XjcZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 424w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 848w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 1272w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XjcZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XjcZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 424w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 848w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 1272w, https://substackcdn.com/image/fetch/$s_!XjcZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99b108d9-7615-4aaa-a19a-91a617cf1f52_2048x1152.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So what comes next? I don&#8217;t know. The three things I keep coming back to are reliable computer use, secure mobile agentic capability, and real document collaboration. I don&#8217;t know if any of them are the thing, but they all feel like potential unlocks for the next step change in token demand.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.forwarddeployed.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Forward Deployed! Subscribe for new posts and episodes at the intersection of AI, engineering, and the enterprise.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[10 Links for Saturday, May 2]]></title><description><![CDATA[The Goal, Flue, postmortems, DESIGN.md, and variance. Welcome to the weekend.]]></description><link>https://www.forwarddeployed.com/p/10-links-for-saturday-may-2</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/10-links-for-saturday-may-2</guid><dc:creator><![CDATA[Noah Brier]]></dc:creator><pubDate>Sat, 02 May 2026 13:50:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/add1cb45-9da0-46ff-b918-ea6c57ffffd8_892x1296.avif" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>You&#8217;re getting this email as a subscriber of ForwardDeployed, a newsletter at the intersection of AI, engineering, and the enterprise. </em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.forwarddeployed.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.forwarddeployed.com/subscribe?"><span>Subscribe now</span></a></p><p><br>Looking for something to read this weekend? Here&#8217;s 10 quick hits:</p><ol><li><p>Taylor Pearson (who will be on the next episode of the podcast with me) <a href="https://x.com/taylorpearsonme/status/2049126686567129370">uses Goldratt&#8217;s </a><em><a href="https://x.com/taylorpearsonme/status/2049126686567129370">The Goal</a></em><a href="https://x.com/taylorpearsonme/status/2049126686567129370">/theory of constraints to question parallel agent usage</a>: &#8220;I suspect a lot of agent usage right now is the same fallacy at higher resolution. Running 20 Claude Code sessions in parallel can feel productive because something is always happening. But, if the bottleneck in your work is judgment about what&#8217;s worth doing, more agents just generate more output for you to wade through.&#8221; (If you haven&#8217;t read <em><a href="https://www.amazon.com/Goal-Process-Ongoing-Improvement/dp/0884271951">The Goal</a></em>, it&#8217;s highly recommended. <a href="https://www.amazon.com/Eliyahu-M-Goldratts-BUSINESS-GRAPHIC/dp/0884272079/ref=pd_lpo_d_sccl_1/357-4617668-0259835?pd_rd_w=ka7NM&amp;content-id=amzn1.sym.4c8c52db-06f8-4e42-8e56-912796f2ea6c&amp;pf_rd_p=4c8c52db-06f8-4e42-8e56-912796f2ea6c&amp;pf_rd_r=JKHQF5DE1BSNZNEKY9D1&amp;pd_rd_wg=g76uA&amp;pd_rd_r=c7833004-b61f-4853-bb63-f7e633dbec11&amp;pd_rd_i=0884272079&amp;psc=1">There&#8217;s even a graphic novel version</a>.)</p></li><li><p><a href="https://maggieappleton.com/zero-alignment">GitHub&#8217;s Maggie Appleton hits on a point</a> I&#8217;ve been making a lot: We&#8217;ve known code wasn&#8217;t the bottleneck for 50 years (see: <em>Mythical Man Month</em>) &#8230; &#8220;Implementation is rapidly becoming a solved problem, right? Writing code is now fast, it&#8217;s getting cheap, and quality is going up and to the right. The hard question is no longer how to build it. It&#8217;s should we build it.&#8221;</p></li><li><p><a href="https://developers.openai.com/codex/app">Codex Desktop</a> is very good. <a href="https://flueframework.com/">Flue looks intriguing</a>: &#8220;Not another SDK. Build powerful, autonomous agents with Flue&#8217;s programmable TypeScript harness. Write once, deploy anywhere.&#8221;</p></li><li><p><a href="https://www.anthropic.com/engineering/april-23-postmortem">Very worth reading the Claude Code postmortem from Anthropic</a>. Lots of lessons about the way small changes can impact agentic systems.</p></li><li><p>Interesting to track some of the follow-ups to the <a href="https://www.youtube.com/watch?v=Hrbq66XqtCo">Jensen/Dwarkesh</a> conversation. I particularly <a href="https://www.chinatalk.media/p/no-jensen-not-all-compute-is-created">liked this ChinaTalk piece</a>.</p></li><li><p>It&#8217;s cool that <a href="https://openai.com/index/introducing-openai-privacy-filter/">OpenAI released this little PII redaction model</a>.</p></li><li><p><a href="https://blog.google/innovation-and-ai/models-and-research/google-labs/stitch-design-md/">I&#8217;d love to see DESIGN.md catch on</a> &#8230; curious to see what happens.</p></li><li><p>Steve Yegge (of <a href="https://gist.github.com/chitchcock/1281611">Stevey&#8217;s platform rant</a> fame) <a href="https://x.com/Steve_Yegge/status/2046260541912707471">punched back when Google folks said he was wrong about internal usage of Gemini models</a>. No clue what&#8217;s real/not real here, but following the way the Mag7 is using AI and how different each company&#8217;s approach is is fun to watch.</p></li><li><p>I&#8217;ve been thinking a lot about my <a href="https://www.alephic.com/variance-spectrum">variance spectrum concept</a> recently and how it relates to the spectrum of testing, linting, acceptance criteria, etc., that we integrate into agentic workflows. I put up a page to cover off on the basics, as it&#8217;s a model I continue to come back to.</p></li><li><p><a href="https://x.com/krishnanrohit/status/2047723242836901965">Aligned agents still build misaligned organizations</a>, from <a href="https://www.strangeloopcanon.com/about">Rohit Krishnan</a> (coming on the podcast soon) is fun. &#8220;So while each role did things that made sense to them, they ended up in a spot where they&#8217;re clearly misleading folks. The headline failure here is that the company&#8217;s billing system ends with the SLA clock stopped when the underlying world clearly says the outage stayed past the trigger when credit and review should have opened. (That is the value the billing system would return to say, an auditor.)&#8221;</p></li></ol><p>That&#8217;s it for now. Have a great weekend.</p><p>&#8212; Noah</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.forwarddeployed.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Forward Deployed! Subscribe and share.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Forward Deployed, Episode 4: The Special Forces Model]]></title><description><![CDATA[What Green Berets, Palantir's FDEs, and film showrunners teach us about building agentic systems]]></description><link>https://www.forwarddeployed.com/p/forward-deployed-episode-4-the-special</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/forward-deployed-episode-4-the-special</guid><dc:creator><![CDATA[Alephic HQ]]></dc:creator><pubDate>Sat, 21 Mar 2026 20:59:34 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/191703198/10c7d4457ee82c5e7a5c1ca21d3c74d0.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome to episode four of Forward Deployed. Noah sits down with Chris Papasadero to explore the deep parallels between Special Forces operations, enterprise software deployment, and creative direction&#8212;and what all of it means for building AI agents that actually work in the real world.</p><h2>Key Topics Covered</h2><ul><li><p>The Forward Deployed Engineer (FDE) model: How Palantir&#8217;s approach to embedding technical experts mirrors Special Forces doctrine</p></li><li><p>Force multiplication: Why Green Berets are designed to produce outsized output from minimal input &#8212; and what that means for AI agents</p></li><li><p>Comfort with ambiguity: The Special Forces selection pipeline, the Star Course, and why a 2% selection rate tests for the right traits</p></li><li><p>Cultural embedding: Why Palantir contractors in Afghanistan succeeded by understanding the operational environment, not just the software</p></li><li><p>Organizational structure and bureaucracy: NCO-led detachments, pushing planning to the lowest level, and the OSS Simple Sabotage Field Manual</p></li><li><p>Three layers of alignment: Shared cultural values, doctrine, and experience &#8212; Chris&#8217;s framework for aligning both teams and AI agents</p></li><li><p>Second and third-order effects: Why software engineering (like warfare) is a creative pursuit, not a six sigma factory process</p></li><li><p>Showrunners and creative direction: The role of holding both operational and creative vision across a large, autonomous team</p></li><li><p>Warhol Factory vs. Ford Factory: Why creative production is a better analogy for agentic systems than industrial automation</p></li></ul><h2>Timestamps</h2><p><em>Note: timestamps are approximate</em></p><ul><li><p>00:00 - Introduction and the origin of &#8220;Forward Deployed&#8221;</p></li><li><p>01:30 - Chris&#8217;s background: Special Forces and Palantir</p></li><li><p>05:00 - The Forward Deployed Engineer (FDE) model at Palantir</p></li><li><p>08:30 - Cultural embedding: Why Palantir worked in Afghanistan</p></li><li><p>12:00 - Special Forces as force multipliers</p></li><li><p>15:00 - The selection pipeline and comfort with ambiguity</p></li><li><p>18:30 - The Star Course: Navigating alone without external guidance</p></li><li><p>21:00 - Maintaining the big picture in the fog of war</p></li><li><p>25:00 - Organizational structure: NCO-led detachments and decentralized planning</p></li><li><p>29:00 - Planning for failure: Incorporating contingencies from the start</p></li><li><p>33:00 - The Simple Sabotage Field Manual and organizational bureaucracy</p></li><li><p>37:00 - Applying military frameworks to AI agents</p></li><li><p>41:00 - Three layers of alignment: Values, doctrine, and experience</p></li><li><p>45:00 - Second and third-order effect analysis</p></li><li><p>49:00 - Software engineering as a creative pursuit</p></li><li><p>52:00 - Showrunners, dailies, and creative direction</p></li><li><p>56:00 - The Warhol Factory model for agentic systems</p></li><li><p>59:00 - Wrap-up and key takeaways</p></li></ul><h2>Links &amp; References</h2><h3>Core References</h3><ul><li><p><a href="https://blog.palantir.com/a-day-in-the-life-of-a-palantir-forward-deployed-software-engineer-45ef2de257b1">Palantir &#8212; Forward Deployed Engineering</a></p></li><li><p><a href="https://www.cia.gov/stories/story/the-art-of-simple-sabotage/">OSS Simple Sabotage Field Manual</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-1-the-bitter">Forward Deployed Episode 1: The Bitter Lesson</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-2-claude">Forward Deployed Episode 2: Claude Code Skills</a></p></li><li><p><a href="https://www.forwarddeployed.com/p/forward-deployed-episode-3-context">Forward Deployed Episode 3: Context Engineering</a></p></li></ul><h3>Concepts &amp; Frameworks</h3><ul><li><p><a href="https://en.wikipedia.org/wiki/United_States_Army_Special_Forces">Special Forces (Green Berets)</a> &#8212; Force Multiplication doctrine</p></li><li><p>The Star Course &#8212; Special Forces land navigation assessment</p></li><li><p>NCO-led detachments &#8212; Decentralized command and planning</p></li><li><p>Andy Warhol&#8217;s Factory &#8212; Creative production model</p></li></ul><h3>Related Content</h3><ul><li><p><a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a> by Richard Sutton &#8212; General methods that leverage computation</p></li><li><p><a href="https://www.forwarddeployed.com/">Previous episodes of Forward Deployed</a></p></li></ul><h2>About the Hosts</h2><p><strong>Noah Brier</strong> is co-founder of <a href="https://alephic.com">Alephic</a>, an AI consulting company helping brands build custom AI systems. He writes about AI strategy and implementation.</p><p><strong>Chris Papasadero</strong> is this episode&#8217;s guest, bringing deep experience at the intersection of Special Forces operations, defense technology, and enterprise software deployment.</p><h2>Connect with the Hosts</h2><p>Noah Brier: <a href="https://linkedin.com/in/noahbrier">LinkedIn</a> | <a href="https://twitter.com/noahbrier">X/Twitter</a></p><p><em>Subscribe for weekly episodes exploring how AI is actually being deployed in the real world.</em></p>]]></content:encoded></item><item><title><![CDATA[Forward Deployed, Episode 3: Context Engineering]]></title><description><![CDATA[Why Karpathy&#8217;s term matters more than prompt engineering for agents]]></description><link>https://www.forwarddeployed.com/p/forward-deployed-episode-3-context</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/forward-deployed-episode-3-context</guid><dc:creator><![CDATA[Alephic HQ]]></dc:creator><pubDate>Sat, 27 Dec 2025 00:10:17 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/182661179/8c869f7a2d41c2eecfc9e9669ec6dda5.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome to episode three of Forward Deployed. Noah and Lance dive deep into context engineering&#8212;the art and science of filling context windows with just the right information for agents to take the next action.</p><p><strong>Key Topics Covered</strong></p><p>- Context engineering: Why Karpathy&#8217;s term matters more than prompt engineering for agents</p><p>- Three core techniques: Reducing context, isolating context, and offloading context to file systems</p><p>- Skills vs MCPs: Why progressive disclosure beats loading everything into context</p><p>- Claude Diary: Building agent memory through reflection and evolving CLAUDE.md</p><p>- Why subagents should isolate context, not anthropomorphize org charts</p><p>- Production patterns from Manus: Compaction, file system offloading, and sandbox architecture</p><p>- The convergence of agent primitives: Read, write, reduce, and recombine</p><p>- Why treating agents like humans works: Dual-use tools and new coworker onboarding</p><p><strong>Timestamps</strong></p><p>- 00:10 - What is context engineering and why should we care?</p><p>- 00:37 - Karpathy&#8217;s definition: Filling context windows with just the right information</p><p>- 01:15 - Context engineering vs prompt engineering: Tools bring context too</p><p>- 02:27 - Three buckets: Reducing, isolating, and offloading context</p><p>- 04:30 - Dynamic context loading: Skills as just-in-time context</p><p>- 06:38 - Skills as context offloading: Progressive disclosure of tools</p><p>- 09:15 - Sandboxes beyond Manus and Claude Code: Application integration patterns</p><p>- 12:12 - Context isolation through subagents: Not org charts</p><p>- 14:41 - Using isolation as a feature: Test-driven development with blank slate agents</p><p>- 16:33 - Real-world subagent challenges: Shared file systems and context bottlenecks</p><p>- 20:47 - Claude Diary: Automatic memory through session reflection</p><p>- 22:53 - CLAUDE.md bloat: Reducing unnecessary words like Strunk and White</p><p>- 28:06 - Dual-use tools: Why Claude Code uses bash, grep, and human-readable commands</p><p>- 31:02 - Skills discovery problems: From manual search to automated finding</p><p>- 34:22 - Agent primitives: Read, write, reduce, and creative recombination</p><p>- 36:32 - Tool search and action space design: How to find the right function</p><p>- 40:50 - Hooks as determinism: Injecting code into agent loops</p><p>- 41:41 - Skills everywhere: ChatGPT&#8217;s progressive disclosure discovery</p><p>- 43:12 - Nano Banana: Text on images unlocks new agentic workflows</p><p><strong>Links &amp; References</strong></p><p><strong>Core References</strong></p><ul><li><p>&#128279; <a href="https://twitter.com/karpathy/status/17932426212001755713">Karpathy tweet on context engineering</a> </p></li><li><p>&#127897;&#65039; <a href="https://www.youtube.com/watch?v=6_BcCthVvb8&amp;utm_source=chatgpt.com">Manus webinar on context engineering</a>  </p></li></ul><div id="youtube2-6_BcCthVvb8" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;6_BcCthVvb8&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/6_BcCthVvb8?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><ul><li><p>&#128196; <a href="https://cognition.ai/blog/dont-build-multi-agents Cognition">Cognition blog post on multi-agents</a>  </p></li><li><p>&#128196; <a href="https://simonwillison.net/2025/Jun/27/context-engineering/ Simon Willison&#8217;s Weblog">Simon Willison on skills and file systems</a> </p></li><li><p>&#128279; <a href="https://www.anthropic.com/news/model-context-protocol?utm_source=chatgpt.com">Anthropic Model Context Protocol announcement</a> </p></li><li><p>&#128279; <a href="https://github.com/barefootford/anthropic-mcp-docs GitHub">GitHub MCP docs (Anthropic</a>) </p></li></ul><p><strong>Technical Resources</strong></p><ul><li><p>&#128279; <a href="https://docs.claude.com/en/docs/claude-code/hooks">Claude Code hooks documentation</a>  </p></li><li><p>&#128279; <a href="https://github.com/anthropics/claude-skills-examples">Anthropic frontend design skills example</a> </p></li><li><p>&#128196; <a href="https://github.com/obra-ai/magic-skills-repo">Obra magic skills repository</a>  </p></li><li><p>&#128279; <a href="https://vercel.com/docs/concepts/edge-functions/sandbox-environment">Vercel sandboxes docs</a> </p></li></ul><p><strong>Tools &amp; Frameworks</strong></p><ul><li><p>&#128279;<a href="https://python.langchain.com/en/latest/index.html"> LangChain Docs (Python)</a></p></li><li><p>&#128279; <a href="https://cerebras.ai/technology">Cerebras Systems &#8211; Technology</a>  </p></li><li><p>&#128279; <a href="https://deepmind.google/blog/article/introducing-nano-banana">DeepMind Nano Banana info</a>  </p></li><li><p>&#128279; <a href="https://platform.openai.com/docs/guides/chat/skills">ChatGPT Skills directory</a> </p></li></ul><p><strong>Related Content</strong></p><p>- &#127897;&#65039; <a href="https://www.dwarkesh.com/p/karpathy">Karpathy on Dwarkesh Podcast</a> - Discussion of AI development principles</p><p>- &#128218; <a href="https://en.wikipedia.org/wiki/The_Elements_of_Style">The Elements of Style by Strunk and White</a> - &#8220;Reduce unnecessary words&#8221; as principle for CLAUDE.md</p><p>- &#128196; <a href="https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf">The Bitter Lesson by Richard Sutton</a> - General methods that leverage computation</p><p><strong>About the Hosts</strong></p><p>Noah Brier is co-founder of <a href="https://www.alephic.com/">Alephic</a>, an AI consulting company helping brands build custom AI systems. He writes about AI strategy and implementation.</p><p>Lance Martin is a founding engineer at <a href="https://www.langchain.com/">LangChain</a>, where he works on developer tools for building AI applications.</p><p><strong>Connect with the Hosts</strong></p><p>- Noah Brier: <a href="https://www.linkedin.com/in/noahbrier">LinkedIn</a> | <a href="https://twitter.com/heyitsnoah">X/Twitter</a> </p><p>- Lance Martin: <a href="https://www.linkedin.com/in/lance-martin">LinkedIn</a> | <a href="https://twitter.com/RLanceMartin">X/Twitter</a></p><p>* * *</p><p>Subscribe for weekly episodes exploring how AI is actually being deployed in the real world.</p>]]></content:encoded></item><item><title><![CDATA[Forward Deployed, Episode 2: Claude Code Skills and the Progressive Disclosure Problem ]]></title><description><![CDATA[Noah walks through the Alephic CLI he&#8217;s been building and reveals a wild hook-based routing system using Cerebras at 3,000 tokens per second.]]></description><link>https://www.forwarddeployed.com/p/forward-deployed-episode-2-claude</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/forward-deployed-episode-2-claude</guid><dc:creator><![CDATA[Alephic HQ]]></dc:creator><pubDate>Mon, 17 Nov 2025 03:26:07 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/179088941/2cde98465ed97fe6d3f1b8fb9b20a297.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome back to episode 2 of Forward Deployed. This week Noah and Lance dive deep into Claude Code skills, a deceptively simple feature that&#8217;s changing how we think about building AI agents. Noah walks through the Alephic CLI he&#8217;s been building and reveals a wild hook-based routing system using Cerebras at 3,000 tokens per second.</p><p>Plus: Why Andreessen thinks AI isn&#8217;t the Internet redux.</p><h2>Key Topics Covered</h2><ul><li><p>Claude Code skills and the progressive disclosure problem</p></li><li><p>How Noah built a hook to solve the 10% skill hit rate</p></li><li><p>Tier 1 vs Tier 2 action space: Why Manus and Anthropic converged on the same architecture independently</p></li><li><p>3,000 tokens per second: Using Cerebras Llama 120B as an invisible routing layer for every user message</p></li><li><p>The MCP/skill/command convergence: Are they all just different flavors of the same primitive?</p></li><li><p>Vision feedback loop: Turning Gemini into a Pentagram creative director to critique Claude&#8217;s web designs</p></li><li><p>Andreessen&#8217;s &#8220;computers v2&#8221; thesis: Why AI isn&#8217;t the Internet redux, it&#8217;s the first von Neumann architecture replacement in 80 years</p></li><li><p>Git workflows with Claude Code: Why Lance and Noah don&#8217;t worry about merge conflicts anymore</p></li></ul><h2>Timestamps</h2><ul><li><p><strong>00:11</strong> &#8211; Welcome to episode 2 on Claude Code skills</p></li><li><p><strong>00:28</strong> &#8211; What are Claude Code skills? Not much more than a folder full of prompts</p></li><li><p><strong>01:10</strong> &#8211; Lance: Skills as &#8220;instructing a new hire&#8221; with subfolder instructions</p></li><li><p><strong>01:25</strong> &#8211; Simon Willison and Jesse Vincent&#8217;s &#8220;superpowers&#8221; discovery</p></li><li><p><strong>04:45</strong> &#8211; Noah demos the Alephic CLI skill directory structure</p></li><li><p><strong>07:32</strong> &#8211; The hook-based skill search system using Cerebras</p></li><li><p><strong>08:19</strong> &#8211; Lance reveals: YAML front matter always loads into system prompt</p></li><li><p><strong>09:32</strong> &#8211; The 10% skill hit rate problem when you have 10+ skills</p></li><li><p><strong>10:08</strong> &#8211; Cerebras Llama 120B running at 3,000 tokens per second for invisible routing</p></li><li><p><strong>13:17</strong> &#8211; The universal pattern: Everyone&#8217;s trying to control context</p></li><li><p><strong>15:59</strong> &#8211; Tier 1 vs Tier 2 action space: Manus and Anthropic converge independently</p></li><li><p><strong>21:29</strong> &#8211; Noah&#8217;s big challenge: Getting models to consistently look for skills</p></li><li><p><strong>28:04</strong> &#8211; Hit rate drops to 10% even with only 4&#8211;5 skills</p></li><li><p><strong>30:54</strong> &#8211; Could progressive disclosure become built-in like chain of thought?</p></li><li><p><strong>34:06</strong> &#8211; Lance on externalizing context to file systems</p></li><li><p><strong>35:15</strong> &#8211; Vision feedback loop: Gemini as Pentagram creative director critiquing Claude&#8217;s designs</p></li><li><p><strong>37:57</strong> &#8211; Andreessen: AI isn&#8217;t the Internet, it&#8217;s computers v2</p></li><li><p><strong>42:08</strong> &#8211; Why Noah and Lance don&#8217;t worry about merge conflicts anymore</p></li></ul><h2>Links &amp; References</h2><h3>Core References</h3><ul><li><p>&#128196; <a href="https://www.anthropic.com/engineering">Anthropic Engineering Blog on Skills</a> &#8211; The engineering blog post Lance mentions about YAML front matter and system prompts</p></li><li><p>&#127897;&#65039; <a href="https://www.dwarkesh.com/podcast">Andrej Karpathy on Dwarkesh Podcast</a> &#8211; Discussion on small models and reasoning engines</p></li><li><p>&#127897;&#65039; <a href="https://www.youtube.com/watch?v=example">Andreessen on Cheeky Pint Podcast</a> &#8211; &#8220;AI is computers v2&#8221; thesis: First von Neumann architecture replacement in 80 years</p></li><li><p>&#127897;&#65039; <a href="https://www.anthropic.com/podcast">Claude Code Podcast with Boris and Kat</a> &#8211; Discussion on dual-use tools</p></li></ul><h3>Tools &amp; Frameworks</h3><ul><li><p>&#128279; <a href="https://claude.ai/code">Claude Code</a> &#8211; Anthropic&#8217;s AI coding assistant</p></li><li><p>&#128279; <a href="https://github.com/jvincent/superpowers">Jesse Vincent&#8217;s Superpowers</a> &#8211; Original skills plugin that inspired Noah</p></li><li><p>&#128279; <a href="https://manus.ai">Manus</a> &#8211; Consumer agent with multi-tier action space architecture</p></li><li><p>&#128279; <a href="https://cerebras.ai">Cerebras</a> &#8211; Llama 120B at 3,000 tokens per second</p></li><li><p>&#128279; <a href="https://github.com/puppeteer/puppeteer">Puppeteer MCP</a> &#8211; Browser automation MCP</p></li><li><p>&#128279; <a href="https://playwright.dev/">Playwright MCP</a> &#8211; Browser automation alternative</p></li><li><p>&#128279; <a href="https://cli.github.com/">GitHub CLI</a> &#8211; Command line tool Lance loves for PR management</p></li></ul><h3>Blog Posts</h3><ul><li><p><a href="https://simonwillison.net/">Simon Willison&#8217;s Blog</a> &#8211; First place Noah saw Claude Code skills coverage</p></li><li><p><a href="https://blog.cloudflare.com/">Cloudflare TypeScript Type Definitions Technique</a> &#8211; Alternative to classic MCP definitions</p></li></ul><h3>Companies Mentioned</h3><ul><li><p><a href="https://www.anthropic.com/">Anthropic</a> &#8211; Claude Code creator</p></li><li><p><a href="https://www.langchain.com/">LangChain</a> &#8211; Where Lance is a founding engineer</p></li><li><p><a href="https://www.alephic.com/">Alephic</a> &#8211; Noah&#8217;s AI consulting company</p></li><li><p><a href="https://www.pentagram.com/">Pentagram</a> &#8211; Design agency Noah used as creative director persona</p></li></ul><h3>Development Tools</h3><ul><li><p><a href="https://obsidian.md/">Obsidian</a> &#8211; Note-taking use case for Claude Code</p></li><li><p><a href="https://git-scm.com/docs/git-worktree">Git Work Trees</a> &#8211; How the team manages multi-branch development</p></li></ul><h2>About the Hosts</h2><p><strong>Noah Brier</strong> is co-founder of <a href="https://www.alephic.com/">Alephic</a>, an AI consulting company helping brands build custom AI systems. He writes about AI strategy and implementation.</p><p><strong>Lance Martin</strong> is a founding engineer at <a href="https://www.langchain.com/">LangChain</a>, where he works on developer tools for building AI applications.</p><h2>Connect with the Hosts</h2><ul><li><p><strong>Noah Brier:</strong> <a href="https://linkedin.com/in/noahbrier">LinkedIn</a> | <a href="https://twitter.com/heynoah">X/Twitter</a></p></li><li><p><strong>Lance Martin:</strong> <a href="https://linkedin.com/in/lancemartin">LinkedIn</a> | <a href="https://twitter.com/RLanceMartin">X/Twitter</a></p></li></ul><p>Subscribe for weekly episodes exploring how AI is actually being deployed in the real world.</p>]]></content:encoded></item><item><title><![CDATA[Forward Deployed, Episode 1: The Bitter Lesson]]></title><description><![CDATA[Welcome to the first episode of Forward Deployed, a podcast exploring the intersection of AI, software development, and the enterprise.]]></description><link>https://www.forwarddeployed.com/p/forward-deployed-episode-1-the-bitter</link><guid isPermaLink="false">https://www.forwarddeployed.com/p/forward-deployed-episode-1-the-bitter</guid><dc:creator><![CDATA[Noah Brier]]></dc:creator><pubDate>Mon, 03 Nov 2025 14:15:06 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/177811633/fb55501d5ba8e0169636bf0292fdef0d.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome to the first episode of Forward Deployed (<a href="https://www.youtube.com/watch?v=bw37s0yxpvY">YouTube</a>), a podcast exploring the intersection of AI, software development, and the enterprise.</p><p>We&#8217;re very excited to have you join us as we tackle the wild world of Forward Deployed engineering. We hope to make this a roughly bi-weekly show and look forward to inviting guests in the future. The idea is to dive deep into the realities of making this stuff work in companies, building on our expertise as builders both inside and outside the enterprise.</p><p>Thanks for listening, and please let us know what you think.</p><p>Thanks,</p><p>Noah &amp; Lance</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.forwarddeployed.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to Forward Deployed to be the first to hear about new episodes and articles.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p>In this inaugural episode, hosts <a href="http://x.com/heyitsnoah">Noah Brier</a> (Co-founder, <a href="http://alephic.com">Alephic</a>) and <a href="https://x.com/RLanceMartin">Lance Martin</a> (Founding Engineer, <a href="http://langchain.com">LangChain</a>) dive deep into one of AI&#8217;s most controversial ideas: The Bitter Lesson. They unpack Richard Sutton&#8217;s famous essay, debate whether LLMs truly follow its principles, and explore what this means for anyone building with AI today.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.youtube.com/watch?v=bw37s0yxpvY&quot;,&quot;text&quot;:&quot;Watch on YouTube&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.youtube.com/watch?v=bw37s0yxpvY"><span>Watch on YouTube</span></a></p><h2>Key Topics Covered</h2><ul><li><p>The Bitter Lesson: Why more compute beats clever algorithms (or does it?)</p></li><li><p>Richard Sutton&#8217;s surprising take on why LLMs aren&#8217;t &#8220;bitter lesson pilled&#8221;</p></li><li><p>The evolution from CNNs to transformers through Lance&#8217;s journey from Stanford to Uber&#8217;s self-driving program to LangChain</p></li><li><p>Chain of thought prompting vs reasoning models - why your prompts might be breaking</p></li><li><p>The real challenges of enterprise AI adoption</p></li><li><p>Why ICs are adopting AI faster than managers</p></li><li><p>Building for imperfection: Why optimizing for today&#8217;s models is a mistake</p></li></ul><h2>Timestamps</h2><ul><li><p>00:00 - Introductions and backgrounds</p></li><li><p>00:53 - Lance&#8217;s journey: Stanford PhD to Uber self-driving to LangChain</p></li><li><p>02:49 - Noah&#8217;s path from marketing to AI obsession</p></li><li><p>04:04 - What &#8220;forward deployed&#8221; really means</p></li><li><p>09:04 - The Bitter Lesson explained</p></li><li><p>11:31 - Why Sutton thinks LLMs aren&#8217;t following the bitter lesson</p></li><li><p>23:09 - Chain of thought prompting and the reasoning model revolution</p></li><li><p>24:19 - Building for future models, not current ones</p></li><li><p>45:20 - ICs vs managers in AI adoption</p></li></ul><h2>About the Hosts</h2><p>Noah Brier is co-founder of <a href="https://www.alephic.com/">Alephic</a>, an AI consulting company working with enterprise clients like PayPal, EY, Meta, and Amazon on AI-powered content intelligence and competitive analysis. Previously founded and sold Percolate (marketing tech). He also runs the <a href="https://brxnd.ai/">BRXND conference series</a> focused on marketing and AI.</p><p>Lance Martin is a founding engineer at <a href="https://www.langchain.com/">LangChain</a> with a PhD from Stanford. Former computer vision lead for Uber&#8217;s self-driving truck program.</p><h2>Links &amp; References</h2><h3>Core References</h3><ul><li><p>&#128196; <a href="https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf">The Bitter Lesson by Richard Sutton (2019)</a></p></li><li><p>&#127897;&#65039; <a href="https://www.dwarkesh.com/p/richard-sutton">Richard Sutton on Dwarkesh Podcast: &#8220;Father of RL thinks LLMs are a dead end&#8221;</a> (September 26, 2025)</p><ul><li><p><a href="https://www.youtube.com/watch?v=21EYKqUsPfg">YouTube version</a></p></li><li><p><a href="https://podcasts.apple.com/us/podcast/richard-sutton-father-of-rl-thinks-llms-are-a-dead-end/id1516093381?i=1000728584744">Apple Podcasts</a></p></li><li><p><a href="https://www.dwarkesh.com/p/thoughts-on-sutton">Dwarkesh&#8217;s follow-up reflections</a></p></li></ul></li><li><p>&#128196; <a href="https://karpathy.github.io/2015/05/21/rnn-effectiveness/">The Unreasonable Effectiveness of Recurrent Neural Networks by Andrej Karpathy (2015)</a></p></li><li><p>&#128196; <a href="https://a16z.com/services-led-growth/">Trading Margin for Moat: Why the Forward Deployed Engineer Is the Hottest Job in Startups (A16Z)</a></p></li></ul><h3>Related Podcast Appearances</h3><ul><li><p>&#127897;&#65039; <a href="https://every.to/podcast/how-to-use-claude-code-as-a-thinking-partner">Noah Brier on Every&#8217;s AI &amp; I: &#8220;Claude Code Can Be Your Second Brain&#8221;</a> (September 10, 2025) - Noah demonstrates his Claude Code-Obsidian setup for research and thinking</p><ul><li><p><a href="https://open.spotify.com/episode/4D8G4ZhgmFrMDNYrrQOTbe">Listen on Spotify</a></p></li><li><p><a href="https://podcasts.apple.com/ca/podcast/claude-code-can-be-your-second-brain/id1719789201?i=1000725911151">Apple Podcasts</a></p></li></ul></li></ul><h3>Blog Posts from the Hosts</h3><ul><li><p><a href="https://rlancemartin.github.io/2025/07/30/bitter_lesson/">Learning the Bitter Lesson</a> - Lance Martin</p></li><li><p><a href="https://rlancemartin.github.io/2025/06/23/context_engineering/">Context Engineering for Agents</a> - Lance Martin</p></li><li><p><a href="https://www.alephic.com/writing/thinking-ahead-building-ahead">Thinking Ahead, Building Ahead</a> - Charles Gallant</p></li><li><p><a href="https://www.alephic.com/writing/the-magic-of-claude-code">The Magic of Claude Code</a> - Noah Brier</p></li><li><p><a href="https://www.alephic.com/writing/strategic-software">Strategic Software</a> - Noah Brier</p></li><li><p><a href="https://www.alephic.com/writing/things-i-think-i-think-about-ai">Things I Think I Think About AI</a> - Noah Brier</p></li></ul><h2>Connect with the Hosts</h2><ul><li><p>Noah Brier: <a href="https://www.linkedin.com/in/noahbrier/">LinkedIn</a> | <a href="https://x.com/heyitsnoah">X/Twitter</a></p></li><li><p>Lance Martin: <a href="https://www.linkedin.com/in/lance-martin-64a33b5/">LinkedIn</a> | <a href="https://x.com/rlancemartin">X/Twitter</a></p></li><li><p>Alephic: <a href="https://www.alephic.com/">alephic.com</a></p></li><li><p>BRXND: <a href="https://brxnd.ai/">brxnd.ai</a></p></li></ul><div><hr></div><p>Subscribe for weekly episodes exploring how AI is actually being deployed in the real world.</p><p>Newsletter: Sign up for updates at <a href="https://forwarddeployed.com/">forwarddeployed.com</a></p>]]></content:encoded></item></channel></rss>