<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Tech Fundamentals]]></title><description><![CDATA[This newsletter breaks down complex tech fundamentals for investors, traders & VCs. You’ll find deep dives, risk analysis, and short and mid-term market context without hype.]]></description><link>https://www.techfundamentals.blog</link><image><url>https://substackcdn.com/image/fetch/$s_!xWDB!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0870fc2b-6ca8-452b-aea6-8572d56e44de_670x670.png</url><title>Tech Fundamentals</title><link>https://www.techfundamentals.blog</link></image><generator>Substack</generator><lastBuildDate>Wed, 15 Apr 2026 20:55:58 GMT</lastBuildDate><atom:link href="https://www.techfundamentals.blog/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Daniel Kolb]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[techfundamentals@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[techfundamentals@substack.com]]></itunes:email><itunes:name><![CDATA[Daniel Kolb]]></itunes:name></itunes:owner><itunes:author><![CDATA[Daniel Kolb]]></itunes:author><googleplay:owner><![CDATA[techfundamentals@substack.com]]></googleplay:owner><googleplay:email><![CDATA[techfundamentals@substack.com]]></googleplay:email><googleplay:author><![CDATA[Daniel Kolb]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Apple Using Google’s AI Isn’t a Surrender But a Pattern]]></title><description><![CDATA[Why Gemini in Siri looks more like Intel-in-Macs than Apple falling behind]]></description><link>https://www.techfundamentals.blog/p/apple-using-googles-ai-isnt-a-surrender</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/apple-using-googles-ai-isnt-a-surrender</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Sun, 18 Jan 2026 15:02:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QZ8j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QZ8j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QZ8j!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QZ8j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7891366,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.techfundamentals.blog/i/184956767?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QZ8j!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QZ8j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e70f9fe-a49e-44e1-a453-f96d33297e54_5184x3456.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Koshiro K - stock.adobe.com</figcaption></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>Apple and Google have confirmed a multi-year partnership in which Apple will use Google&#8217;s Gemini models to power parts of future versions of Siri and Apple Intelligence. The confirmation came via a rare joint statement from both companies <a href="https://blog.google/company-news/inside-google/company-announcements/joint-statement-google-apple/">https://blog.google/company-news/inside-google/company-announcements/joint-statement-google-apple/</a>.</p><p>The immediate reaction framed this as another checkpoint in the AI race. Google &#8220;wins&#8221; Apple. Apple &#8220;falls behind&#8221; hyperscalers. Stocks move accordingly.</p><p>That framing is understandable, but it misses a recurring Apple pattern. Apple often uses external technology as a bridge, not an endpoint. This is very important to remember since we can learn from how Apple approached new technologies in the history of the company.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Tech Fundamentals is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div><hr></div><h2><strong>Apple: Using External Models Without Giving Up Control</strong></h2><p>Apple has never optimized for owning every layer from day one. Historically, Apple has been comfortable relying on partners while it builds internal capability - and then switching once it can control the stack end to end.</p><p>The most obvious example is Intel. Apple shipped Macs on Intel CPUs for years before transitioning to Apple Silicon. That move wasn&#8217;t rushed. It happened only once Apple could deliver better performance per watt, tighter integration, and more predictable economics with its own M-series chips. The same pattern showed up with Google Maps before Apple Maps, and with GPUs before Apple&#8217;s custom graphics pipelines matured.</p><p>Seen in that context, Gemini looks like a familiar Apple move. Apple is accessing state-of-the-art reasoning capability now, while continuing to invest in its own foundation models and silicon-level acceleration. <strong>Apple was explicit that Gemini will not behave like a generic cloud API bolted onto Siri</strong>. Apple Intelligence still runs primarily on-device, with fallback to Apple&#8217;s own Private Cloud Compute when needed. Apple and Google emphasized that user data is not broadly shared and that Apple&#8217;s privacy guarantees remain intact<a href="https://observervoice.com/google-offers-privacy-assurance-to-iphone-users-following-gemini-partnership-with-apple-173649/">https://observervoice.com/google-offers-privacy-assurance-to-iphone-users-following-gemini-partnership-with-apple-173649/</a>.</p><p>From a systems perspective, Apple is borrowing capability without surrendering orchestration. Gemini provides model intelligence today. Apple Silicon, the Neural Engine, Core ML, and the OS decide where inference runs, how often, and under what constraints. If and when Apple&#8217;s own models reach sufficient maturity, swapping out the backend becomes a product decision rather than a structural rewrite.</p><p>For Apple stock, that matters. Temporary dependency is very different from structural dependency. Apple is not locking itself into a cost structure or vendor relationship it can&#8217;t later unwind.</p><div><hr></div><h2><strong>Google: Model Validation and Strategic Positioning</strong></h2><p>Google still gets something meaningful out of this arrangement, even if it isn&#8217;t permanent: validation.</p><p>In the joint statement, Apple concluded that Gemini provides &#8220;the most capable foundation&#8221; for its next generation of AI features <a href="https://blog.google/company-news/inside-google/company-announcements/joint-statement-google-apple/">https://blog.google/company-news/inside-google/company-announcements/joint-statement-google-apple/</a>. That endorsement carries weight precisely because Apple is selective and historically willing to walk away once it has an internal alternative.</p><p>For Google, this is less about locking Apple in and more about proving Gemini is good enough to sit at the core of a massive consumer platform. Even if Apple eventually transitions to its own models, Google benefits from the reputational signal and near-term strategic relevance.</p><p>For Alphabet investors, this looks like a positioning win rather than a long-term capture of Apple&#8217;s AI economics. Google strengthens its claim to frontier leadership, but without controlling Apple&#8217;s destiny.</p><div><hr></div><h2><strong>Microsoft: Still Central, But Not the Default Choice</strong></h2><p>Microsoft isn&#8217;t directly affected by the Gemini deal, but it does lose the assumption that Azure plus OpenAI automatically becomes the foundation for every major consumer assistant.</p><p>Apple already integrates ChatGPT for certain Siri queries, but it chose Gemini as the broader foundation layer. That suggests Apple evaluated multiple options and picked what worked best for its current constraints, not what aligned best with a long-term dependency.</p><p>This mirrors Apple&#8217;s past behavior with Intel. Apple partnered deeply, extracted value, and then moved on once it had a better internal alternative. Microsoft&#8217;s strength remains enterprise AI and developer ecosystems, but this deal reinforces that Apple will always keep optionality.</p><div><hr></div><h2><strong>Amazon: Indirect Exposure and Long-Term Implications</strong></h2><p>Amazon sits slightly outside this specific partnership, but the broader implication still applies.</p><p>Always-on consumer AI struggles with cloud-first economics. Apple&#8217;s decision to keep inference mostly on-device while temporarily sourcing model capability externally reinforces that lesson. It&#8217;s a reminder that centralized inference scales poorly when AI becomes ambient rather than occasional.</p><p><strong>For Amazon, this echoes some of the challenges Alexa faced</strong>. It also hints that future consumer AI systems may blend local execution with selective cloud augmentation, rather than relying entirely on hyperscaler infrastructure.</p><div><hr></div><h2><strong>OpenAI: Important, but Potentially Transitional</strong></h2><p>OpenAI remains part of Apple&#8217;s AI stack. Apple continues to use ChatGPT for certain requests that require broad world knowledge or creative generation<a href="https://wandb.ai/byyoung3/ml-news/reports/Apple-turns-to-Google-for-AI--VmlldzoxNTYxNDk2Nw">https://wandb.ai/byyoung3/ml-news/reports/Apple-turns-to-Google-for-AI--VmlldzoxNTYxNDk2Nw</a>.</p><p>But just like Gemini, OpenAI&#8217;s role looks more like a component than a permanent foundation. Apple&#8217;s architecture already supports multiple models, and its tooling is designed to swap backends as capabilities evolve.</p><p><strong>If Apple follows its historical playbook, both Gemini and OpenAI could eventually be replaced; Not abruptly, but once Apple&#8217;s internal models are good enough for its specific use cases.</strong></p><div><hr></div><h2><strong>How This Fits Apple&#8217;s Broader AI Architecture</strong></h2><p>This partnership fits cleanly into Apple&#8217;s long-term approach: control the stack, even if you don&#8217;t own every piece yet.</p><p>Apple isn&#8217;t trying to outspend hyperscalers on centralized compute. It&#8217;s trying to avoid inheriting a variable cost structure that grows with every user interaction. By pushing inference onto devices it already sells, Apple turns AI execution into a fixed, amortized cost tied to hardware cycles.</p><p>External models like Gemini fill the gap while Apple&#8217;s own foundation models mature. Once they do, Apple can internalize more of the stack without retraining users or rethinking the product.</p><p>This is the same transition Apple executed with CPUs, GPUs, and mapping infrastructure. The difference is that AI is more visible, but the strategy is the same.</p><div><hr></div><h2><strong>Implications for Apple, Hyperscalers, and AI Platforms</strong></h2><p>This deal doesn&#8217;t redraw the AI landscape overnight, but it clarifies incentives.</p><p>Apple buys time and capability without committing long term. Google gains validation and near-term relevance. Microsoft and Amazon see confirmation that platform control is never guaranteed. OpenAI remains important, but likely not permanent.</p><p>More broadly, this reinforces that AI is becoming infrastructure. As that happens, temporary partnerships, modular architectures, and cost containment matter more than public dominance narratives.</p><div><hr></div><h2><strong>What This Means for the Stocks Involved</strong></h2><p>There&#8217;s no single winner here.</p><p>Apple reduces risk around Siri while preserving optionality. Google gains credibility. The hyperscalers are reminded that consumer AI is expensive and that Apple will not accept open-ended cost exposure.</p><p>For investors, the key insight is that Apple is behaving exactly as it has before: rely on partners when necessary, build internally in parallel, and switch once control and economics improve.</p><div><hr></div><h2><strong>Closing Perspective</strong></h2><p>Apple&#8217;s decision to use Gemini for Siri looks less surprising when you view it through the company&#8217;s own history.</p><p>Apple has repeatedly used external technology as a bridge rather than a destination. Intel CPUs, Google Maps, even early GPUs followed the same pattern. Gemini fits neatly into that lineage.</p><p>Apple isn&#8217;t conceding the AI stack. It&#8217;s buying time while keeping control of where costs, latency, and privacy live. If history is a guide, this partnership may last only as long as it takes Apple to ship something better for its own constraints.</p><p>That&#8217;s not an exciting story. But in technology and investing, those are often the stories that matter.</p><p>If you want to learn more about Apple&#8217;s approach to AI, feel free to check out my article from a few weeks ago: <a href="https://substack.com/@techfundamentals/p-183697737">https://substack.com/@techfundamentals/p-183697737</a></p>]]></content:encoded></item><item><title><![CDATA[AI Agents Are About to Spend Real Money]]></title><description><![CDATA[Agentic Commerce and the Bull Case for Shopify, Amazon and eBay]]></description><link>https://www.techfundamentals.blog/p/ai-agents-are-about-to-spend-real</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/ai-agents-are-about-to-spend-real</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Tue, 13 Jan 2026 13:06:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jeAr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jeAr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jeAr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 424w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 848w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 1272w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jeAr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp" width="1456" height="628" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:628,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1089704,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.techfundamentals.blog/i/184430774?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jeAr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 424w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 848w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 1272w, https://substackcdn.com/image/fetch/$s_!jeAr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aa79dbc-ae19-4caa-bd36-5b4471f47774_4640x2000.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>Agentic commerce sounds abstract, but the underlying idea is straightforward. Instead of software that helps you shop by surfacing options or recommendations, you now have software that can actually act. An AI agent like ChatGPT or Gemini can search, compare, decide, and complete a purchase based on rules the user sets ahead of time.</p><p>That shift matters because ecommerce has always been organized around human-driven steps. Search, browse, click, compare, checkout. Once software starts collapsing those steps into a single automated flow, the question becomes less about user experience and more about which systems the agents trust to execute reliably.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Tech Fundamentals is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2><strong>Shopify: Designing for AI as the Customer</strong></h2><p>Shopify has been unusually clear about where it wants to sit in this new setup. In its announcement on <strong>AI commerce at scale</strong>, Shopify focused less on storefront design and more on making commerce primitives readable and usable by AI systems.</p><p>The idea is that an agent should be able to access a merchant&#8217;s catalog, understand pricing and availability, and complete a transaction without forcing the merchant to rebuild anything or manage one-off integrations. Shopify wants to handle that complexity once, at the platform level, and let everyone else plug in.</p><p>As Shopify VP Vanessa Lee put it:</p><blockquote><p>&#8220;Agentic commerce has so much potential to redefine shopping, and we want to make sure it can scale to every product a customer might want to purchase.&#8221;<br><br>- Shopify News, <em>AI commerce at scale</em> Source: <a href="https://www.shopify.com/news/ai-commerce-at-scale">https://www.shopify.com/news/ai-commerce-at-scale</a></p></blockquote><p>What&#8217;s interesting here is that Shopify isn&#8217;t trying to own the AI interface itself. It&#8217;s positioning itself as the execution layer that sits underneath whatever interface the user prefers. From an investment perspective, that matters because it changes Shopify&#8217;s role over time. Early on, it&#8217;s a merchant tool. At scale, it starts to look more like shared infrastructure that other systems depend on.</p><p>If agentic commerce grows, merchants won&#8217;t want to think about which AI model, search engine, or assistant their customers are using. They&#8217;ll want a backend that &#8220;just works&#8221; across all of them, and that&#8217;s the slot Shopify is aiming for.</p><div><hr></div><h2><strong>Adobe (Magento): Capable, but Harder to Standardize</strong></h2><p>Adobe&#8217;s exposure to ecommerce comes mainly through Magento, which serves a different part of the market. Magento is flexible and powerful, especially for large brands with custom needs, but that flexibility comes at the cost of uniformity.</p><p>That tradeoff becomes more visible in an agent-driven world. Autonomous systems work best with predictable, standardized surfaces. Magento deployments tend to vary widely, with integrations handled by agencies or internal teams and business logic spread across multiple layers.</p><p>Adobe&#8217;s broader AI push is focused on content, personalization, and analytics, which are valuable but sit upstream from the actual transaction. For agentic commerce, the bottleneck isn&#8217;t creativity or targeting, it&#8217;s execution.</p><p>As a result, agentic commerce is more of an indirect tailwind for Adobe. It supports demand for orchestration tools, but unless Adobe aggressively simplifies how Magento exposes commerce functionality to agents, it&#8217;s unlikely to become the default transaction backend in the same way Shopify is trying to be.</p><div><hr></div><h2><strong>Amazon: Built for This, but With Different Tradeoffs</strong></h2><p>Amazon already operates close to what agentic commerce looks like in practice. The company has spent years optimizing for speed, convenience, and minimal user effort, all backed by a tightly integrated logistics and payment system.</p><p>The potential tension isn&#8217;t about whether AI agents increase ecommerce volume. It&#8217;s about who controls the interaction layer. If purchasing increasingly happens inside AI assistants or search tools run by others, Amazon may still fulfill the order but lose some control over the customer relationship.</p><p>That helps explain why Amazon has been cautious and sometimes defensive around third-party AI shopping tools. It suggests the company sees agentic commerce as strategically important, but also as something that needs to be shaped carefully to avoid turning Amazon into a commoditized backend.</p><p>From an investor&#8217;s point of view, this points to durability rather than disruption. Amazon still owns the hardest assets in ecommerce, especially logistics. Agentic interfaces change how demand is routed, but they don&#8217;t make those assets less valuable.</p><div><hr></div><h2><strong>eBay: A Neutral Marketplace That Fits Agents Well</strong></h2><p>eBay sits in an interesting middle ground when it comes to agentic commerce. It doesn&#8217;t own logistics like Amazon, and it doesn&#8217;t act as a full-stack merchant backend like Shopify. Instead, it runs a large, relatively neutral marketplace with standardized listings, pricing, and checkout flows.</p><p>That neutrality may actually be an advantage in an agent-driven world. AI agents are good at structured comparison, and eBay&#8217;s inventory is already organized around searchable attributes, condition, price ranges, and seller reputation. From an agent&#8217;s perspective, that&#8217;s a clean environment to operate in.</p><p>eBay management has hinted in the past that AI-driven discovery and automation are areas of focus, especially around search relevance and buyer efficiency. The company&#8217;s incentives are also simpler: more completed transactions, regardless of where discovery happens. Unlike Amazon, eBay has less reason to protect a single dominant interface, and unlike Shopify, it doesn&#8217;t need to serve millions of independent backends.</p><p>For eBay stock, agentic commerce isn&#8217;t a dramatic growth narrative, but it could quietly improve conversion efficiency. If agents increasingly act as comparison engines that route demand toward the best value option, eBay&#8217;s broad, price-transparent marketplace could benefit without requiring heavy reinvention. The risk is mostly one of execution: eBay needs to make sure its APIs, listings, and checkout flows are easy for agents to interact with, or it risks being bypassed in favor of more standardized platforms.</p><p>In that sense, eBay feels less like a headline winner and more like a system that could age well if agentic commerce grows gradually rather than all at once.</p><div><hr></div><h2><strong>Infrastructure Starts to Matter More Than Storefronts</strong></h2><p>One side effect of agentic commerce is that attention shifts away from front-end design and toward the pipes underneath. Payments, identity, fraud prevention, and authorization become more central when purchases happen with fewer explicit human actions.</p><p>Market observers have started to describe this as a convergence of decision-making and execution, where systems need to be both intelligent and trustworthy. Analysts at McKinsey have highlighted agentic commerce as an area where standardised transaction infrastructure becomes increasingly important as autonomy increases.</p><p>This tends to reward platforms that are consistent and boring in the right ways. Clean APIs, stable data models, and predictable behaviour matter more than flashy features when software is the one doing the buying.</p><div><hr></div><h2><strong>What to Take Away</strong></h2><p>Agentic commerce doesn&#8217;t remove ecommerce platforms from the picture. It changes which layers capture leverage.</p><p>Interfaces become more fluid and interchangeable, while execution concentrates in systems that agents can rely on without special handling. Shopify is leaning directly into that shift by trying to be the neutral execution layer. Adobe has strong pieces but less cohesion at the transaction boundary. Amazon remains dominant, though more constrained by its own scale and incentives.</p><p>For investors, it&#8217;s worth thinking of agentic commerce less as a feature upgrade and more as an interface transition. Over time, the companies that quietly control how intent turns into a paid order are the ones that tend to compound, even if they don&#8217;t own the most visible part of the experience.</p>]]></content:encoded></item><item><title><![CDATA[After AI, Investors Are Eyeing Quantum Computing. Here’s What Actually Matters.]]></title><description><![CDATA[Which stocks to watch and why the next great computing breakthrough could reward patience more than hype]]></description><link>https://www.techfundamentals.blog/p/after-ai-investors-are-eyeing-quantum</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/after-ai-investors-are-eyeing-quantum</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Fri, 09 Jan 2026 16:16:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gI0G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gI0G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gI0G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gI0G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1969386,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.techfundamentals.blog/i/184036405?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gI0G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!gI0G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f979cb5-a474-464d-9058-700c5ecc6d77_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p><br>Quantum computing sits in a strange place. It&#8217;s early, it&#8217;s hard, and it&#8217;s absolutely not following a clean startup-to-scale story. At the same time, it&#8217;s one of the few areas in tech where the upside isn&#8217;t just better software, but access to entirely new classes of problems.</p><p>That combination makes it frustrating and compelling at the same time. If you&#8217;re an investor, the question usually isn&#8217;t &#8220;does this work?&#8221; but &#8220;where does value realistically accumulate while this is still messy?&#8221;</p><p>I&#8217;m going to explain quantum computing in simple terms for you and will describe the risks and opportunities from a technical perspective. Of course, I&#8217;ll also mention some stocks that are worth looking into if you want participate in the future of quantum computing.</p><div><hr></div><h2><strong>What quantum computing actually is</strong></h2><p>Quantum computing is an attempt to compute using physical systems that behave fundamentally differently from classical electronics. Instead of abstracting everything down to bits, quantum machines lean directly into quantum mechanical behaviour.</p><p>The important implication is that progress is constrained by nature, not just engineering talent. If quantum computing ever becomes commercially useful at scale, it won&#8217;t be because someone shipped a clever app. It will be because a small number of teams solved problems that most companies cannot even meaningfully attempt.</p><p>From an investment perspective, that matters. <strong>Hard problems tend to concentrate value.</strong></p><div><hr></div><h2><strong>How quantum computing works, in very simple terms</strong></h2><p>A classical computer explores one path at a time, very quickly.</p><p>A quantum computer explores many possible paths at once, but only for very specific kinds of problems. It uses &#8220;Qubits&#8221; for that. Qubits can exist in overlapping states, and groups of qubits can be linked together in ways that allow correlations classical systems cannot replicate.</p><p>When everything lines up, this lets quantum machines search complex solution spaces far more efficiently than brute-force classical approaches.</p><p><strong>That&#8217;s why areas like materials science, chemistry simulation, cryptography, and optimization keep coming up</strong>.</p><p>The downside is: Qubits don&#8217;t like the real world. Keeping them coherent requires extreme conditions and constant correction. This makes quantum computing very tricky to control and develop and many of the brightest minds are working actively on its problems.</p><div><hr></div><h2><strong>Quantum computing and AI: complementary, not competitive</strong></h2><p>Quantum computing isn&#8217;t going to replace GPUs for training AI models. That&#8217;s not its role.</p><p>Where quantum could matter is in areas adjacent to AI: optimization, sampling, and simulation problems that are hard even for large classical clusters. Think better materials for chips, better batteries, more efficient logistics, or improved models for physical systems.</p><p>There&#8217;s also a feedback loop forming. AI is increasingly used to design quantum experiments, tune control systems, and search for better quantum algorithms. As quantum hardware improves, that loop tightens.</p><p>This is less about one technology eating the other, and more about both expanding the set of problems we can realistically attempt.</p><div><hr></div><h2><strong>Why quantum computing is still risky, but less fragile than it looks</strong></h2><p>Investing in Quantum Computing is risky, but not in the way most people might think.</p><p>The risk isn&#8217;t that quantum is fake or that progress will suddenly stop. The real risk is mistaking technical progress for commercial readiness, or assuming timelines compress the way software timelines do.</p><p>That said, there&#8217;s a more positive angle that often gets overlooked. <strong>Much of the foundational work has already been done</strong>. Error rates are improving. Tooling is better and hybrid classical-quantum workflows are becoming more practical. Governments and large enterprises are no longer just funding research, they&#8217;re testing early use cases.</p><p>This doesn&#8217;t mean revenue explodes this or next year. It does mean the probability that quantum becomes <em>useful</em> keeps going up, even if adoption stays slow and uneven.</p><p>For <strong>patient value investors</strong>, that distinction matters.</p><div><hr></div><h2><strong>The real risk: quantum becoming the next place speculation hides</strong></h2><p>The biggest risk around quantum computing isn&#8217;t that the technology is imaginary, it&#8217;s that capital is very good at outrunning reality. Over the last few decades, markets have repeatedly shown that hype often produces better short-term returns than legitimate businesses, at least until the story collapses under its own weight. Dot-coms, crypto, NFTs, and now generative AI all followed a similar pattern: plausible technology, enormous expectations, weak economics, and capital piling in long before usefulness was proven.</p><p>AI is now starting to show those cracks. Despite hundreds of billions in investment, productivity gains remain narrow and difficult to generalize, while costs scale aggressively. Several studies (like the one from MIT) suggest that most AI pilots fail to deliver meaningful profitability or output improvements, and even widely adopted products struggle to break even without optimistic pricing assumptions. As improvement curves flatten and unit economics remain uncomfortable, capital naturally looks for the next narrative that can absorb belief without immediate verification.</p><p>Quantum computing fits that role almost too well. It is real, genuinely hard, deeply technical, and extremely difficult for outsiders to falsify. Hardware progress is visible and easy to market, while the true bottleneck, useful software and algorithms, remains slow, uncertain, and largely resistant to brute-force investment. Even with functional quantum hardware, quantum computers cannot run conventional code and only outperform classical systems on very specific problem classes. For many of the areas they are most often associated with, including AI training and general machine learning, there is still no clear evidence that relevant quantum algorithms even exist.</p><p>This creates a familiar mismatch. Scientific progress continues, but commercial usefulness lags far behind expectations. Timelines stretch into decades, while public markets price outcomes years in advance. In that gap, speculation thrives. The danger for investors isn&#8217;t that quantum fails, but that it gets used as an escape hatch for disappointed AI expectations, pulling future possibilities forward to justify present valuations.</p><p>Quantum computing can still become important, even transformative, without rewarding most early capital. That&#8217;s the uncomfortable part. The technology doesn&#8217;t owe investors a clean adoption curve, and history suggests it rarely provides one.</p><div><hr></div><h2><strong>How investors can participate</strong></h2><p>This is where nuance helps. The most visible quantum stocks are not interchangeable. They represent very different philosophies. Here are some of the more famous ones:</p><h3><strong>IonQ (IONQ): betting on physics elegance and long-term scalability</strong></h3><p><strong>IonQ</strong> is often seen as the &#8220;cleanest&#8221; quantum story. Their approach uses trapped ions, which tend to have very high qubit fidelity and long coherence times.</p><p>From the outside, IonQ looks slow. From the inside, they&#8217;re optimising for scalability and error reduction rather than rushing qubit counts. That&#8217;s appealing if you believe stable qubits matter more than flashy numbers.</p><p>IonQ also benefits from a relatively hardware-light model compared to some competitors. Their systems integrate well with cloud platforms, which makes experimentation easier for customers and lowers friction for adoption.</p><p>As an investment, IonQ is a long-duration bet on doing things the hard way early so scaling hurts less later. It won&#8217;t pay off quickly, but if trapped-ion architectures win, the payoff could be meaningful.</p><div><hr></div><h3><strong>Rigetti Computing (RGTI): vertical integration and iteration speed</strong></h3><p><strong>Rigetti Computing</strong> takes a different approach. They build superconducting qubit systems and control much more of the stack themselves, from chip fabrication to software.</p><p>That vertical integration gives Rigetti speed and flexibility. They can iterate quickly, experiment with architectures, and optimize across layers. The downside is cost and complexity. Superconducting systems are demanding, and scaling them reliably is non-trivial.</p><p>Rigetti feels more like a traditional hardware startup: faster feedback loops, higher burn, more visible stumbles. For investors, this is a higher-volatility bet, but also one with clearer signals along the way as systems improve or don&#8217;t.</p><p>If quantum adoption accelerates through cloud-accessible, general-purpose systems, Rigetti is positioned to benefit.</p><div><hr></div><h3><strong>D-Wave Quantum (QBTS): the most commercially grounded path</strong></h3><p><strong>D-Wave Quantum</strong> is often misunderstood because it doesn&#8217;t build universal quantum computers in the same way as IonQ or Rigetti. Instead, D-Wave focuses on quantum annealing, which is specialized but useful for optimization problems.</p><p>This specialization is a feature, not a flaw. D-Wave already has customers using their systems for real-world optimization tasks. Revenue exists. Use cases are concrete. The tradeoff is that annealing won&#8217;t solve everything quantum can theoretically do.</p><p>From an investor&#8217;s perspective, D-Wave is less about moonshots and more about pragmatic adoption. It&#8217;s a bet that early, narrow utility can compound into broader relevance over time.</p><div><hr></div><h3><strong>NVIDIA (NVDA): quantum doesn&#8217;t replace GPUs, it leans on them</strong></h3><p>NVIDIA is not a quantum computing company in the traditional sense, but it&#8217;s deeply embedded in the quantum ecosystem.</p><p>Most quantum work today still depends heavily on classical computing. You simulate quantum systems, design quantum circuits, run error-correction layers, and coordinate experiments using massive amounts of classical compute. GPUs sit right in the middle of that workflow.</p><p>NVIDIA&#8217;s role is subtle but powerful. They sell the picks and shovels for quantum research, from simulation libraries to high-performance hardware used by labs, startups, and hyperscalers. Even if large-scale quantum computing takes longer than expected, NVIDIA still benefits as long as money keeps flowing into research and experimentation.</p><p>From an investor perspective, this is attractive because NVIDIA doesn&#8217;t need a quantum breakthrough. Incremental progress is enough to justify continued spending on tooling and infrastructure.</p><div><hr></div><h3><strong>Microsoft (MSFT): quantum as a platform feature, not a product</strong></h3><p>Microsoft approaches quantum the same way it approaches most infrastructure problems: abstract it, wrap it in tooling, and make it accessible through a platform.</p><p><a href="https://azure.microsoft.com/en-us/solutions/quantum-computing">Azure Quantum</a> is less about owning the &#8220;best&#8221; quantum hardware and more about being the place where quantum experiments happen. Microsoft connects multiple quantum hardware providers, classical compute, developer tools, and enterprise workflows under one roof.</p><p>What&#8217;s interesting here is the incentive structure. Microsoft doesn&#8217;t need to guess which quantum architecture wins. If quantum becomes useful at all, customers are likely to access it through existing cloud relationships. That gives Microsoft leverage without forcing them to bet everything on one technical path.</p><p><strong>This makes quantum at Microsoft feel less like a moonshot and more like a long-dated platform option that fits naturally into their cloud strategy.</strong></p><div><hr></div><h3><strong>IBM (IBM): slow, steady, and deeply committed</strong></h3><p>IBM has been working on quantum longer than almost anyone else, and it shows in how methodical their approach is.</p><p>IBM treats quantum as an extension of its research-first DNA. They publish roadmaps, expose systems to developers, and push heavily on education and ecosystem building. Their superconducting qubit approach is well understood, even if it&#8217;s technically demanding.</p><p>From an investment lens, IBM&#8217;s strength is credibility. Enterprises, governments, and universities already trust IBM with long-term infrastructure projects. Quantum fits naturally into that relationship model.</p><p>IBM is unlikely to surprise anyone with sudden breakthroughs, but it&#8217;s also unlikely to abandon the field. If quantum adoption happens through conservative, institutional channels, IBM is positioned to be a core beneficiary.</p><div><hr></div><h3><strong>Google (GOOG): quantum as a long-term research advantage</strong></h3><p>Google treats quantum computing as fundamental research with potential strategic upside, not as a near-term business line.</p><p>Google&#8217;s quantum work is deeply tied to their strengths in physics, systems engineering, and large-scale infrastructure. Their focus has been on proving hard technical milestones, even when those milestones don&#8217;t immediately translate into products.</p><p>What makes Google interesting is optionality. If quantum ends up mattering for materials science, optimization, or future compute workloads, Google has the talent and infrastructure to integrate it internally before it ever becomes a commercial offering.</p><p>For investors, this means quantum success would likely show up indirectly, through better internal tools, improved efficiency, or future services layered into Google Cloud rather than as a standalone revenue line.</p><div><hr></div><h2><strong>Closing thoughts</strong></h2><p>Quantum computing is neither a punchline nor a saviour. It&#8217;s a real technological frontier with legitimate long-term potential, operating under constraints that don&#8217;t bend to capital or narrative pressure. The physics is difficult, the engineering is slow, and the software problem is still largely unsolved. None of that makes quantum uninteresting. It just makes it incompatible with the way markets like to price stories.</p><p>What complicates things now is timing. As confidence in near-term AI payoffs softens, quantum risks absorbing expectations that have nowhere else to go. That doesn&#8217;t change what quantum can eventually become, but it does increase the chance that valuations get ahead of usefulness, especially for companies whose entire identity rests on long-dated breakthroughs.</p><p>For investors, the challenge isn&#8217;t deciding whether quantum matters. It&#8217;s recognising where value can realistically accrue while timelines stretch and narratives rotate.</p><p>Quantum computing will likely matter in ways that don&#8217;t look obvious today. The mistake is assuming that inevitability translates cleanly into investable returns, or that the market will wait patiently for reality to catch up.</p>]]></content:encoded></item><item><title><![CDATA[Apple’s AI Strategy Isn’t Loud. It’s Durable.]]></title><description><![CDATA[Why on-device intelligence favors control over spectacle and why this is a good thing in an AI first world]]></description><link>https://www.techfundamentals.blog/p/apples-ai-strategy-isnt-loud-its</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/apples-ai-strategy-isnt-loud-its</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Tue, 06 Jan 2026 19:34:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9DAR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9DAR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9DAR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9DAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg" width="1456" height="903" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:903,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2435279,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183697737?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9DAR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9DAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F798fec17-d778-4f41-ac28-e1fcc8dc3e26_7996x4957.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">gguy - stock.adobe.com</figcaption></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>Most of the AI conversation today is framed as a race. Bigger models, more parameters, faster benchmarks, larger clusters. If you look at it that way, Apple often seems like it&#8217;s lagging behind the hyperscalers like Google, Amazon or Microsoft.</p><p>But that framing misses what Apple is actually optimizing for.</p><p>Apple isn&#8217;t trying to build the smartest model in the data center. It&#8217;s trying to make intelligence cheap, predictable, and always available on devices that already sit in people&#8217;s pockets. That leads to very different technical choices, especially around where inference runs, how costs scale, and which risks are acceptable.</p><p>Once you stop treating AI as a product category and start treating it as infrastructure, Apple&#8217;s focus on on-device AI starts to look less conservative. This article looks at those choices from a systems and fundamentals perspective, and what they imply for Apple, hyperscalers (like Google / Amazon / Microsoft), and investors.</p><div><hr></div><h2><strong>How is Apple&#8217;s AI approach different from other Big Tech companies?</strong></h2><p>Apple&#8217;s AI strategy emphasizes <strong>on-device intelligence</strong> rather than relying on massive cloud inference. This approach is rooted in real <strong>technical and systems trade-offs</strong> &#8212; latency, privacy, bandwidth, and hardware economics &#8212; that become especially important at consumer scale.</p><div><hr></div><h2><strong>The Hardware Foundation: Apple Silicon and the Neural Engine</strong></h2><p>Central to Apple&#8217;s strategy is the integration of <strong>specialized neural acceleration in Apple Silicon</strong>, beginning with the A-series and continuing through the famous M-series chips. If you own an Apple device you&#8217;ve surely come across Apple Intelligence and that it runs and is secured on your device - that&#8217;s what I&#8217;m referring to here.</p><p>The <strong>Apple Neural Engine (ANE)</strong> is a dedicated neural processing unit designed to accelerate neural network inference efficiently and with low power consumption. Although Apple doesn&#8217;t publish full architectural details, projects analyzing the ANE describe it as a <strong>neural processing unit tuned for matrix operations and machine learning tasks</strong>, distinct from GPUs or CPUs.</p><p>Research and developer documentation from Apple shows that machine learning models &#8212; including <strong>transformers and other AI workloads &#8212; can be optimized and deployed locally</strong> using the Neural Engine, CPU, and GPU together.</p><p>Apple has also released posts showing <strong>on-device execution of LLMs</strong>, such as an example where a mid-sized 8B-parameter model runs locally on Apple silicon at usable real-time throughput. (<strong><a href="https://machinelearning.apple.com/research/core-ml-on-device-llama">Source</a></strong>)</p><div><hr></div><h2><strong>Delivering Low and Predictable Latency</strong></h2><p>One of the biggest practical advantages of on-device inference is <strong>latency elimination</strong>. This means the time it takes the AI to execute a task.</p><p>Cloud-based AI like ChatGPT and Gemini inevitably on the other hand introduces network hops and queuing delays, which are particularly noticeable in interactive UI features: If you&#8217;ve been using ChatGPT before, you probably know the typical loading spinner that appears while ChatGPT is processing your chat - that&#8217;s the latency.</p><p>When <strong>inference</strong> (meaning the execution of the AI model like processing a chat instead of training it on new things) runs locally, there&#8217;s <strong>no network dependency</strong>, so response time becomes driven purely by local hardware performance rather than internet speed or server load. This matters most when AI functionality is tied to the <strong>user experience itself</strong> &#8212; for example, in instant text completion or real-time image analysis &#8212; where even small delays are perceptible and degrade UX.</p><div><hr></div><h2><strong>Privacy as an Architectural Constraint</strong></h2><p>Apple&#8217;s privacy commitments are not just marketing language; they affect how the entire AI stack is designed. According to Apple&#8217;s privacy materials:</p><blockquote><p><em>Apple products, including Apple Intelligence, are engineered to protect user privacy by keeping as much processing on device as possible and minimizing the data sent to servers. (<strong><a href="https://www.apple.com/privacy/features">Source</a></strong>)</em></p></blockquote><p>Technical documentation shows that <strong>Core ML and the machine learning frameworks are explicitly designed to run models entirely on device</strong>, with no network connection needed, which supports both latency and privacy goals. (<strong><a href="https://developer.apple.com/machine-learning/core-ml">Source</a></strong>)</p><p>Developer-focused analyses similarly emphasize that <strong>&#8220;every inference happens on your device using the Neural Engine, CPU, and GPU in concert,&#8221;</strong> meaning personal data stays local and isn&#8217;t sent to remote servers for processing. (<strong><a href="https://medium.com/%40cognidownunder/apple-just-handed-every-developer-a-3-billion-parameter-ai-model-no-cloud-required-b2118eaac574">Source: Medium</a></strong>)</p><p>Apple&#8217;s broader AI architecture &#8212; including its Private Cloud Compute infrastructure &#8212; is built around the dual modalities of on-device processing and secure offload only when needed, ensuring that data remains private and ephemeral even in cloud interactions. (<strong><a href="https://security.apple.com/documentation/private-cloud-compute">Source: Apple Security Research</a></strong>)</p><div><hr></div><h2><strong>Bandwidth and Cost: Inference Cost</strong></h2><p>At scale - in Apple&#8217;s case that means <strong>hundreds of millions of active devices</strong> - the economics of AI shift dramatically. In cloud-first systems, each <strong>inference</strong> call (meaning executing a AI task) adds cost: compute, bandwidth, and storage. For every photo analyzed, text processed, or voice snippet interpreted, there&#8217;s a per-request cost in elastic cloud infrastructure.</p><p>When <strong>inference</strong> happens on the device itself in comparison to cloud based services like the ones from OpenAI or Google, these are the results:</p><ul><li><p><strong>Bandwidth costs drop to near zero</strong> after models are shipped or updated.</p></li><li><p><strong>Per-request variable costs are eliminated.</strong></p></li><li><p><strong>Network usage is only triggered for non-AI data or optional syncs.</strong></p></li></ul><p>This turns AI execution into a <strong>fixed cost</strong> (upfront hardware and model distribution) rather than a <strong>continuously variable cloud bill</strong>, which is preferable when the scale is enormous and the unit economics matter.</p><p><strong>This means Apple prepares for a future where AI runs at scale on every of their devices.</strong></p><div><hr></div><h2><strong>Software Infrastructure: Core ML and Developer Toolchains</strong></h2><p>Apple doesn&#8217;t just provide hardware; it provides optimized tooling. The <strong>Core ML framework</strong> is engineered to run models directly on the device with <strong>automated optimization and efficient hardware utilization</strong>, abstracting over CPU, GPU, and Neural Engine resources.</p><p>Core ML explicitly supports <strong>advanced AI models with compression and efficient execution</strong>, which makes it feasible to run larger transformer-style models locally without network dependency.</p><p>Additionally, Apple&#8217;s <strong>Foundation Models framework</strong> &#8212; released as part of its developer ecosystem &#8212; enables on-device large language models to power intelligent app features while still preserving privacy and offline capability.</p><div><hr></div><h2><strong>Offline and Reliability Benefits</strong></h2><p>On-device AI <strong>doesn&#8217;t break when connectivity does</strong>. Features stay responsive whether or not the user is on Wi-Fi, cellular, or completely offline. This robustness isn&#8217;t just a convenience; it&#8217;s a <strong>reliability constraint baked into product design</strong>, especially for critical features (e.g., local text understanding, image analysis) that are expected to work anywhere.</p><p>This offline capability expands where and how AI can be used - from high-latency rural networks to airplanes in flight - without degrading quality or predictability.</p><div><hr></div><h2><strong>Tradeoffs and Strategic Implications</strong></h2><p>On-device AI isn&#8217;t a completely free lunch, otherwise everyone would do it. Apple is effectively trading <strong>cloud flexibility and massive model scale</strong> for <strong>predictability, privacy, and consistent UX</strong>. Smaller parameter counts and optimized models constrain peak generative capabilities compared to cloud LLMs, but this trade-off aligns with Apple&#8217;s design philosophy: <strong>intelligence that&#8217;s ambient, private, and reliable on the device itself</strong>.</p><p>This architectural choice shapes the unit economics and operational risk profile dramatically: rather than paying per inference or relying on external infrastructure providers, Apple compounds <strong>hardware amortization and device-level execution</strong> &#8212; costs paid once and leveraged per user for years.</p><div><hr></div><h2><strong>Recap Of Why Apple Runs AI On-Device</strong></h2><p>Apple&#8217;s emphasis on on-device AI is not a superficial positioning statement &#8212; it&#8217;s a <strong>systems-level engineering choice</strong> that tightly aligns hardware, software, and privacy constraints. By localizing inference, Apple minimizes latency, reduces dependence on network bandwidth and cloud resources, preserves privacy, and shifts AI costs to predictable, device-specific investments rather than ongoing cloud bills. These factors fundamentally shape how AI powers everyday features on Apple platforms and define a very different set of economic constraints than cloud-centric alternatives.</p><div><hr></div><h2><strong>What This Means for Apple vs. Hyperscalers in Today&#8217;s AI Race</strong></h2><p>This means that Apple&#8217;s on-device AI approach puts it on a very different path than the hyperscalers, and the gap matters far more at the <strong>systems and cost level</strong>.</p><h3><strong>Two Very Different Ideas of Where Intelligence Lives</strong></h3><p>Hyperscalers like Google, Microsoft, and Amazon are building what are essentially centralized intelligence factories. The assumption is simple: intelligence lives in the data center, users and companies call it remotely, and the economics work because costs can be passed on through usage fees, subscriptions, or platform lock-in.</p><p>That model makes a lot of sense when AI is scarce, expensive, and clearly differentiated. It also fits naturally with how cloud businesses already operate, where variable cost and elastic pricing are accepted facts of life.</p><p><strong>Apple is betting on a different future. One where intelligence becomes frequent, embedded, and mostly invisible. In that world, sending every small AI interaction back to a data center starts to look wasteful, slow, and hard to justify economically.</strong></p><h3><strong>Centralized Compute vs. Intelligence at the Edge</strong></h3><p>From a technical perspective, this is really a story about where compute sits.</p><p>Hyperscalers concentrate intelligence where power, GPUs, and networking already exist. That gives them huge advantages in model scale, iteration speed, and enterprise-grade workloads. If you need massive models or fast experimentation, centralized infrastructure wins.</p><p>Apple flips the problem around. It pushes intelligence out to the edge, onto devices people already own, and spreads compute across hundreds of millions of phones and laptops. That gives Apple consistency in latency, privacy by default, and very predictable costs once the hardware is shipped.</p><p>Neither approach is universally better. They&#8217;re optimized for different bottlenecks. Hyperscalers optimize for capability and flexibility. Apple optimizes for scale without surprises.</p><h3><strong>When Cost Curves Start to Matter</strong></h3><p><strong>Early in the AI cycle</strong>, hyperscalers look clearly ahead. Bigger models, faster progress, and obvious monetization paths through cloud APIs and enterprise deals.</p><p>But as AI moves from &#8220;cool feature&#8221; to something people use constantly, the cost curves start to change. Hyperscalers pay for inference every single time a model runs. Apple pays once, upfront, when it designs the chip and sells the device. Basically, Apple makes their customers pay the AI bill upfront.</p><p>At low usage, that difference barely matters. At high, habitual usage, it becomes the whole game. This is where hyperscalers obsess over batching, utilization, and efficiency, while Apple benefits from inference that is effectively free at the margin.</p><h3><strong>Control Versus Ongoing Exposure</strong></h3><p>The other big difference is where risk sits.</p><p>Hyperscalers are continuously exposed to GPU supply cycles, energy prices, and the need to keep building and upgrading data centers. Those costs don&#8217;t stop once a model ships. They show up every quarter.</p><p>Apple takes on a different kind of risk. It concentrates it upfront in chip design, silicon investment, and longer iteration cycles. That risk is real, but it&#8217;s bounded. Once the hardware is out in the world, the economics are largely locked in.</p><p>From an investor&#8217;s point of view, that means Apple&#8217;s AI risk is front-loaded and familiar, while hyperscaler AI risk is ongoing and variable.</p><h3><strong>What &#8220;Winning&#8221; the AI Race Actually Means</strong></h3><p>If winning means the biggest models, the most parameters, and the flashiest benchmarks, Apple will often look behind.</p><p>If winning means stable margins at massive scale, AI quietly embedded into everyday behavior, and fewer operational surprises, Apple looks much more competitive than it gets credit for.</p><p>The key point is that Apple isn&#8217;t trying to beat hyperscalers at their own game. It&#8217;s opting out of that game and playing one where distribution, hardware amortization, and control over the stack matter more than raw model size.</p><div><hr></div><h2><strong>What This Means for You If You Hold Apple Stock</strong></h2><h3><strong>AI Isn&#8217;t Changing Apple&#8217;s Business Model. That&#8217;s the Point.</strong></h3><p>If you zoom out a bit, Apple&#8217;s on-device AI push doesn&#8217;t really change <em>what</em> the company is. It changes <em>how much risk</em> it&#8217;s willing to take on.</p><p>Apple isn&#8217;t trying to bolt a new AI business onto the side of its P&amp;L. It&#8217;s folding AI into the same (very profitable) machine it&#8217;s been running for years: sell premium hardware, spread silicon and R&amp;D costs over massive volume, and use software to make each new generation feel meaningfully better than the last one.</p><p>On-device inference fits that model cleanly. There&#8217;s no new per-request cost that scales with usage, and no sudden dependence on cloud pricing, GPUs, or energy markets. That&#8217;s a very Apple move.</p><h3><strong>This Is About Protecting Margins, Not Creating a New Growth Story</strong></h3><p>If you&#8217;re looking for an obvious &#8220;AI revenue line&#8221; in Apple&#8217;s numbers, you&#8217;re probably going to be disappointed. That&#8217;s not what this is.</p><p>Apple&#8217;s AI strategy is mostly about <em>not</em> letting margins get chipped away over time. If everyday intelligence had to run through the cloud, Apple would eventually be paying a variable cost every time users typed, searched, spoke, or edited photos. At Apple&#8217;s scale, that would add up fast.</p><p>By keeping inference on the device, Apple avoids that slow margin bleed. It&#8217;s defensive, but in a good way. For a company this large, keeping margins stable often matters more than finding a new flashy growth lever.</p><h3><strong>Apple Is Avoiding the Messy Part of the AI Stack</strong></h3><p>A lot of companies jumping into AI are signing up for things they don&#8217;t fully control: GPU supply cycles, cloud vendor pricing, energy costs, and constant infrastructure upgrades.</p><p>Apple mostly sidesteps that. It puts the risk where it&#8217;s already comfortable: chip design, long product cycles, and tight integration. That doesn&#8217;t mean zero risk, but it&#8217;s familiar risk. Investors usually underestimate how valuable that is.</p><p>This is Apple choosing the boring problems it already knows how to solve instead of the exciting ones that blow up later.</p><h3><strong>AI Helps Sell Hardware, Quietly</strong></h3><p>The real upside from AI for Apple isn&#8217;t some standalone AI service. It&#8217;s hardware upgrades.</p><p>When AI features are tied to newer chips, older devices don&#8217;t just feel slower, they feel outdated. That nudges replacement cycles without Apple having to say much about it. Over time, that shows up as better product mix, more high-end devices sold, and stronger ecosystem lock-in.</p><p>None of that makes headlines, but it compounds.</p><h3><strong>Control Keeps the Numbers Predictable</strong></h3><p>Because Apple controls the silicon, the OS, and distribution, it can decide how capable its AI features are and how expensive they&#8217;re allowed to be. That matters in a world where AI costs are still moving targets.</p><p>Predictability doesn&#8217;t sound exciting, but it&#8217;s exactly what long-term investors want. Fewer surprises means steadier cash flows. Steadier cash flows mean Apple can keep doing what it&#8217;s always done: buy back stock, pay dividends, and invest without drama.</p><p>In my opinion this aligns with a future where AI will ultimately be boring infrastructure business.</p><h3><strong>The Bet You&#8217;re Really Making as an Investor</strong></h3><p>Owning Apple in the AI era isn&#8217;t a bet on winning benchmarks or shipping the biggest models. It&#8217;s a bet that AI becomes everywhere, that costs matter more than raw capability once that happens, and that tight integration beats brute-force compute over time.</p><p>Apple isn&#8217;t trying to win the AI race loudly. It&#8217;s trying to not lose quietly. From a fundamentals point of view, that&#8217;s usually a better position than it looks at first glance.</p><div><hr></div><h2><strong>Closing Thoughts</strong></h2><p>Apple&#8217;s AI strategy makes a lot more sense once you stop treating AI as a product category and start treating it as infrastructure.</p><p>Hyperscalers are building centralized intelligence because that&#8217;s where their strengths already are. They&#8217;re optimized for scale, iteration speed, and selling compute as a service. Apple is optimizing for something else entirely: making intelligence cheap, predictable, and quietly embedded into devices people already use all day.</p><p>From the outside, that can look conservative, even underwhelming. From the inside, it&#8217;s a very deliberate choice about where costs should live, which risks are acceptable, and which ones are better avoided altogether. Apple is choosing to pay upfront, lock in economics early, and minimize ongoing exposure as AI usage grows.</p><p>For investors, the mistake would be to judge Apple by the same metrics used for hyperscalers like Google / Amazon / Microsoft. Apple doesn&#8217;t need to win benchmarks or dominate cloud APIs to &#8220;win&#8221; in AI. It needs AI to strengthen its existing flywheel without destabilizing margins or introducing new dependencies.</p><p>If AI becomes everywhere, the loudest players won&#8217;t necessarily be the best-positioned ones. The companies that quietly control costs, distribution, and failure modes tend to age better. Apple&#8217;s on-device approach is less about racing ahead and more about making sure it doesn&#8217;t get dragged into a cost structure it can&#8217;t control later.</p><p>That&#8217;s not an exciting ending. But in tech, and especially in investing, boring endings are often the ones that compound.</p>]]></content:encoded></item><item><title><![CDATA[Why Google Built TPUs and Why Investors Should Care]]></title><description><![CDATA[Inference costs, custom silicon, and what it signals for NVIDIA, Nebius, and AI investors]]></description><link>https://www.techfundamentals.blog/p/why-google-built-tpus-and-why-investors</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/why-google-built-tpus-and-why-investors</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Mon, 05 Jan 2026 16:33:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hJQO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hJQO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hJQO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hJQO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6116729,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183480558?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hJQO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hJQO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo Agency - stock.adobe.com</figcaption></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>There&#8217;s currently a lot of buzz around Google&#8217;s <strong>TPU</strong>, especially among investors and here on Substack. This - and given your interest in my last post about AI hardware - I think it&#8217;s a good idea to elaborate a bit on what TPUs are, their history and their influence on AI in today&#8217;s post.</p><div><hr></div><h2>A short history of TPUs: why they were built</h2><p>To understand why <strong>Tensor Processing Units (TPUs)</strong> matter for AI infrastructure investing, it helps to start with the problem they were designed to solve.</p><p>By the early 2010s, Google was already running neural networks in production across search ranking, ads, speech recognition, and image classification. These models worked, but the cost of running them at scale was becoming a structural issue. CPUs were too slow and inefficient. GPUs helped, but they were designed for graphics first and carried a lot of flexibility that translated into wasted power and higher operating costs when used continuously for inference.</p><p>This is the context in which TPUs emerged.</p><h3>Training vs inference: What you need to know to understand TPUs</h3><p>Before talking about TPUs in detail, it is worth slowing down on a distinction that shapes almost every AI cost curve: <strong>training versus inference</strong>.</p><p><strong>Training</strong> is the phase where a model learns. It is compute-intensive, episodic, and highly visible. Large clusters spin up, capital is deployed in bursts, and progress is easy to market. This is the phase that produces headlines, benchmark charts, and capex shock numbers. From the outside, it looks like the center of gravity in AI.</p><p>A concrete example is large language models like <strong>OpenAI</strong>&#8217;s GPT series. Training a new generation involves running massive datasets through thousands of GPUs or accelerators for weeks or months. This is where you see headline numbers around capex, energy usage, and hardware shortages. Once the run is finished, the cluster can be reallocated or powered down.</p><p>Another example is recommendation models at companies like <strong>Netflix</strong> or <strong>Meta</strong>. These models are retrained periodically as user behavior shifts. The training jobs are large but scheduled. They are part of an operational cycle, not a constant drain.</p><p><strong>Inference</strong> is what happens after the model exists. Every prompt, every search ranking, every recommendation, every generated token runs through inference. It is continuous, predictable, and quietly expensive. Inference does not happen once per quarter. It happens millions or billions of times per day, and it shows up directly in cost of revenue.</p><p>Every time someone types a query into a search engine at <strong>Google</strong>, multiple models run inference to rank results, filter spam, translate languages, and personalize outcomes. Each individual query is cheap. Billions of them are not.</p><p>When a user asks a question in a chatbot powered by models from <strong>OpenAI</strong>, inference runs for every token generated. The longer the answer, the higher the cost. This happens continuously, regardless of whether a new model is being trained that week.</p><p>In e-commerce, companies like <strong>Amazon</strong> run inference constantly to power recommendations, search ranking, fraud detection, and pricing systems. These models may be retrained occasionally, but inference never stops. It is part of the baseline cost of doing business.</p><p>At small scale, this distinction barely matters. Training dominates costs, inference is negligible, and general-purpose hardware feels flexible enough. At large scale, the balance flips. Training becomes a smaller share of lifetime cost, while inference becomes the permanent load the business has to carry.</p><p>This is the point where incentives change and why TPUs were invented:</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K24O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K24O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 424w, https://substackcdn.com/image/fetch/$s_!K24O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 848w, https://substackcdn.com/image/fetch/$s_!K24O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 1272w, https://substackcdn.com/image/fetch/$s_!K24O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K24O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp" width="1456" height="986" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:986,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:121788,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183480558?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K24O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 424w, https://substackcdn.com/image/fetch/$s_!K24O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 848w, https://substackcdn.com/image/fetch/$s_!K24O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 1272w, https://substackcdn.com/image/fetch/$s_!K24O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98cd0e12-39ee-4725-95b8-4efafaf6fc0e_2400x1626.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Google presenting TPU v4 in 2021</figcaption></figure></div><h3>What a TPU actually is</h3><p>A <strong>Tensor Processing Unit (TPU)</strong> is a custom-built <strong>application-specific integrated circuit (ASIC)</strong> designed specifically to accelerate machine learning workloads, especially dense tensor operations such as matrix multiplication. Unlike CPUs, which are general-purpose processors, or GPUs, which are broad parallel accelerators, TPUs are intentionally narrow. They remove features that are not useful for neural networks in order to maximize efficiency per watt and per dollar.</p><p>Google&#8217;s own documentation describes them this way:</p><blockquote><p>&#8220;Tensor Processing Units (TPUs) are Google&#8217;s custom-developed, application-specific integrated circuits (ASICs) used to accelerate machine learning workloads.&#8221;</p><p>Source: Google Cloud TPU documentation<br><a href="https://docs.cloud.google.com/tpu/docs/intro-to-tpu">https://docs.cloud.google.com/tpu/docs/intro-to-tpu</a></p></blockquote><p>TPUs are not consumer chips. They are deployed inside Google data centers and offered through Google Cloud, typically in large groups called TPU pods, because their economic advantage only appears at scale.</p><h3>Why Google built TPUs in the first place</h3><p>Google began developing TPUs internally around 2013, long before AI became a mainstream investment theme. The motivation was not performance leadership for its own sake, but cost containment. Neural network inference was growing faster than hardware efficiency, and relying on general-purpose chips meant accepting a cost curve that would only get worse with scale.</p><p>An early Google engineering reflection makes this explicit:</p><blockquote><p>&#8220;We thought we&#8217;d maybe build under 10,000 of them. We ended up building over 100,000 to support all kinds of great stuff including Ads, Search, speech projects, AlphaGo.&#8221;<br>Andy Swing, Principal Engineer, Google TPU team<br>Source: Google Cloud blog<br><a href="https://cloud.google.com/transform/ai-specialized-chips-tpu-history-gen-ai">https://cloud.google.com/transform/ai-specialized-chips-tpu-history-gen-ai</a></p></blockquote><p>The first publicly disclosed TPU, announced in 2016, was built primarily for <strong>inference</strong>, not training. That detail is easy to overlook, but it matters. Inference is the always-on cost center. Training is episodic. Google was optimizing the part of the AI lifecycle that shows up every second in operating expenses.</p><h3>What the efficiency gains looked like</h3><p>In Google&#8217;s original academic paper introducing TPUs, the authors quantified just how inefficient general-purpose hardware had become for neural network workloads:</p><blockquote><p>&#8220;The TPU delivered 15&#8211;30&#215; higher performance and 30&#8211;80&#215; higher performance-per-watt than contemporary CPUs and GPUs.&#8221;<br>Norman P. Jouppi et al., &#8220;In-Datacenter Performance Analysis of a Tensor Processing Unit&#8221;<br>Source: <a href="https://en.wikipedia.org/wiki/Tensor_Processing_Unit">https://en.wikipedia.org/wiki/Tensor_Processing_Unit</a></p></blockquote><p>Those numbers were not about peak benchmarks. They reflected real production inference workloads inside Google&#8217;s data centers.</p><p>As TPUs evolved, later generations expanded into training and large-scale model deployment, but the design center remained the same: predictable performance, high utilization, and lower marginal cost at scale. The most recent generations, such as Google&#8217;s inference-focused TPU platforms, continue to emphasize efficiency rather than raw flexibility.</p><p>Industry analysts like UncoverAlpha on Substack have summarized this positioning succinctly:</p><blockquote><p>&#8220;TPUs are not faster GPUs. They are cheaper answers to a narrower question.&#8221;<br>Source: Uncover Alpha</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:179815720,&quot;url&quot;:&quot;https://www.uncoveralpha.com/p/the-chip-made-for-the-ai-inference&quot;,&quot;publication_id&quot;:237593,&quot;publication_name&quot;:&quot;UncoverAlpha&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!YsyF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28f3c972-e191-4eec-857d-3ccfda10a107_227x227.png&quot;,&quot;title&quot;:&quot;The chip made for the AI inference era &#8211; the Google TPU&quot;,&quot;truncated_body_text&quot;:&quot;&quot;,&quot;date&quot;:&quot;2025-11-24T13:54:19.214Z&quot;,&quot;like_count&quot;:141,&quot;comment_count&quot;:3,&quot;bylines&quot;:[{&quot;id&quot;:22294341,&quot;name&quot;:&quot;UncoverAlpha&quot;,&quot;handle&quot;:&quot;uncoveralpha&quot;,&quot;previous_name&quot;:&quot;Rihard Jarc&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5dbed507-03bf-4b83-9728-ce17e75bf4b8_227x227.png&quot;,&quot;bio&quot;:&quot;Tech investor, CIO, and former CEO &amp; co-founder of an AI startup (sold to a public software company). &quot;,&quot;profile_set_up_at&quot;:&quot;2022-04-17T09:19:27.899Z&quot;,&quot;reader_installed_at&quot;:&quot;2024-08-20T05:17:29.956Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:238524,&quot;user_id&quot;:22294341,&quot;publication_id&quot;:237593,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:true,&quot;publication&quot;:{&quot;id&quot;:237593,&quot;name&quot;:&quot;UncoverAlpha&quot;,&quot;subdomain&quot;:&quot;uncoveralpha&quot;,&quot;custom_domain&quot;:&quot;www.uncoveralpha.com&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Deep dives/analyses of the technology companies and tech sub-industries. Mostly about AI, semiconductor, cloud, software, and ad tech sectors.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/28f3c972-e191-4eec-857d-3ccfda10a107_227x227.png&quot;,&quot;author_id&quot;:22294341,&quot;primary_user_id&quot;:22294341,&quot;theme_var_background_pop&quot;:&quot;#00C2FF&quot;,&quot;created_at&quot;:&quot;2020-12-12T11:35:53.306Z&quot;,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Rihard Jarc&quot;,&quot;founding_plan_name&quot;:&quot;Founding Member&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;enabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;newspaper&quot;,&quot;is_personal_mode&quot;:false}},{&quot;id&quot;:1917668,&quot;user_id&quot;:22294341,&quot;publication_id&quot;:1927580,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:1927580,&quot;name&quot;:&quot;Uncover Tech Insights&quot;,&quot;subdomain&quot;:&quot;uncovertechinsights&quot;,&quot;custom_domain&quot;:&quot;www.uncovertechinsights.com&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;A weekly newsletter with the most impactful news in tech. From financial news outlets to alternative sources such as interviews with former employees, customers, and industry experts.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fec638f8-c833-41ca-927a-32745f6cff59_465x465.png&quot;,&quot;author_id&quot;:22294341,&quot;primary_user_id&quot;:null,&quot;theme_var_background_pop&quot;:&quot;#25BD65&quot;,&quot;created_at&quot;:&quot;2023-09-05T11:33:12.700Z&quot;,&quot;email_from_name&quot;:&quot;Uncover Tech Insights&quot;,&quot;copyright&quot;:&quot;Rihard Jarc&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;newspaper&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:100,&quot;status&quot;:{&quot;bestsellerTier&quot;:100,&quot;subscriberTier&quot;:1,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:{&quot;type&quot;:&quot;bestseller&quot;,&quot;tier&quot;:100},&quot;paidPublicationIds&quot;:[6349492],&quot;subscriber&quot;:null}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.uncoveralpha.com/p/the-chip-made-for-the-ai-inference?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!YsyF!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28f3c972-e191-4eec-857d-3ccfda10a107_227x227.png" loading="lazy"><span class="embedded-post-publication-name">UncoverAlpha</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The chip made for the AI inference era &#8211; the Google TPU</div></div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">5 months ago &#183; 141 likes &#183; 3 comments &#183; UncoverAlpha</div></a></div></blockquote><h3>Why this history matters for investors</h3><p>This history clarifies intent: <strong>TPUs were not a reaction to hype, nor an attempt to compete in the merchant chip market</strong>. They were a response to an internal cost curve that threatened long-term scalability.</p><p>From an investing perspective, that distinction is critical. When a company commits to custom silicon, it is usually because renting general-purpose infrastructure no longer works economically. TPUs are a visible example of how AI, once it reaches scale, stops behaving like software and starts behaving like infrastructure.</p><div><hr></div><h2>AI looks like software, but behaves like infrastructure</h2><p>At small scale, AI behaves like software. You rent GPUs, ship features, and costs feel elastic. You can always optimize later. This phase is intoxicating, and it produces most of the narratives investors hear.</p><p>At scale, AI behaves like infrastructure. Power, memory bandwidth, utilization, and latency move from engineering trivia into margin drivers. Inference does not spike once a quarter like training runs. It runs every minute of every day, and it shows up directly in cost of revenue.</p><blockquote><p>TPUs exist because this transition happens faster than people expect. Google did not build them to win benchmarks. It built them because renting general-purpose accelerators at hyperscale turns variable costs into a permanent tax.</p></blockquote><div><hr></div><h2>Training gets the headlines, inference pays the bills</h2><p>Training dominates public discussion because it is visible. Giant clusters, massive capex numbers, new model releases. Inference is quieter and far more important economically.</p><p>Once a model is deployed, inference becomes the long tail of cost. Every user query, every autocomplete, every recommendation request burns compute. Even modest inefficiencies compound brutally when multiplied by billions of requests.</p><p>TPUs are increasingly tuned for this reality. They are optimized around throughput per watt, stable latency, and predictable scaling rather than raw flexibility. That tells you something about where Google sees the long-term cost center.</p><p>For investors, the implication is simple. Any AI business that cannot structurally reduce inference cost over time is running uphill. Pricing pressure eventually arrives, and when it does, gross margin is the only shock absorber.</p><div><hr></div><h2>Control is the real asset</h2><p>From the outside, TPUs look like a technical preference. From the inside, they are about control.</p><p>By building its own silicon, <strong>Alphabet</strong> controls its cost curve, its hardware roadmap, and its deployment cadence. It is not exposed to supplier pricing power in the same way a pure renter is. That does not mean TPUs are cheaper in every scenario. It means the company decides when and how cost improvements show up.</p><p>This distinction matters more than people realize. Companies that rent their entire AI stack inherit someone else&#8217;s incentives. When demand spikes, they pay more. When supply tightens, they wait. When pricing changes, margins move whether they want them to or not.</p><div><hr></div><h2>The GPU monoculture problem</h2><p>The current AI stack has a monoculture risk. A single architecture dominates, a single ecosystem defines tooling, and a single supplier captures most of the value. From a momentum and economic perspective, this looks unbeatable.</p><p>From a systems perspective, it is fragile.</p><p>When everyone depends on the same supply chain, pricing power concentrates. When demand overshoots supply, margins compress downstream. When innovation slows, the entire ecosystem feels it.</p><p>TPUs exist partly as an escape hatch. Not a replacement for GPUs everywhere, but a pressure valve. They remind investors that diversification at the infrastructure layer is not about performance bragging rights. It is about resilience.</p><p>This is where understanding the role of <strong>NVIDIA</strong> becomes more nuanced. Owning the toll road is incredibly lucrative, but toll roads are cyclical assets when alternatives mature.</p><div><hr></div><h2>How TPUs could break out beyond Google&#8217;s environment</h2><p>Historically, TPUs have lived almost entirely inside Google&#8217;s own walls. Unlike GPUs, which are sold broadly and show up in every major cloud and data center, TPUs have been tightly coupled to Google&#8217;s internal infrastructure and to Google Cloud. That constraint is not accidental. It reflects a deliberate strategy: TPUs were designed first to solve Google&#8217;s own cost problems, not to become a merchant chip.</p><p>What has changed is not the architecture, but the <strong>pressure on the rest of the ecosystem</strong>.</p><h3>From internal cost lever to external option</h3><p>Google has already crossed one important threshold by offering TPUs to third parties via Google Cloud. That move alone shifted TPUs from &#8220;internal optimization&#8221; to &#8220;commercial infrastructure,&#8221; even if still within a controlled environment.</p><p>As Google itself describes it:</p><blockquote><p>&#8220;Cloud TPUs are designed to accelerate machine learning workloads and make Google&#8217;s internal ML infrastructure available to customers.&#8221;<br>Source: Google Cloud TPU documentation<br><a href="https://docs.cloud.google.com/tpu/docs/intro-to-tpu">https://docs.cloud.google.com/tpu/docs/intro-to-tpu</a></p></blockquote><p>The more interesting question for investors is whether TPUs remain a Google-only cloud feature or evolve into something closer to a <strong>multi-tenant, multi-hyperscaler option</strong>.</p><h3>Signals from hyperscaler-scale customers</h3><p>In late 2024 and 2025, reports began to surface suggesting that very large AI users were actively exploring TPUs as an alternative to exclusive GPU dependence.</p><p>Reuters reported that Google was working with Meta to reduce Nvidia&#8217;s software advantage by improving TPU support for widely used frameworks:</p><blockquote><p>&#8220;Google is working with Meta Platforms to improve support for AI software on its in-house Tensor Processing Units, a move aimed at reducing Nvidia&#8217;s dominance in AI chips.&#8221;<br>Source: Reuters<br><a href="https://www.reuters.com/business/google-works-erode-nvidias-software-advantage-with-metas-help-2025-12-17/">https://www.reuters.com/business/google-works-erode-nvidias-software-advantage-with-metas-help-2025-12-17/</a></p></blockquote><p>That article matters less for the Meta angle and more for what it implies: Google is investing in <strong>ecosystem compatibility</strong>, not just hardware performance. That is a prerequisite for TPUs to travel beyond Google-native workloads.</p><p>Separate industry reporting has also noted that Meta has explored large-scale TPU usage through Google Cloud as part of its broader effort to diversify away from single-vendor GPU exposure.</p><blockquote><p>&#8220;Meta is reportedly considering Google&#8217;s TPU chips as part of its effort to reduce reliance on Nvidia for AI infrastructure.&#8221;<br>Source: AI Certs industry report<br><a href="https://www.aicerts.ai/news/meta-eyes-google-tpu-chips-in-high-stakes-ai-partnership/">https://www.aicerts.ai/news/meta-eyes-google-tpu-chips-in-high-stakes-ai-partnership/</a></p></blockquote><p>None of this means TPUs are about to be sold like off-the-shelf GPUs. But it does suggest that, at hyperscaler scale, <strong>the desire for architectural leverage is growing</strong>.</p><h3>Software is the real gatekeeper</h3><p>One of the biggest historical blockers for TPUs outside Google was tooling. NVIDIA&#8217;s CUDA became the de facto standard for AI development, and most production pipelines were built around it. TPUs, by contrast, were initially optimized for TensorFlow and Google-internal stacks.</p><p>Google has been explicit about trying to lower this barrier:</p><blockquote><p>&#8220;We&#8217;re investing heavily in making TPUs easier to use with popular frameworks so customers don&#8217;t have to rewrite their models.&#8221;<br>Source: Google Cloud blog<br><a href="https://cloud.google.com/blog/products/ai-machine-learning/an-in-depth-look-at-googles-first-tensor-processing-unit-tpu">https://cloud.google.com/blog/products/ai-machine-learning/an-in-depth-look-at-googles-first-tensor-processing-unit-tpu</a></p></blockquote><p>From an investor perspective, this is the most important signal. If TPUs remain software-isolated, they stay niche. If they become genuinely first-class citizens in frameworks like PyTorch, they become an <strong>economic alternative</strong>, not just a technical one.</p><h3>Why a TPU breakout would matter</h3><p>If TPUs ever move beyond Google Cloud into colocated environments, joint ventures, or tightly integrated partnerships with other hyperscalers, the impact would not be about performance leadership. It would be about <strong>negotiating power</strong>.</p><p>For the AI infrastructure market, that would mean:</p><ul><li><p>More leverage against single-vendor pricing like NVIDIA</p></li><li><p>More pressure on GPU margins at the largest scale</p></li><li><p>More incentive for diversified compute strategies</p></li></ul><p>This does not &#8220;kill&#8221; <strong>NVIDIA</strong>. But it does introduce a credible reminder that its largest customers are not passive. They are actively looking for ways to internalize costs and reduce dependency.</p><p>TPUs escaping Google&#8217;s silo would not be a sudden disruption. It would be a slow, structural shift, visible first in contracts and capex decisions rather than benchmarks. For investors, those are exactly the signals that tend to matter most over the long run.</p><div><hr></div><h2>How TPUs could affect NVIDIA and Nebius</h2><p>Understanding TPUs is useful not because they replace GPUs everywhere, but because they change the <strong>power balance</strong> in the AI infrastructure stack. That shift has very different implications for companies like <strong>NVIDIA</strong> and <strong>Nebius</strong>.</p><h3>NVIDIA: pricing power with an expiration date, not a cliff</h3><p>NVIDIA&#8217;s position today is extraordinary. It controls the dominant hardware platform, the dominant software ecosystem, and most of the marginal dollars flowing into AI compute. From an investor perspective, that looks like a textbook toll booth.</p><p>TPUs do not invalidate this. In fact, in the near term they arguably reinforce it. The existence of TPUs signals just how expensive and strategically important AI compute has become. That validates NVIDIA&#8217;s margins rather than undermining them.</p><p>Every large-scale TPU deployment is a reminder that NVIDIA&#8217;s biggest customers are also its most motivated future competitors. Hyperscalers do not build custom silicon because GPUs are bad products. They do it because paying a perpetual margin to an external supplier becomes painful once inference volume stabilizes and grows predictably.</p><p>For NVIDIA, this creates a subtle but important dynamic: Revenue growth can remain strong while pricing power slowly caps out and unit volumes can rise even as hyperscalers work to reduce long-term dependence.</p><p>This means the the risk is not sudden displacement, but <strong>margin compression</strong>, starting with the largest, most sophisticated buyers.</p><p>From the outside, NVIDIA still looks untouchable. From the inside, TPUs represent the point where customers stop asking &#8220;how fast can we scale?&#8221; and start asking &#8220;how much control do we really have?&#8221;</p><p>That does not break the NVIDIA story, but it reframes it as a cyclical infrastructure business with unusually strong near-term leverage rather than an infinitely compounding software-like asset.</p><h3>Nebius: caught between simplicity and dependence</h3><p>Nebius sits on the opposite side of the control equation.</p><p>As a cloud and AI infrastructure provider without proprietary silicon, Nebius benefits from NVIDIA&#8217;s ecosystem and performance leadership, but it also inherits NVIDIA&#8217;s cost structure. Every improvement in GPU pricing or efficiency helps Nebius. Every supply constraint or pricing shift flows directly into its margins.</p><p>TPUs matter here not because Nebius will deploy them tomorrow, but because they highlight the <strong>structural disadvantage</strong> of being a compute renter in a world where AI inference becomes the dominant workload instead of training new models.</p><p>If hyperscalers can run inference more cheaply on their own silicon, they gain room to compete more aggressively on price for large customers and absorb cost volatility internally.</p><p>Nebius, by contrast, must either pass costs through or accept margin pressure. At small scale, this is manageable. At large scale, it becomes a strategic ceiling.</p><p>From an investor perspective, this does not mean Nebius cannot succeed. It means its upside is more tightly coupled to NVIDIA&#8217;s roadmap and pricing discipline. TPUs expose that dependency clearly. Nebius is building on top of someone else&#8217;s product.</p><div><hr></div><h3>The broader takeaway for investors of NVIDIA &amp; Nebius</h3><p>TPUs do not &#8220;kill&#8221; GPU companies, and they do not automatically doom GPU-based clouds. What they do is draw a clean line between <strong>infrastructure owners</strong> and <strong>infrastructure renters</strong>.</p><ul><li><p>NVIDIA sits upstream, monetizing everyone&#8217;s urgency to deploy AI.</p></li><li><p>Hyperscalers with TPUs are trying to internalize that urgency into controllable costs.</p></li><li><p>Providers like Nebius live in the middle, exposed to both sides.</p></li></ul><p>For investors, the mistake is treating these businesses as if they compound in the same way. TPUs make it clear that AI infrastructure rewards control first, efficiency second, and flexibility last.</p><p>Once AI becomes boring and inference is becoming a bigger topic, the companies that own their cost curves tend to age better than the ones that rent them.</p><div><hr></div><h2>Closing thoughts</h2><p>Once you see TPUs as an economic tool rather than a chip, several things shift.</p><p>You stop evaluating AI companies solely on model quality and start asking who controls their cost base. You become skeptical of businesses whose gross margins depend on someone else&#8217;s roadmap. You treat infrastructure ownership as a form of pricing power, even when it does not show up immediately.</p><p>Most importantly, you stop assuming that AI compounding is automatic. Technology overwrites itself constantly. Cost structures are what survive.</p><p>TPUs matter because they sit at that intersection. They are not the future of all AI compute. They are proof that, at scale, fundamentals reassert themselves.</p>]]></content:encoded></item><item><title><![CDATA[Intro To AI Hardware For Investors: GPUs Won. That’s Why Everyone Is Hedging Them]]></title><description><![CDATA[GPUs, CPUs, RAM, Silicon, ARM explained. Plus: Margins, and the quiet shift in AI hardware]]></description><link>https://www.techfundamentals.blog/p/intro-to-ai-hardware-for-investors</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/intro-to-ai-hardware-for-investors</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Sun, 04 Jan 2026 16:06:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!McA6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!McA6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!McA6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 424w, https://substackcdn.com/image/fetch/$s_!McA6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 848w, https://substackcdn.com/image/fetch/$s_!McA6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!McA6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!McA6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg" width="1456" height="1090" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1090,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:9135354,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183448524?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!McA6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 424w, https://substackcdn.com/image/fetch/$s_!McA6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 848w, https://substackcdn.com/image/fetch/$s_!McA6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!McA6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F173760ca-b260-4cd7-8b21-aa3f382fb345_5264x3942.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>An investor&#8217;s lens on AI hardware</h2><p>Most AI investment narratives start in the wrong place. They focus on models, benchmarks, and headline-grabbing chip launches, as if the main question is who has the fastest technology. That&#8217;s rarely where long-term outcomes are decided.</p><p>For investors, AI hardware matters for a simpler reason: it&#8217;s becoming one of the largest and least flexible cost lines in modern tech businesses. Compute, memory, power, and supply constraints don&#8217;t scale like software, and once AI usage moves from experimentation to production, those constraints start showing up directly in margins, pricing power, and strategic optionality.</p><p>This is where a lot of confusion creeps in. Terms like GPUs, ARM, and &#8220;custom silicon&#8221; get thrown around as if they&#8217;re competing bets on the same axis, when in reality they sit at very different layers of the system and serve very different economic purposes. Some choices buy flexibility, others buy efficiency, and a few quietly buy leverage.</p><p>This article isn&#8217;t about predicting which chip wins or which company ships the fastest accelerator. It&#8217;s about understanding how AI hardware actually behaves once it&#8217;s embedded into real products at scale, who controls the economics when that happens, and why many of the most important moves in this space are defensive rather than disruptive.</p><div><hr></div><h2>What AI hardware actually is</h2><p>When people talk about AI hardware, they are almost always talking about GPUs, even if they don&#8217;t explicitly say it, and that&#8217;s already a bit misleading because GPUs were never built for AI in the first place. They were built for graphics, games, video rendering, and other visual workloads, but it turns out that the kind of math you need to draw millions of pixels on a screen looks very similar to the math modern neural networks rely on.</p><p>At a high level, AI models are doing massive amounts of repetitive math, mostly matrix multiplications, nothing conceptually complicated, just a huge volume of it. That&#8217;s where GPUs happen to shine, because instead of being optimized to do a few complex tasks very quickly, they are built to do the same simple operation thousands of times in parallel.</p><p>This is the key difference compared to CPUs. A CPU is great at decision-making, branching, and juggling many different tasks at once, which is why it&#8217;s perfect for running an operating system. A GPU is more like an assembly line, where every worker repeats the same motion over and over again, which sounds limiting until you realize that this is exactly what AI workloads want.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zsk3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zsk3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zsk3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4460522,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183448524?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zsk3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Zsk3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7067bc0-0fad-4c03-9dfa-5688135497fe_5376x3584.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>How GPUs quietly became the backbone of AI</h2><p>No one sat down and designed a &#8220;perfect AI chip&#8221; at the beginning. Researchers simply used the hardware that was already available and eventually realized that GPUs could run neural network workloads much faster than CPUs, which triggered a chain reaction that still shapes the industry today.</p><p>Once AI frameworks started targeting GPUs, model architectures adapted to GPU behavior, tooling improved, and developers learned to think in GPU-friendly ways. Over time, the GPU stopped being an optimization choice and became an assumption, something everything else was built around rather than something you consciously evaluated.</p><p>That&#8217;s why GPUs didn&#8217;t win because they were theoretically ideal. They won because they were good enough early on and then accumulated an ecosystem so large that replacing them became a tooling, cultural, and economic problem, not just a technical one.</p><div><hr></div><h2>About training and inference</h2><p>From the outside, all AI compute looks roughly the same, but once you&#8217;re paying for it the difference between <strong>training</strong> and <strong>inference</strong> becomes very obvious. Training is expensive and painful, but it usually happens in bursts, while inference runs continuously in the background and scales directly with how successful your product becomes.</p><p>Every chatbot reply, every generated image, every recommendation is <strong>inference</strong>, and once users expect those features to work instantly and all the time, you don&#8217;t get to pause or batch that workload later. This is where hardware flexibility starts to matter more than peak efficiency.</p><p>GPUs ended up in a sweet spot because they can handle both training and inference reasonably well, even as models and workloads keep changing, and that adaptability turns out to be more valuable in practice than squeezing out a bit more theoretical performance.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4nOm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4nOm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 424w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 848w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4nOm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7820401,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183448524?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4nOm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 424w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 848w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!4nOm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F124f7693-eb39-4faf-857c-840a586cff69_5568x3712.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Memory: the part that quietly dominates everything</h2><p>If GPUs are the muscle of AI systems, memory is the thing that actually keeps them alive, and this is the part many non-hardware people underestimate. AI models are physically large, not just conceptually complex, because all of those parameters have to live somewhere and be moved around constantly while the model runs.</p><p>If that movement is slow, the GPU ends up waiting, and a GPU that&#8217;s waiting is still drawing power, still occupying rack space, and still showing up on the bill. In many real-world systems, the bottleneck isn&#8217;t how fast the GPU can calculate but how fast it can get data.</p><p>That&#8217;s why memory decisions matter just as much as compute decisions, even though they usually get far less attention.</p><div><hr></div><h2>System RAM and GPU memory</h2><p>There&#8217;s system RAM, which sits next to the CPU and is relatively cheap and flexible, and then there&#8217;s GPU memory, often called VRAM or high-bandwidth memory, which is much faster, much closer to the GPU cores, and dramatically more expensive per gigabyte.</p><p>AI workloads strongly prefer to keep the entire model inside GPU memory. The moment parts of it spill into system RAM, performance drops sharply, latency increases, and power usage creeps up in ways that don&#8217;t show up in benchmarks but absolutely show up in cloud invoices.</p><p>This is why AI GPUs are expensive. You&#8217;re not just paying for compute, you&#8217;re paying for fast, scarce memory that&#8217;s tightly coupled to that compute, and once you hit memory limits you can&#8217;t fix the problem cheaply without changing the whole performance profile of the system.</p><div><hr></div><h2>Memory bandwidth: the quiet limiter</h2><p>In practice, many AI workloads are memory-bound rather than compute-bound, meaning the GPU could theoretically do more work but spends a lot of time waiting for data to arrive. That unused capacity is still paid for in electricity and capital, which is why it becomes an economic problem, not just a technical one.</p><p>This also explains why newer generations of GPUs focus so heavily on memory bandwidth and interconnects. Faster memory leads to higher real utilization, which lowers cost per request and often matters more than raw compute improvements.</p><p>From the outside, this looks like a minor spec detail. From the inside, it&#8217;s one of the main drivers of margins.</p><div><hr></div><h2>Why memory makes AI infrastructure sticky</h2><p>Once an AI system is built around a certain memory setup, it tends to stick, because models get tuned to specific sizes, engineers optimize around bandwidth assumptions, and lots of small, invisible decisions accumulate over time.</p><p>Switching hardware later is rarely a clean swap. It usually means retuning models, rewriting performance-critical code, and accepting new edge cases and failure modes, which is why hardware choices in AI tend to last longer than software choices and why GPUs, with their tightly integrated compute and memory, became so deeply entrenched.</p><p>This stickiness is also what makes any serious attempt to replace GPUs much harder than it looks at first glance.</p><p>With GPUs and Memory covered, I want to talk about CPUs next, which in the AI space is often erroneously referred to &#8220;silicon&#8221;.</p><div><hr></div><h2>What &#8220;silicon&#8221; actually means in AI conversations</h2><p>When people say &#8220;silicon&#8221; in AI discussions, they&#8217;re usually not talking about a specific kind of chip at all. It&#8217;s shorthand for hardware a company designs itself instead of buying off the shelf. Almost every modern processor is made of silicon, whether it&#8217;s a CPU, a GPU, or an AI accelerator, so the material isn&#8217;t the point. What matters is who controls the design and what assumptions are baked into it.</p><p>When you hear phrases like &#8220;Apple&#8217;s silicon&#8221; or &#8220;custom silicon for AI,&#8221; what&#8217;s really being described is a decision to take ownership over part of the hardware stack rather than accept someone else&#8217;s roadmap, pricing, and constraints. Therefore silicon isn&#8217;t a category of hardware. </p><p>I often hear ARM as equally misunderstood term in discussions, so let&#8217;s tackle that next:</p><div><hr></div><h2>ARM is not a chip, it&#8217;s the language CPUs speak</h2><p>ARM is often described as a chip company, but that&#8217;s slightly misleading. ARM mostly doesn&#8217;t build chips. It defines an instruction set, essentially the language a CPU speaks. That language specifies how software talks to hardware, how instructions are structured, and how memory is accessed. ARM became popular because it&#8217;s efficient, modular, and easy to license, which is why it dominates phones and is now spreading into laptops and servers.</p><p>When companies design ARM-based chips, they aren&#8217;t inventing CPUs from scratch. They&#8217;re starting from a shared language and building their own cores, memory systems, and accelerators around it. ARM answers one specific question: how the CPU part of a chip behaves. That&#8217;s important, but it&#8217;s not where most AI computation happens.</p><div><hr></div><h2>CPUs still matter, just not for the heavy AI lifting</h2><p>CPUs, whether ARM-based or not, are still essential. They run the operating system, handle networking, schedule workloads, and keep the system stable. What they don&#8217;t do well is the massive parallel math modern AI models rely on, which is why GPUs took over training and large-scale inference.</p><p>When people talk about &#8220;ARM for AI,&#8221; they usually mean ARM CPUs coordinating AI workloads, not replacing GPUs. The CPU is the conductor, not the orchestra.</p><div><hr></div><h2>Custom AI chips are purpose-built accelerators</h2><p>Custom AI chips live in a different category altogether. These are purpose-built accelerators designed to run neural networks efficiently under specific assumptions. They support a narrower set of operations, trade flexibility for efficiency, and are usually aimed at inference workloads where models are stable and traffic is predictable.</p><p>They don&#8217;t replace CPUs or GPUs. They sit alongside them and handle the parts of the system where efficiency matters more than adaptability.</p><div><hr></div><h2>How these pieces fit together in real systems</h2><p>In real AI systems, these components stack rather than compete. A <strong>CPU</strong> handles orchestration and system logic, a <strong>GPU</strong> handles flexible high-performance compute, and a custom accelerator handles narrow, high-volume inference paths.</p><blockquote><p>Once you look at the system this way, the current hardware strategies become much easier to understand. Companies aren&#8217;t choosing between ARM, GPUs, or custom chips. They&#8217;re deciding which layers of the stack they want to control directly and which ones they&#8217;re willing to rent.</p></blockquote><div><hr></div><h2>Why all of this quietly pushes companies toward custom silicon</h2><p>Once you really sit with how tightly compute and memory are tied together in modern AI systems, a lot of current hardware decisions start to look less mysterious. GPUs didn&#8217;t just win because they were fast, they won because they bundled compute, memory, software tooling, and developer habits into one package that&#8217;s extremely hard to unwind without breaking things in subtle, expensive ways.</p><p>That stickiness is great when you&#8217;re building on top of it and terrible once you&#8217;re dependent on it.</p><p>At small scale, none of this feels dramatic. You spin up instances, pay the bill, and focus on shipping product. But as usage grows, inference runs constantly, memory inefficiencies start showing up in real money, and suddenly hardware stops being an abstract infrastructure layer and starts shaping what you can and can&#8217;t afford to do.</p><p>At that point, GPUs stop being &#8220;just the best option&#8221; and start becoming a structural dependency. Pricing is set somewhere else. Supply constraints are out of your control. Hardware roadmaps begin to leak into product roadmaps in ways software teams usually don&#8217;t like to admit.</p><p>This is where the conversation shifts, not toward replacing GPUs outright, because that&#8217;s far harder than it sounds, but toward asking a quieter question: <strong>how much of this dependency are we actually comfortable with over the long term?</strong></p><p>That&#8217;s where custom silicon shows up.</p><p>The mistake is to think custom chips are about building something fundamentally better than GPUs. Most of the time they aren&#8217;t. They&#8217;re about creating an escape hatch, even if it&#8217;s narrow, limited to certain workloads, or only economically sensible at very large scale.</p><p>Once a company can say that not every workload needs to run on GPUs, the balance of power changes a bit. Pricing discussions look different. Shortages hurt less; if you&#8217;re into gaming or AI infrastructure you surely know what I&#8217;m talking about. Roadmaps become something you negotiate around instead of something you simply accept.</p><p>In that sense, custom silicon isn&#8217;t a revolution or a bold bet on a new future. It&#8217;s what happens when GPUs and memory become important enough that relying on a single external supplier starts to feel risky.</p><div><hr></div><h2>This isn&#8217;t theoretical. Big tech is already acting this way.</h2><p>Once you look at AI hardware through the lens of dependency and leverage, the current moves by big tech stop looking experimental and start looking inevitable. None of the major players are behaving as if the GPU era is about to end, but none of them are behaving as if full dependency on a single hardware supplier is acceptable either.</p><p>They&#8217;re all converging on the same strategy from different angles: keep GPUs central, but reduce how much power any one supplier has over the system as a whole.</p><div><hr></div><h2>Apple: control the stack or don&#8217;t play the game</h2><p><strong>Apple</strong> is often held up as the example of how custom silicon can be transformative, and that&#8217;s true, but it&#8217;s also why Apple is such a bad comparison for almost everyone else.</p><p>Apple didn&#8217;t build its own chips because GPUs were too expensive or because it wanted better benchmarks. In my opinion, it built them because Apple hates being dependent on anyone else&#8217;s roadmap. Performance per watt, battery life, thermals, form factor, software APIs, developer tooling, all of that becomes much easier to reason about once the silicon is yours.</p><p>What&#8217;s interesting is how directly this thinking maps to AI, even though Apple&#8217;s AI story looks very different from cloud providers. Running more inference on-device only makes sense if you control compute and memory tightly, because latency, power usage, and privacy constraints all collapse into hardware decisions.</p><p>Apple&#8217;s takeaway isn&#8217;t &#8220;custom silicon beats GPUs.&#8221;<br>It&#8217;s &#8220;control compounds when you own the whole system.&#8221;</p><div><hr></div><h2>Microsoft: hedge the dependency, don&#8217;t fight it</h2><p><strong>Microsoft</strong> sits in a very different position. It doesn&#8217;t control consumer hardware at scale, but it does run one of the largest cloud platforms in the world, and that makes GPU dependency a balance sheet problem rather than a technical curiosity.</p><p>Azure&#8217;s AI push is deeply GPU-centric, and Microsoft is one of the biggest customers <strong>NVIDIA</strong> will ever have. At the same time, Microsoft is very clearly investing in its own silicon, not because it believes it can replace GPUs overnight, but because relying exclusively on someone else&#8217;s pricing, supply, and roadmap is a long-term risk.</p><p>Exactly, this is the hedge in its purest form.</p><p>Some workloads stay on GPUs because flexibility matters. Others slowly move to internal accelerators where performance is predictable and margins matter more than raw capability. The goal isn&#8217;t technical dominance, but it&#8217;s optionality.</p><p>Microsoft doesn&#8217;t need custom silicon to win. It needs it so that GPUs don&#8217;t get to set all the rules.</p><div><hr></div><h2>NVIDIA: leaning into dominance without pretending it&#8217;s forever</h2><p>What makes <strong>NVIDIA</strong> interesting is that it&#8217;s not behaving like a company that thinks this dependency will never be challenged. NVIDIA knows better than anyone how fragile dominance can be in hardware.</p><p>Instead of just selling chips, NVIDIA keeps expanding the surface area of what &#8220;a GPU&#8221; even means. Software stacks, networking, interconnects, developer tooling, and increasingly entire data center reference architectures. The goal is to make the GPU not just a component but the default way AI systems are built.</p><p>At the same time, NVIDIA is pushing aggressively into higher memory bandwidth, tighter integration, and system-level solutions, which is exactly what you&#8217;d expect if you understand that memory and utilization, not just compute, are where the real leverage lives.</p><p>NVIDIA isn&#8217;t ignoring custom silicon. It&#8217;s pricing, bundling, and integrating in a way that makes leaving as expensive and risky as possible.</p><div><hr></div><h2>Different strategies, same direction</h2><p>Apple, Microsoft, and NVIDIA are doing very different things on the surface, but underneath, the incentives line up neatly.</p><ul><li><p><strong>Apple</strong> wants total control, because it can afford to build everything itself.</p></li><li><p><strong>Microsoft</strong> wants leverage and margin protection at massive scale.</p></li><li><p><strong>NVIDIA</strong> wants to remain unavoidable, even as customers hedge against it.</p></li></ul><p>None of these companies are betting on a clean break from GPUs. None of them are acting like custom silicon is a silver bullet. They&#8217;re all behaving as if AI hardware is now core infrastructure, closer to energy or networking than to normal software, and infrastructure rewards control, redundancy, and leverage.</p><p>Custom silicon isn&#8217;t showing up because GPUs failed, but it&#8217;s showing up because GPUs succeeded so completely that depending on them alone became risky.</p><p>From here, the remaining question isn&#8217;t whether custom silicon &#8220;wins,&#8221; but how much dependency reduction is enough before the complexity stops being worth it.</p><div><hr></div><h2>Closing thoughts</h2><p>If you strip away the hype, the benchmarks, and the chip launch headlines, what&#8217;s happening in AI hardware is actually pretty familiar for the development of new systems in tech. Systems that start out flexible and cheap eventually become critical infrastructure, and once that happens, the conversation shifts from performance to control, cost, and risk.</p><p>GPUs didn&#8217;t become central to AI because they were perfect. They became central because they were available, flexible, and good enough early on, and then everything else quietly aligned around them. Memory, tooling, developer habits, pricing models, and product expectations all piled on top until walking away stopped being a technical decision and started being an economic one.</p><p>Custom silicon is a reaction to that. Not as a rebellion against GPUs, and not because everyone suddenly discovered a better way to do math, but because depending entirely on someone else&#8217;s hardware, memory roadmap, and pricing power feels fine at small scale and dangerous at large scale. The bigger the AI footprint gets, the more that risk matters.</p><p>That&#8217;s why the most interesting signal right now isn&#8217;t who claims to have the fastest chip, but who is quietly building options. </p><p>From an investing perspective, this is where a lot of narratives break. Custom silicon doesn&#8217;t automatically mean disruption, and GPU dominance doesn&#8217;t automatically mean permanent pricing power. What matters is who controls enough of the stack to survive margin pressure (see my other articles), supply shocks, and the slow grind of inference costs over time.</p><p>AI hardware is starting to look less like software and more like utilities. Boring, expensive, politically constrained, and incredibly hard to unwind once you&#8217;re in too deep (except if you read and subscribe to my Substack of course).</p><p>I hope this article could give you a good overview of hardware involved in AI and you&#8217;ll understand hardware news better now. Don&#8217;t forget to like and share this article. If you have any questions or comments, don&#8217;t hesitate to post them.</p><p><strong>Update: </strong>If you want to learn more about a specific custom silicon from Google called <strong>TPU</strong>, I got you covered:</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;b392b2c9-5da8-48a5-b855-29c5d25d9b88&quot;,&quot;caption&quot;:&quot;Disclaimer: This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Why Google Built TPUs and Why Investors Should Care&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:428975524,&quot;name&quot;:&quot;Daniel Kolb&quot;,&quot;bio&quot;:&quot;Software engineer investing in tech stocks and crypto with a focus on fundamentals, system design, and real-world usage.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/129784c0-a568-4531-b741-47c6984b24da_638x638.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-01-05T16:33:48.988Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/$s_!hJQO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81ec762d-b6ab-4acd-93b0-c2eb24aa5f23_9216x5184.jpeg&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://techfundamentals.substack.com/p/why-google-built-tpus-and-why-investors&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:183480558,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:3,&quot;comment_count&quot;:1,&quot;publication_id&quot;:7381602,&quot;publication_name&quot;:&quot;Tech Fundamentals&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!xWDB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0870fc2b-6ca8-452b-aea6-8572d56e44de_670x670.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div>]]></content:encoded></item><item><title><![CDATA[AI in 2026: When the Demos Stop Carrying the Story]]></title><description><![CDATA[From wow moments to economic reality]]></description><link>https://www.techfundamentals.blog/p/ai-in-2026-when-the-demos-stop-carrying</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/ai-in-2026-when-the-demos-stop-carrying</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Thu, 01 Jan 2026 07:00:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cW8L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cW8L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cW8L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cW8L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg" width="1456" height="660" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:660,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1439769,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183107616?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cW8L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cW8L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bfc30d7-1c3a-40a7-8ef1-c6b4cf9308dc_3046x1380.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p><br><strong>Happy new year everyone!</strong> <br><br>For the last few years, AI has mostly been judged by how impressive it looks in public. First it was <strong>OpenAI</strong> and <strong>ChatGPT</strong> suddenly sounding smarter than anyone expected, then <strong>Sora</strong> arrived and made it feel like video production itself was about to be rewritten. Capital (and especially the NVIDIA stock) reacted immediately. Big tech spent hundreds of billions on GPUs, data centers, long-term energy contracts, and custom chips, all on the assumption that intelligence would naturally translate into revenue at scale.</p><p>From where I sit, that assumption has always felt fragile. As a software engineer working with these systems daily, and as someone who has built startups and runs an agency that ships production software for paying clients, I&#8217;ve learned that technical capability and economic leverage rarely move in sync. AI looks magical in isolation, but once it lands inside real environments with SLAs, compliance constraints, edge cases, and accountability, the story becomes far less clean.</p><p>So far, the numbers reflect that. AI revenue remains small relative to the balance sheets now supporting it. Even compared to companies like <strong>Apple</strong> or <strong>Alphabet</strong>, AI is strategically important but still not a standalone cash engine. And almost none of the infrastructure spend accrues to the average AI startup. Most live one layer above the models, packaging the same intelligence in slightly different forms and hoping focus or UX creates leverage.</p><h2>Adoption looks big until it meets real workflows</h2><p>This is why 2026 feels like a turning point. Not because the models stop improving (they won&#8217;t), but because attention is shifting from demos to deployment. From the outside, adoption looks everywhere. Engineers rely on AI daily, marketers treat it as default tooling, students grow up with it, and many people quietly get more done because of it.</p><p>Inside organizations, progress is slower and messier. Formal adoption is uneven, pilots are common, and many fade out without drama, sometimes because it might reflect bad on employees to speak loud about AI failures. I see this constantly in agency work. Teams build a proof of concept, show early promise (especially by euphorized team members), then hit harder questions than expected. Who signs off on outputs that are sometimes wrong. How failures are handled when automation breaks something subtle. How probabilistic systems fit into organizations built around certainty and blame.</p><p>AI often gets close enough to be useful without ever fully owning the outcome. Responsibility does not transfer along with the output. Legal risk, on-call duty, customer fallout, and reputational damage still sit with humans. That gap between assistance and ownership is where most AI ROI disappears, and it is also where many AI startups quietly stall.</p><p>Even focused vertical tools like <strong>Harvey AI (</strong>AI for lawyers) or <strong>Sierra</strong> (AI for customer service) run into this reality. Their success depends less on model quality and more on whether organizations are willing to change workflows, incentives, and responsibility boundaries. That kind of change is slow, political, and consistently underestimated.</p><h2>AI wrappers: Startups that don&#8217;t own their tech &amp; what it means for Big Tech</h2><p>There is a seductive belief that AI finally restores the compounding effect to software. Smarter tools, faster iteration and more leverage per person. From my experience and in practice, AI dramatically lowers the cost of building while simultaneously shortening the half-life of advantage for anyone who does not control their economics.</p><p>This is why <strong>AI wrapper economics</strong> should feel familiar to anyone who lived through previous platform shifts. It resembles businesses built entirely on APIs whose fate depended on someone else&#8217;s roadmap. Many of those AI wrapper companies were well executed. Some became case studies.</p><p>All of this matters because AI is no longer a side theme in markets. It is a core assumption. By late 2025, AI-related companies made up a massive share of the <strong>S&amp;P 500</strong>, and current valuations assume that adoption accelerates and margins expand.</p><p>A large share of what gets labeled an AI company today is effectively a wrapper. The core intelligence comes from <strong>OpenAI</strong>, <strong>Anthropic</strong>, or a hyperscaler. The startup builds a thin product layer on top. From the outside, it looks like SaaS. Under the hood, it behaves like a variable-cost service.</p><p>When these systems run at scale, the tension becomes obvious. Usage does not follow neat pricing assumptions. Customers expect more volume for less money. Model behavior changes without warning and margins compress in ways pitch decks never modeled. In 2026, more investors will realize that many AI startups are not failing because adoption is slow, but because the economics were fragile from the start.</p><p>This puts a lot of pressure on the big AI providers. To live up to the expectations, they need to push into real production. According to a MIT research from July 2025, 95% of of businesses&#8217; AI pilots failed to generate a return at all. That means Big Tech needs to act fast.</p><h2>Might agents be the solution to a stalling adoption curve?</h2><p>To push past the adoption friction, the industry is leaning into agents. Systems that take tasks and execute them end to end. From a buyer&#8217;s perspective, outcomes are easier to understand than prompts. From a founder&#8217;s perspective, results are easier to sell than assistance. It&#8217;s as simple as replacing humans, not augmenting their capabilities.</p><p>From an operator&#8217;s perspective, agents surface the hardest questions immediately. Responsibility, failure modes, and substitution stop being abstract. Startups like <strong>Artisan</strong> framing agents as replacements for human beings rather than tools may gain attention, but they also turn technical products into cultural statements.</p><h2>What 2026 actually sorts out</h2><p>By the end of 2026, it will not be clear whether AI &#8220;won,&#8221; but it will be clear which advantages survived contact with a reset. Some companies will have restructured workflows deeply enough to make AI stick. Some startups will prove they actually control their cost curves. Many others will turn out to be compounding inside someone else&#8217;s system, on borrowed time.</p><p>From my perspective, AI does not need to fail to disappoint. It only needs to scale more slowly than expected, cost more than hoped, and concentrate value higher up the stack than most narratives assume.</p><p>That is why 2026 feels less like a breakthrough year and more like a sorting year. Not between AI and non-AI companies, but between businesses that can survive repeated resets and those that quietly plateau while the story moves on.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Why VC Math Encourages Fragile AI Startups]]></title><description><![CDATA[I recently reread The Power Law, and it clicked in a very uncomfortable way with what I see every day working on AI products.]]></description><link>https://www.techfundamentals.blog/p/why-vc-math-encourages-fragile-ai</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/why-vc-math-encourages-fragile-ai</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Wed, 31 Dec 2025 14:46:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!OKRQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OKRQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OKRQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 424w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 848w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OKRQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2649573,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183058689?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OKRQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 424w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 848w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!OKRQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F246123d1-131d-416f-a1e7-36c398e0cdcf_5824x3264.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>For context, I&#8217;m a software engineer by background, but I&#8217;ve also built startups, run a boutique agency for years, and spend most of my days designing and shipping software products, nowadays often with AI components. I&#8217;m not watching this market from the sidelines. I&#8217;m inside the systems, the invoices, the latency issues, the model limits, and the cost surprises. That proximity changes how you interpret the stories we tell about AI startups.</p><p>Especially the venture-backed ones.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>Power law returns shape everything downstream</h3><p>The core argument of <em>The Power Law</em> is not just that VC returns are uneven. It&#8217;s that they are violently uneven. One or two companies can return an entire fund, while most others go to zero or limp to an acquihire.</p><p>Once you internalize that, a lot of otherwise confusing behavior starts to make sense.</p><p>VCs are not optimizing for average outcomes. They are optimizing for outliers. That means they are structurally indifferent to fragility as long as there is a credible path to extreme upside.</p><p>This matters because founders adapt to incentives. Consciously or unconsciously, you build what gets funded. And in AI right now, what gets funded is speed, surface-level magic, and the appearance of leverage.</p><h3>Why &#8220;just build on top of the best model&#8221; is rational</h3><p>From a purely technical perspective, building an AI wrapper can feel irresponsible. You don&#8217;t control the model. You don&#8217;t control pricing. You don&#8217;t control rate limits. You don&#8217;t control whether a new API release makes your core feature obsolete overnight.</p><p>But from a founder trying to raise venture capital, it&#8217;s often the most rational move available.</p><p>Wrappers let you move fast. They let you demo something impressive with a small team and they let you show usage graphs before you&#8217;ve solved any hard infrastructure problems. But most importantly, they let you tell a story that fits power law thinking.</p><p>&#8220;If this category explodes, we could be the category leader.&#8221;</p><p>That sentence matters more than unit economics in early-stage VC. And AI wrappers are excellent vehicles for that narrative.</p><p>I don&#8217;t say this as an outsider criticizing naive founders. I&#8217;ve been on both sides of this. I know how tempting it is to defer hard problems when momentum is rewarded more than resilience.</p><h3>The illusion of leverage in AI products</h3><p>AI makes this even trickier because it creates the illusion of leverage everywhere.</p><p>A small team can now ship something that would have taken a large organization a few years ago. That&#8217;s real. But the leverage often belongs to the model provider, not the startup.</p><p>When working on AI products for my company and clients, I see the same pattern over and over. Early costs look trivial. Latency feels acceptable. The system works well enough. Then usage or requirements grow, edge cases pile up, prompts get more complex, and suddenly the cost curve shows up.</p><p>At that point, if you don&#8217;t control the underlying system, your margins are someone else&#8217;s variable.</p><p>From a VC perspective, that&#8217;s fine. If the company breaks, it breaks. The fund only needs one winner. From a founder or operator perspective, it&#8217;s terrifying.</p><h3>Fragility is not accidental, it&#8217;s selected for</h3><p>One thing <em>The Power Law</em> makes very clear is that venture capital doesn&#8217;t accidentally produce fragile companies. It selects for them.</p><p>Speed beats correctness. Growth beats efficiency. Story beats control.</p><p>If you try to build something slower, more robust, and more boring, you often get punished in fundraising. Your charts look worse and your pitch sounds less exciting. Your upside feels capped, even if your downside is far better managed.</p><p>Again, I&#8217;ve felt this tension personally. Running an agency forces you to think in terms of cash flow, trust, and systems that don&#8217;t fall apart when something upstream changes. Those instincts actively work against the kind of risk-taking VC wants to see.</p><p>AI just amplifies that mismatch.</p><h3>The dependency nobody prices correctly</h3><p>The quiet problem with many AI startups is not that they will fail. Failure is normal. It&#8217;s that their dependency structure caps their long-term value even if they succeed.</p><p>If your core capability is rented, your differentiation has a ceiling. You can have great UX, strong distribution, and excellent branding, but if the underlying intelligence is commoditized and controlled by someone else, your negotiating power erodes over time and you can get easily replaced by competitors or evolving AI agents.</p><p>I see a lot of founders assume they&#8217;ll &#8220;figure it out later.&#8221; Replace the model. Negotiate pricing. Build proprietary data moats. Maybe all of that happens. But maybe it doesn&#8217;t.</p><p>From the VC side, that uncertainty is acceptable. From the founder side, it&#8217;s existential.</p><h3>How this changed how I invest and build</h3><p>As an investor or customer of an AI project, I&#8217;ve become much more skeptical of AI companies that look impressive quickly. I don&#8217;t ask whether the demo works. I ask if the team behind the product is able to adapt swiftly. This assumption is especially important when I meet AI companies without a technical cofounders - or especially - vibe coded AI products built by marketing and sales people who obviously don&#8217;t know what&#8217;s happening under the hood.</p><p>As a builder, I&#8217;m increasingly biased toward projects that look boring in the early days. Things with real constraints. Things that don&#8217;t explode on Twitter. Things where the hardest work happens before the story gets interesting.</p><p>Power law math encourages founders to chase optionality over durability. That&#8217;s not a moral failure. It&#8217;s a structural one.</p><p>But if you&#8217;re an operator, an employee, or an investor who actually cares about long-term outcomes, you need to see that clearly. Most AI startups are not designed to last. They&#8217;re designed to be asymmetric bets.</p><p>Once you see that, the current AI boom looks less like a gold rush and more like a very efficient machine for producing impressive demos, fragile businesses, and a small number of enormous winners.</p><p>And the winners probably won&#8217;t look like wrappers at all. They&#8217;ll look slow, capital-intensive, and unsexy for a very long time. Which is exactly why most people will miss them.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How Compounding Breaks When Technology Resets Every Cycle]]></title><description><![CDATA[A guide for stock traders & VCs inspired by my career spent rebuilding on moving ground and adapting when the wind changes]]></description><link>https://www.techfundamentals.blog/p/how-compounding-breaks-when-technology</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/how-compounding-breaks-when-technology</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Wed, 31 Dec 2025 10:19:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-AR5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-AR5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-AR5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-AR5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg" width="1456" height="888" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:888,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2437660,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/183040788?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-AR5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-AR5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fda343f-4943-465b-98de-4740d143d93d_3612x2204.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>The longer my career as a founder and engineer gets, the more one idea from <em>The Psychology of Money</em> keeps resurfacing: long-term success isn&#8217;t about brilliance, timing, or even hard work in isolation. It&#8217;s about staying in the game long enough for compounding to matter. That&#8217;s also a lot what I read about here on Substack.</p><p><strong>Compounding</strong> is slow, boring, and deeply unfair in the short term. That&#8217;s why it works.</p><p>But the more years you spend inside technology - not just investing in it, but building on top of it - the clearer it becomes that tech is one of the most hostile environments for clean compounding.</p><p>Not because growth isn&#8217;t real, but because the rules reset far more often than our mental models assume.</p><div><hr></div><h3>Technology doesn&#8217;t compound, it overwrites</h3><p>Most technological progress isn&#8217;t additive. It replaces what came before.</p><p>Desktop computing didn&#8217;t smoothly compound into mobile. Mobile shifted where value was created and erased entire categories of software businesses. On-prem infrastructure didn&#8217;t evolve gracefully into cloud services; it became stranded capital almost overnight. Traditional software isn&#8217;t slowly absorbing AI - it&#8217;s being repriced, re-scoped, and in some cases made irrelevant.</p><blockquote><p>Each cycle quietly wipes out part of the previous advantage stack. Experience becomes table stakes and tooling turns into legacy debt. The Moats shrink into features and what once differentiated you becomes the minimum requirement just to stay in the conversation.</p></blockquote><p>You don&#8217;t go back to zero, but you also don&#8217;t start from where your intuition tells you that you should. And that gap, between perceived leverage and actual leverage, is where compounding quietly dies.</p><div><hr></div><h3>I&#8217;ve lived this personally as a consultancy founder</h3><p>Running an IT consultancy looks, on paper, like a textbook compounding business. Experience accumulates. Reputation spreads. Client relationships deepen. Every year should make the next one easier.</p><p>That&#8217;s the story. The lived reality is far less linear.</p><p>Every few years, the technical foundation underneath the business shifts. And when that happens, a meaningful portion of accumulated advantage simply loses relevance.</p><p>I&#8217;ve watched this play out repeatedly. From LAMP stacks to MERN. From server-rendered applications to API-driven frontends. From Magento-heavy projects to Shopify-centric ecosystems. From owning and tuning infrastructure to living entirely on managed platforms.</p><p>Each transition forces uncomfortable questions. Who do we hire now? What do we even sell? How do we price work when platforms abstract away complexity we used to bill for?</p><p>The business doesn&#8217;t reset to zero - but it absolutely does not compound smoothly. It feels more like climbing a staircase in poor lighting. You take steady steps for a while, then suddenly the geometry changes. Miss one transition and you don&#8217;t crash dramatically. You just stop ascending while others quietly move ahead.</p><p>That&#8217;s what broken compounding looks like from the inside in my opinion.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/p/how-compounding-breaks-when-technology?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/p/how-compounding-breaks-when-technology?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.techfundamentals.blog/p/how-compounding-breaks-when-technology?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div><hr></div><h3>Public tech companies aren&#8217;t immune, they just hide it better</h3><p>It&#8217;s tempting to believe this problem only applies to small and mid sized companies or agencies like mine. Public tech companies, we tell ourselves, are different. They have scale, capital, and optionality.</p><p>But even the best examples show how fragile compounding really is.</p><p><strong>Microsoft</strong> didn&#8217;t smoothly compound from Windows dominance into cloud leadership. There was a long stretch where the company looked culturally and strategically stuck. Azure wasn&#8217;t an incremental extension, but it was a full reset that happened just in time.</p><p><strong>IBM</strong> survived multiple computing eras, but survival isn&#8217;t the same as compounding. Its influence and growth flattened long before the narrative caught up, precisely because each platform shift diluted prior advantages.</p><p><strong>Adobe</strong> is often held up as a compounding success story thanks to its subscription transition. And it deserves credit. But that transition wasn&#8217;t optional. Without it, decades of dominance would&#8217;ve turned into structural obsolescence very quickly (looking at you Sketch, Affinity and Figma).</p><p>Even companies often seen as perpetual compounders tell the same story under the surface. <strong>Amazon</strong> didn&#8217;t compound by staying still. It repeatedly rebuilt itself, using cash flow from one layer to finance reinvention in another. <strong>Meta</strong> rode social platforms to enormous scale, only to discover how fragile that compounding becomes when user behavior or platform paradigms shift.</p><p>From the outside, these stories look inevitable, but from the inside, they are a sequence of narrowly avoided dead ends.</p><div><hr></div><h3>Survivorship bias is brutal in technology</h3><p>What we mostly see are the winners who crossed multiple resets. What we don&#8217;t see are the far more common outcomes.</p><p>Companies that didn&#8217;t fail loudly. They just stalled.</p><p>The product still works and the team is still competent. Revenue still grows a little, but margins compress. Hiring gets harder. Strategy becomes reactive instead of intentional. The company turns into a maintenance machine for decisions made in a previous cycle.</p><p>This is the most dangerous failure mode for founders because it doesn&#8217;t feel like failure. And it&#8217;s the most dangerous failure mode for investors because it looks like stability.</p><p>Compounding doesn&#8217;t always end with a crash. Often it ends with a long, quiet plateau.</p><div><hr></div><h3>AI doesn&#8217;t fix this - it accelerates it</h3><p>There&#8217;s a seductive belief that AI finally restores compounding to software. Smarter tools, faster iteration, more leverage per person - what could go wrong?</p><p>Working with AI systems on a daily basis tells a different story.</p><p>AI dramatically lowers the cost of building, which is real progress. But it also dramatically shortens the half-life of advantage for anyone who doesn&#8217;t control their own economics. Many AI products grow quickly while sitting on top of models, infrastructure, and pricing decisions owned by someone else.</p><div><hr></div><h3>This looks familiar if you&#8217;ve lived through platform shifts</h3><p>If you&#8217;ve been building long enough, AI wrapper economics should feel uncomfortably familiar. </p><blockquote><p>Want to know what AI wrappers are? I&#8217;ve written another post about it: <a href="https://substack.com/home/post/p-182769365">https://substack.com/home/post/p-182769365</a></p></blockquote><p>It feels like businesses built entirely on social media APIs or mobile apps whose fate depended on App Store policy changes. Or SaaS products whose distribution vanished when a platform reprioritized its roadmap.</p><p>Those companies weren&#8217;t stupid. Many were well executed. Some even became case studies for a while. But when the platform shifted, past effort stopped compounding. Execution quality mattered less than position in the stack.</p><p>AI is replaying this pattern at a much faster pace, with far more capital and far less patience.</p><div><hr></div><h3>The real compounding question founders should ask</h3><p>Re-reading <em>The Psychology of Money</em> pushed me to reframe the core question.</p><p>It&#8217;s not &#8220;how fast can this grow?&#8221;<br>It&#8217;s &#8220;what survives when the reset happens?&#8221;</p><p>What still compounds when the underlying platform changes? What advantage carries forward? And what quietly evaporates?</p><p>The companies that manage to compound across cycles aren&#8217;t the ones that avoid change. They&#8217;re the ones that can absorb it without destroying their economics. They own something fundamental enough that resets don&#8217;t erase their leverage.</p><p>Everyone else is compounding inside someone else&#8217;s system, on borrowed time, under assumptions they don&#8217;t control.</p><div><hr></div><h3>A more honest model of tech compounding</h3><p>Compounding in technology isn&#8217;t a smooth exponential curve. It&#8217;s a staircase.</p><p>You accumulate quietly for years, then hit sharp transitions where the rules change. Some founders step up. Some slip. Some realize, too late, that the staircase moved while they were still climbing the previous step.</p><p>As a founder, this means staying paranoid about relevance even when things are going well. Comfort is often a lagging indicator of decay. As an investor, it means distrusting stories that project yesterday&#8217;s advantage too cleanly into tomorrow.</p><p>Compounding still matters. But in technology, it only belongs to those who survive and that group is far smaller than the success stories make it seem.</p><div><hr></div><h3>What this means for investing in tech companies</h3><p>This compounding problem doesn&#8217;t stop at founders. It leaks directly into how tech gets funded - and why so much capital underperforms expectations across cycles.</p><p>For <strong>VCs</strong>, the core mistake is often assuming that early traction plus a big market automatically implies a long compounding runway. In reality, a lot of venture-scale outcomes are really bets on <em>timing a cycle</em>, not owning a durable advantage.</p><p>Many startups look like compounders only because the platform underneath them is still stable. When that platform shifts - cloud primitives change, distribution moves, AI pricing resets - the fund isn&#8217;t underwriting growth anymore. It&#8217;s underwriting the company&#8217;s ability to survive a reset it doesn&#8217;t control.</p><p>This is why so many VC portfolios end up with a strange shape: a few extreme winners that managed to align with or control a platform transition, and a long tail of companies that never quite die but also never deliver venture returns. They didn&#8217;t fail at execution. They failed at carrying leverage across cycles.</p><p>For <strong>public market investors</strong>, the problem shows up differently but just as painfully.</p><p>Public tech investing is full of implicit compounding assumptions. Revenue growth gets extrapolated. Margins are modeled as improving over time. Moats are assumed to widen with scale. But in technology, scale often <em>increases exposure</em> to resets instead of protecting against them.</p><p>A SaaS company compounding at 30% looks amazing, until a platform change turns pricing power into a negotiation. An AI-heavy product looks defensible, until inference costs or model access terms change. Suddenly the long-term margin story that justified the multiple doesn&#8217;t exist anymore.</p><p>This is why so many tech stocks don&#8217;t collapse, they just de-rate. Growth slows a bit. Margins disappoint slightly. Guidance gets conservative. The stock goes sideways for years while investors wait for compounding that never really resumes.</p><p>The uncomfortable truth is that many tech investments aren&#8217;t long-term compounding bets at all. They&#8217;re <strong>cycle bets disguised as forever businesses</strong>.</p><p>The investors who tend to do better over long horizons are usually asking different questions. Not &#8220;how big can this get?&#8221; but &#8220;what does this company still control after the next reset?&#8221; Not &#8220;is this growing fast?&#8221; but &#8220;where does the cost curve live, and who owns it?&#8221;</p><p>Because in technology, returns don&#8217;t accrue to the companies that grow the fastest inside a cycle. They accrue to the ones that still matter when the cycle ends.</p><p>And that&#8217;s a much rarer and much harder thing to underwrite.</p><div><hr></div><h3>Closing thoughts</h3><p><strong>The biggest mistake we make with technology, whether as founders, operators, or investors, is assuming that time automatically works in our favor.</strong></p><p>In many industries, it does. Experience stacks. Advantages harden. Compounding feels almost mechanical. In technology, time is far more conditional. It only helps if you are aligned with where the system is going next, not where it has already been.</p><p>Building in tech means repeatedly letting go of things that once worked. Investing in tech means accepting that many &#8220;great&#8221; businesses are only great within the boundaries of a specific cycle. When those boundaries move, the compounding story often breaks long before the narrative does.</p><blockquote><p>The uncomfortable reality is that technology rewards adaptability more than consistency, and positioning more than effort. That doesn&#8217;t make compounding impossible, but it does make it rare, fragile, and uneven.</p></blockquote><p><strong>If there is a single mental shift worth making, it is this: stop asking whether something is a good business today. Start asking whether its advantages survive a reset it does not control.</strong></p><p>Because in technology, the future rarely belongs to the best operators of the current system. It belongs to the ones who are still standing, and still relevant, after the system forgets its own past.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[The AI Boom Has a Graveyard No One Talks About]]></title><description><![CDATA[A Software Engineer's Perspective - The Hidden Dependency Behind Most AI Products And Why Many Will Fail In The Long Run]]></description><link>https://www.techfundamentals.blog/p/the-ai-boom-has-a-graveyard-no-one</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/the-ai-boom-has-a-graveyard-no-one</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Sun, 28 Dec 2025 14:20:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JcXP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JcXP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JcXP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JcXP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg" width="1456" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2876348,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://techfundamentals.substack.com/i/182769365?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JcXP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JcXP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5e77b5b-95ac-4c0f-9d3a-5c65c162e399_4486x2671.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p><br>I&#8217;ve been rereading <strong>The Psychology of Money</strong> by <strong>Morgan Housel</strong> recently, and it&#8217;s been hard not to see the parallels with how people are investing in AI right now. Housel&#8217;s core point is that most mistakes aren&#8217;t about missing information. They&#8217;re about how we interpret success after the fact.</p><p>That hit a little closer to home because I spend my days working as a software engineer, building and integrating AI products. I&#8217;m inside these systems constantly. I see where the power lives, where the costs show up, and how much of what looks like &#8220;AI innovation&#8221; is actually dependency dressed up as product.</p><p>AI investing feels like a textbook case of psychology getting ahead of system reality.</p><div><hr></div><h3>The Part of the AI Story That Feels Too Obvious</h3><p>Most AI conversations today sound strangely settled. People point to a few winners, talk about adoption curves, and speak as if the market structure was inevitable. Of course these companies won. Of course this is where the value landed.</p><p>That sense of inevitability is exactly what Housel warns about. When outcomes feel obvious in hindsight, we stop asking what had to go right and what risks are being quietly ignored.</p><div><hr></div><h3>Most AI Products Don&#8217;t Own the Intelligence</h3><p>This is something that becomes very clear when you actually work with AI systems every day.</p><p>Most AI products do not own the intelligence that makes them work. They sit on top of it.</p><p>A large share of AI startups are thin layers on top of a small number of foundational providers, most commonly OpenAI. They add a workflow, a UI, some prompt logic, maybe light fine-tuning, and wrap it into something usable. That can still be valuable. But it&#8217;s very different from owning the core system.</p><p>When people say &#8220;AI company,&#8221; what they often mean is &#8220;company that calls an API and makes it feel nice.&#8221;</p><div><hr></div><h3>Dependency Is Not a Phase You Grow Out Of</h3><p>From a software perspective, dependency always has a cost. In AI, it defines the business.</p><p>If you rely on an upstream model provider, you don&#8217;t control pricing, inference efficiency, performance improvements, or roadmap direction. Your margins and differentiation are downstream of decisions you don&#8217;t make.</p><p>That&#8217;s not a scaling issue you solve later. That&#8217;s the structure you&#8217;re building on.</p><p>One thing Housel emphasizes is how easy it is to underestimate risks that don&#8217;t show up immediately. Dependency feels fine early. It only becomes painful once you&#8217;re successful enough for it to matter.</p><div><hr></div><h3>Why So Many AI Companies Stall Instead of Fail</h3><p>This is why most AI companies don&#8217;t fail dramatically.</p><p>Early on, costs are falling, credits are generous, and competition is light. Usage grows and everything looks validated. Then scale arrives. Inference costs become very real. Customers start questioning pricing. Features you thought were differentiators turn into baseline expectations as models improve.</p><p>From the inside, nothing breaks. From the outside, growth just stops being impressive.</p><p>That&#8217;s survivorship bias at work. We remember the few that broke through and forget the many that quietly stalled.</p><div><hr></div><h3>The AI Graveyard Is Quiet</h3><p>Working with AI products daily makes this especially obvious. Most projects don&#8217;t collapse. They just hit economic ceilings.</p><p>Margins never show up. The product remains useful but not defensible. Eventually the company becomes strategically irrelevant or gets absorbed. These outcomes don&#8217;t make headlines, which is why the graveyard is easy to miss.</p><p>But it&#8217;s full.</p><div><hr></div><h3>The Wrapper Problem Isn&#8217;t an Insult, It&#8217;s a Constraint</h3><p>Calling something an AI &#8220;wrapper&#8221; isn&#8217;t an insult. It&#8217;s a description of where power sits.</p><p>If the underlying model improves, your product gets commoditized. If pricing goes up, your margins compress. If the provider launches something adjacent, your differentiation disappears. Switching costs are lower than they look once customers understand what&#8217;s underneath.</p><p>From a software standpoint, you&#8217;re competing on UX, speed, and distribution, not on system control. That can work for a while. It rarely compounds.</p><div><hr></div><h3>AI Feels Like Software, but Behaves Like Infrastructure</h3><p>This is where a lot of expectations break.</p><p>AI feels like software because you can ship fast and iterate. But once usage grows, it behaves like infrastructure. Hardware efficiency, energy costs, networking, and inference economics dominate outcomes. You can&#8217;t refactor your way out of those constraints.</p><p>If you don&#8217;t sit close to the cost curve, scale amplifies the problem instead of solving it.</p><div><hr></div><h3>Why Infrastructure Owners Sit in a Different Position</h3><p>This is why companies closer to the metal operate in a different reality.</p><p>NVIDIA and large cloud platforms like Amazon Web Services, Microsoft Azure, and Google Cloud sit where costs are set, not just passed through. That doesn&#8217;t mean they&#8217;re guaranteed great returns. It does mean they&#8217;re not structurally dependent in the same way most AI products are.</p><p>From an engineering perspective, owning infrastructure is painful. From an investing perspective, not owning it caps upside.</p><div><hr></div><h3>A Simple Test I Keep Coming Back To</h3><p>One simple mental model I keep using, both as an engineer and as an investor, is this:</p><p>What happens to this business if its core AI provider changes pricing or launches a competing feature?</p><p>If the answer is &#8220;that would really hurt,&#8221; then you&#8217;re not looking at a moat. You&#8217;re looking at a dependency that only feels manageable while conditions are unusually friendly.</p><div><hr></div><h3>Closing Thoughts</h3><p>AI is real. The technology is powerful. Working with these systems daily makes that very clear.</p><p>What <em>The Psychology of Money</em> helped reinforce for me is how much of today&#8217;s AI market is shaped by survivorship bias and hidden dependencies. A small group of companies own critical layers of the stack. A much larger group depends on them and hopes the economics work out.</p><p>Survivorship bias makes it feel like everyone has a shot.<br>In reality, most of the graveyard is made up of companies that never failed loudly. They just never escaped the stack they were built on.</p><p>The hardest part of AI isn&#8217;t building something impressive.<br>It&#8217;s building something that still works when the psychology shifts and the easy phase is over.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.techfundamentals.blog/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Tech Fundamentals! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Shopify: A Software Platform Disguised as a Retail Stock]]></title><description><![CDATA[Why Shopify is still misunderstood by the market]]></description><link>https://www.techfundamentals.blog/p/shopify-a-software-platform-disguised</link><guid isPermaLink="false">https://www.techfundamentals.blog/p/shopify-a-software-platform-disguised</guid><dc:creator><![CDATA[Daniel Kolb]]></dc:creator><pubDate>Fri, 26 Dec 2025 14:54:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vva8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vva8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vva8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 424w, https://substackcdn.com/image/fetch/$s_!vva8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 848w, https://substackcdn.com/image/fetch/$s_!vva8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!vva8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vva8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg" width="1456" height="858" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:858,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!vva8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 424w, https://substackcdn.com/image/fetch/$s_!vva8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 848w, https://substackcdn.com/image/fetch/$s_!vva8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!vva8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a4a08e-9043-4d44-af21-9e7768e14ad7_2000x1179.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p><strong>Disclaimer: </strong>This publication and its authors are not licensed investment professionals. Nothing posted on this blog should be construed as investment advice. Do your own research.</p></blockquote><p>At first glance, <strong>Shopify</strong> looks like a leveraged bet on consumer spending. When retail weakens, the stock sells off. When e-commerce optimism returns, it rallies. That surface-level framing misses what Shopify actually is.</p><p>Shopify is not a retailer. It is not a marketplace. It is not a logistics company in the Amazon sense. Shopify is an operating system for commerce. It is a modular, API-driven software platform that lets millions of small and mid-sized businesses run online commerce without building infrastructure themselves.</p><p>Understanding Shopify as a software platform, not a retail proxy, changes how you should think about its durability, margins, risks, and long-term upside.</p><p>This post breaks Shopify down from a technical and systems perspective and explains why the stock continues to confuse both bulls and bears.</p><div><hr></div><h2>Shopify&#8217;s core product is abstraction</h2><p>Shopify&#8217;s real product is not storefront templates or checkout buttons. Its real product is abstraction.</p><p>It abstracts away:</p><ul><li><p>Payments infrastructure</p></li><li><p>Tax compliance</p></li><li><p>Hosting and scalability</p></li><li><p>Fraud prevention</p></li><li><p>Inventory logic</p></li><li><p>App integrations</p></li><li><p>Cross-border commerce complexity</p></li></ul><p>For a merchant, Shopify replaces what would otherwise require multiple vendors, custom engineering, and ongoing operational risk. For a developer, Shopify provides a stable, opinionated platform with APIs that rarely break and documentation that actually works.</p><p>This abstraction is where the moat begins. Once a merchant builds workflows, themes, apps, and operational habits around Shopify, switching costs become real. Not contractual, but practical.</p><div><hr></div><h2>The merchant base is fragmented by design</h2><p>A common criticism of Shopify is merchant quality. Many stores fail. Many merchants are small. Average revenue per merchant looks unimpressive compared to enterprise SaaS.</p><p>This critique misunderstands Shopify&#8217;s strategy.</p><p>Shopify intentionally targets a fragmented, long-tail market. That fragmentation protects Shopify from customer concentration risk and from enterprise procurement cycles. A single merchant failing does not matter. Millions of merchants experimenting does.</p><p>Shopify is closer to a protocol than a traditional SaaS vendor. It monetizes experimentation at scale.</p><p>This also explains why Shopify optimizes for self-serve onboarding and product-led growth rather than sales-led expansion. The platform is designed to let merchants try, fail, retry, and grow without human intervention.</p><p>That design choice limits short-term monetization but increases long-term surface area.</p><div><hr></div><h2>Payments is the economic engine</h2><p>Shopify&#8217;s most important business decision was pushing hard into payments.</p><p>Shopify Payments is not just a convenience feature. It is the primary mechanism for monetization expansion. Every dollar of Gross Merchandise Volume that flows through Shopify Payments generates revenue with minimal incremental cost.</p><p>This shifts Shopify&#8217;s revenue mix away from flat subscription fees toward usage-based economics.</p><p>From an investor perspective, this matters for two reasons.</p><p>First, Shopify&#8217;s take rate improves naturally as merchants grow. No upsell call required.</p><p>Second, Shopify becomes structurally tied to merchant success rather than merchant count. This aligns incentives in a way many SaaS companies fail to achieve.</p><p>The trade-off is exposure to consumer spending cycles. When GMV slows, revenue growth slows. That volatility is real and should not be ignored.</p><div><hr></div><h2>Shopify versus Amazon is the wrong comparison</h2><p>Shopify is often compared to <strong>Amazon</strong>. This comparison is mostly useless.</p><p>Amazon is a centralized marketplace. Shopify is decentralized infrastructure.</p><p>Amazon owns the customer relationship, the discovery layer, the fulfillment stack, and the data. Merchants rent access.</p><p>Shopify explicitly does the opposite. Merchants own the customer, the brand, the data, and the traffic acquisition strategy. Shopify stays in the background.</p><p>This difference matters because it defines the ceiling and the risks.</p><p>Amazon can extract more value per transaction but faces regulatory scrutiny, merchant resentment, and internal conflicts of interest.</p><p>Shopify monetizes less aggressively but avoids platform abuse accusations and antitrust pressure. It is harder to regulate infrastructure than a marketplace that sets prices and competes with its own sellers.</p><p>Shopify&#8217;s model scales more quietly and more globally.</p><div><hr></div><h2>Fulfillment was a strategic misstep and a valuable lesson</h2><p>Shopify&#8217;s attempt to build a first-party fulfillment network was widely criticized, and correctly so.</p><p>Logistics is capital-intensive, margin-thin, and operationally complex. It conflicts with Shopify&#8217;s asset-light software DNA.</p><p>The decision to divest the fulfillment business was not a failure. It was a course correction.</p><p>What Shopify learned is important. Merchants want integration, not ownership. They want software orchestration, not Shopify-branded warehouses.</p><p>Shopify&#8217;s current approach focuses on connecting merchants to third-party logistics providers through software rather than competing with them.</p><p>This keeps Shopify aligned with its core competency while still participating in the value chain.</p><div><hr></div><h2>Operating leverage is real but delayed</h2><p>Shopify&#8217;s income statement has frustrated investors for years. Revenue grows, but margins lag. Stock-based compensation remains high. Profitability appears optional.</p><p>This is partially true, and partially misleading.</p><p>Shopify deliberately reinvested heavily in product, international expansion, and ecosystem tooling during a period of unusually cheap capital. That era is over.</p><p>The important question is not whether Shopify can be profitable. It is whether profitability scales faster than revenue once investment slows.</p><p>Because Shopify&#8217;s infrastructure costs scale sublinearly relative to GMV, operating leverage should emerge over time. Payments, apps, and services carry much higher incremental margins than subscriptions.</p><p>The risk is timing. Investors expecting near-term margin expansion may be disappointed. Investors with a longer horizon should focus on unit economics rather than quarterly optics.</p><div><hr></div><h2>The developer ecosystem is the quiet moat</h2><p>Shopify&#8217;s app ecosystem is one of its least discussed advantages.</p><p>Thousands of developers build specialized tools on top of Shopify. These apps solve niche problems that Shopify itself should never prioritize. The result is a marketplace of functionality that increases platform stickiness without increasing internal headcount.</p><p>This is classic platform leverage.</p><p>Every successful app makes Shopify more valuable. Every failed app costs Shopify nothing.</p><p>From a systems perspective, this is a compounding advantage that is difficult to replicate. Competing platforms need both scale and trust to attract developers. Shopify already has both.</p><div><hr></div><h2>Where Shopify can fail</h2><p>A serious analysis requires acknowledging failure modes.</p><p>Shopify is exposed to:</p><ul><li><p>Prolonged consumer spending weakness</p></li><li><p>Payment margin compression from competitors</p></li><li><p>Increased regulatory scrutiny on payments and data</p></li><li><p>Merchant acquisition costs rising due to ad platform consolidation</p></li><li><p>Platform fatigue if complexity grows faster than usability</p></li></ul><p>Shopify also sits downstream from platforms like <strong>Meta</strong> and <strong>Google</strong> for traffic. Changes in ad pricing or tracking rules indirectly affect merchant success and therefore Shopify&#8217;s GMV.</p><p>None of these risks are existential, but they cap short-term optimism.</p><div><hr></div><h2>Why Shopify remains interesting as a long-term investment</h2><p>Shopify is not a clean story. It is cyclical, volatile, and often expensive on traditional multiples.</p><p>But from a product and systems perspective, Shopify has built something rare. A globally scalable commerce operating system with embedded payments, strong developer gravity, and aligned incentives.</p><p>It benefits from:</p><ul><li><p>Long-term shift toward independent brands</p></li><li><p>Increasing complexity of global commerce</p></li><li><p>Merchant preference for ownership over marketplaces</p></li><li><p>Software-driven operating leverage</p></li></ul><p>Shopify is unlikely to dominate headlines the way AI infrastructure companies do. It compounds quietly, tied to real economic activity rather than speculative demand.</p><p>For investors willing to tolerate volatility and think in systems rather than narratives, Shopify remains one of the more structurally interesting platforms in public markets.</p><p>Not because it is perfect, but because it is hard to replace.</p>]]></content:encoded></item></channel></rss>