<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[ROCH Labs]]></title><description><![CDATA[Data driven investment analysis for AI and Crypto]]></description><link>https://www.rochlabs.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Q9Il!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2717745f-e35e-4e1b-b3ab-7ff3399c683f_768x768.png</url><title>ROCH Labs</title><link>https://www.rochlabs.com</link></image><generator>Substack</generator><lastBuildDate>Tue, 14 Apr 2026 03:38:01 GMT</lastBuildDate><atom:link href="https://www.rochlabs.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rohit Chauhan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rochlabs@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rochlabs@substack.com]]></itunes:email><itunes:name><![CDATA[Rohit Chauhan]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rohit Chauhan]]></itunes:author><googleplay:owner><![CDATA[rochlabs@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rochlabs@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rohit Chauhan]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Tether AI: The Genius of Low-Rank Adapation or LoRA]]></title><description><![CDATA[Tether's QVAC Fabric dropped today: the world's first cross-platform BitNet + LoRA Framework]]></description><link>https://www.rochlabs.com/p/tether-ai-the-genius-of-low-rank</link><guid isPermaLink="false">https://www.rochlabs.com/p/tether-ai-the-genius-of-low-rank</guid><dc:creator><![CDATA[Rohit Chauhan]]></dc:creator><pubDate>Tue, 17 Mar 2026 18:30:59 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/27cb179d-cff3-447c-bed8-e1f6197857b8_1280x720.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<ul><li><p>Earlier today Tether&#8217;s founder and CEO, Paolo Ardoino announced Tether&#8217;s QVAC Fabric breakthrough. </p></li><li><p>The QVAC Fabric LLM is a framework from Tether Data (the team behind the stablecoin,</p><p><a href="https://x.com/search?q=%24USDT&amp;src=cashtag_click">$USDT</a></p><p>) to decentralize AI fine tuning, and enabling private AI.</p></li><li><p>Consider this, if you&#8217;re a MacBook Air M1 user, this framework is particularly relevant to you because the M1 chip&#8217;s Unified Memory Architecture is exactly the kind of &#8220;ordinary device&#8221; this framework is designed to exploit. </p></li><li><p>The QVAC Fabric is a free, open source software framework (think &#8220;toolkit&#8221; or &#8220;engine&#8221;) which lets anyone run and personalize powerful AI models directly.</p></li></ul><p>Usually, if you want a model to know your specific writing style (like your twitter voice) or your specific niche (mine is financial data analysis), you have to upload that data to a cloud provider like Microsoft Azure, which powers OpenAI&#8217;s infrastructure.)</p><p>To receive persaonlized responses from an LLM provider like OpenAI, your data must be sent to and stored on their remote servers.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><blockquote><p><em><strong>QVAC changes this as the framework allows data to stay on your device, and bypass the cloud entirely because the computation happens locally on your hardware. </strong></em></p></blockquote><p>While for most things, it probably doesn&#8217;t matter for the cloud to have access to all your data. But for sensitive stuff, like financial audits, or proprietary investment research, using a local framework like QVAC will allow you to retain 100% custody.</p><p>A pre-trained LLM like Llama, Qwen is like a super-smart but generic university professor who knows everything about the world in general. If you want this professor to become your personal doctor AI (that understands your specific health history, diet, symptoms, and speaks in your style) you need to teach them your private medical notes, blood tests, Whoop data etc.</p><blockquote><p><em><strong>teaching = fine-tuning = updating the model&#8217;s knowledge with your data.</strong></em></p></blockquote><p> In the old way, we would retrain the entire professor from head to toe, and change very single connection in their brain. For a 13B parameter model, that&#8217;s billions of numbers to update. This needs huge servers, costs a fortune in electricity/GPUs, takes days/weeks, and your private data has to leave your phone. </p><p>This is where LoRA (Low-Rank Adaptation) is genius. LoRA says don&#8217;t touch the professor&#8217;s core brain at all. Instead, give them a tiny set of clip-on cheat sheets or post-it notes that they can consult everytime they answer. </p><p>These post-its only contain a very small number of new instructions (often just 0.1% to 1% of the original model&#8217;s size - up to 99% fewer parameters to actually train.)</p><p>The original professor stays frozen (unchanged, safe, no risk of &#8220;forgetting&#8221; general knowledge.)</p><p>When the professor speaks, they combine: original brain (frozen) + tiny LoRA post-its (your personal tweaks). </p><p>The answer feels fully personalized to you, but almost no extra compute/memory is needed. </p><p>Mathematically its clever. The update to any big weight matrix is approximated as the product of two much smaller matrices (low rank = skinny rectangles instead of huge squares). So instead of changing 10,000 x 10,000 numbers, you only train something like 100 x 10,000 + 10,000 x 100 = ~2 million numbers. Huge savings. </p><p>BitNet already makes the base model tiny and fast on phones (weights are just -1/0/+1 instead of full floating point numbers which already saves ~70-90% of memory &amp; compute).</p><p>But fine-tuning BitNet normally was still hard/impossible on phones because even &#8220;updates&#8221; were too heavy. </p><blockquote><p><em><strong>LoRA slashes the update size so dramatically that now the whole personalization process fits in a phone&#8217;s limited RAM and memory.</strong></em></p></blockquote><p>Think of your phone&#8217;s AI as a very compressed zip file of a genius librarian (BitNet makes the zip super small so it fits on phone.)</p><p>LoRA = keep the zipped library untouched, just add a tiny bookmark file with your personal notes/index. Your phone can handle adding/using the bookmark easily, and when you ask questions, the librarian reads the main zip + your bookmark instantly. </p><p><strong>It turns AI from &#8220;rented cloud services that sees all your data&#8221; into &#8220;your own private brain extension that lives on your device and learns only from you.</strong></p><p>No more paying $20/month to OpenAI, no data sent anywhere, no censorship, works offline. </p><p>Tether/QVAC solved making LoRA + BitNet work cross-platform (AMD/Intel/Nvidia/Apple/Mobile GPUs) for the first time. </p><blockquote><p><strong>LoRA is the unlock button that lets billion parameter personalization escape data centers and live in your pocket. That&#8217;s why Paolo and the team are calling it the start of &#8220;Stable intelligence&#8221;. If this clicks now, the rest of the announcement will feel much more exciting and logical. </strong></p></blockquote><p>Now that you understand LoRA, the post should no longer feel &#8220;cool tech jargon + big numbers&#8221; shill. Impressive but abstract! </p><p>Now that you grasp LoRA as the &#8220;tiny post&#8217;it notes&#8221; trick that freezes the huge core model and only trains a minisule add-on layer (up to 99% frwer paramters to update), the whole thing snaps into place and becomes genuinely revolutionary.</p><p>The &#8220;billion-parameter training on a phone&#8221; claim stops sounding impossible.</p><p>Fine-tuning a 13B model normally means updating billions of numbers which needs massive VRAM and power that no phone has.</p><p>But with LoRA, you&#8217;re only training maybe millions (or fewer) of tiny adapter parameters. Add BitNet&#8217;s extreme compression (weights as simple as -1/0/+1), and suddenly that tiny update process fits in a phone&#8217;s limited RAM/battery.</p><blockquote><p><strong>The iPhone demo (13B fine-tine), and Samsung/Pixel 3.8B ones aren&#8217;t hype, they&#8217;re the direct result of LoRA slashing the workload by orders of magnitude.</strong></p></blockquote><p>The &#8220;biggest unlock&#8221; line hits hard. Paolo calls heterogeneous GPU fine-tuning the biggest unlock because LoRA + BitNet now works everywhere (not just Nvidia CUDA). Before, even if BitNet existed, personalizing it required expensive Nvidia rigs or cloud. Now your AMD laptop, Intel PC, Apple Silicon Mac, or flagship phone can do it.</p><p>LoRA is the key that democratizes that persaonlization step azcross hardware - no more &#8220;sorry, only works on $10K GPU&#8221;.</p><p>Privacy and &#8220;serve the people&#8221; vision feels tangible, not fluffy.</p><p>Without LoRA, &#8220;local private AI&#8221; would mean running a generic frozen model (no real personalization) or shipping your data to the cloud for fine-tuning.</p><p><strong>LoRA lets the model learn deeply from your emails/docs/health data entirely on-device, with almost no extra cost. Your AI becomes truly &#8220;yours&#8221; - customized to your life, offline, private forever.</strong></p><p>That&#8217;s why Paolo frames it as the &#8220;first real-world signal of a local private AI that can truly serve the people.&#8221;</p><p>&#8220;Era of Stable Intelligence&#8221; lands as a real shift.</p><blockquote><p><strong>Tether AI is the stablecoin moment for decentralized artificial Intelligence</strong></p></blockquote><p>&#8220;Stable&#8221; here echoes stablecoins: reliable, decentralized, user-controlled, and no central gatekeepers. LoRA + BitNet + cross-platform Fabric =  the technical foundation to escape cloud dependency.  What used to be locked in data centers (personalized frontier AI) now &#8220;escapes&#8221; to your pocket.</p><p>It&#8217;s not just faster/cheaper but it&#8217;s a philosophical unlock toward AI sovereignty.</p><p>In short, LoRA isn&#8217;t just a technique, it&#8217;s the unlock button that turns theoretical edge AI into practical reality today.</p><p>The announcement isn&#8217;t about raw model size anymore. It is about who controls the intelligence (you, on your device) instead of renting it from Big Tech.</p><p>With this lens, the post reads like a declaration of independence for personal AI, and the demos prove it&#8217;s not vaporware.</p><p>Super exciting if you&#8217;re into decentralization, privacy, or just hating monthly API bills. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Wattage Waterfall of Compute]]></title><description><![CDATA[A single GB200 NVL72 rack draws ~120kW. A standard office building draws ~50-100kW. One AI training rack consumes more power than the entire building you're sitting in.]]></description><link>https://www.rochlabs.com/p/the-wattage-waterfall-of-compute</link><guid isPermaLink="false">https://www.rochlabs.com/p/the-wattage-waterfall-of-compute</guid><dc:creator><![CDATA[Rohit Chauhan]]></dc:creator><pubDate>Thu, 05 Mar 2026 05:56:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bdfc3279-7fc8-4d2c-a660-e4ae371bf620_2752x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In <a href="https://www.rochlabs.com/p/the-physics-behind-the-ai-trade">Part 1</a> we established that every logic operation costs energy and generates heat, a constraint thermodynamics imposes permanently.</p><p>Today we follow that heat all the way up the stack; from a single chip to a full-scale data center running AI workloads around the clock.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The Wattage Waterfall</h1><p>A single logic operation at the level of a chip generates heat that demands cooling. A cluster of ten thousand chips demand a power grid. A training run demands that the grid run without interruption for weeks, and once a model deploys, inference demand runs the grid forever.</p><h2>The Chip: The micro-unit of compute</h2><p>A single H100 draws 700W of power at peak. One H100 server (8 GPUs) draws 10.2kW of total system power. A single GB200 NVL72 rack - Nvidia&#8217;s current Blackwell configuration draws ~120kW. A standard office building draws roughly 50-100 kW which means one AI training rack consumes more power than an entire office building. (1), (2), (3)</p><h2>The Cluster: Where scale meets physics</h2><p>A frontier model training cluster runs 10k to 25k+ GPUs simultaneously. At 700W per H100, a 10k GPU cluster draws 7MW continuously. xAI&#8217;s colossus cluster in Memphis is ~100,000 H100s drawing ~150-200 MW continuously. (1), (4), (5)</p><h2>The training run level: The wattage it takes to bring GPT-4 to life</h2><p>GPT-4 is estimated to consume ~50 GWh of power for training. This is equivalent to powering ~4,800 average US homes for a full year. For xAI&#8217;s Grok 4, estimates show ~310 GWh of power consumed for training, a 5-6x increase in magnitude over GPT-4. As models become bigger, the training compute will continue to expand, and Epoch AI estimates training compute to scale ~4x per year and by extension the energy cost of training scales at roughly the same rate. (5), (6), (7), (8)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6kbR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6kbR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 424w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 848w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 1272w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6kbR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png" width="1420" height="922" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:922,&quot;width&quot;:1420,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6kbR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 424w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 848w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 1272w, https://substackcdn.com/image/fetch/$s_!6kbR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff577eb4c-ee3b-43ae-8965-5a4bd9b3fb3a_1420x922.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The inference level: Where energy demand becomes permanent</h2><p>OpenAI CEO Sam Altman disclosed in a June 2025 blog post that a single ChatGPT query uses ~0.34 watt-hours. Data compiled by Epoch AI independently corroborated Altman&#8217;s own estimates and calculates ~0.3 Wh per query. Altman equated this to an oven in usage for a little over one second, or a high-efficiency lightbulb in use for a couple of minutes in his blog. (6), (9), (10)</p><p>At 2.5B daily queries (or prompts) reported by TechCrunch citing OpenAI in July 2025 allows us to infer that the daily energy demand from inference alone costs 0.34 x 2.5B watt hours of electricity. In other words, ~850 MWh per day, which works to an annual run rate of 850 MWh x 365, ~310 GWh per year. (11)</p><p>As noted above, GPT-4 model training consumed ~50 GWh of electricity. A simple division shows that inference would consume roughly the equivalent of the electricity consumed in the overall training process in ~59 days (50,000 MWh &#247; 850 MWh/day = ~59 days). MIT further states that Inference&#8217;s share of the total AI lifecycle energy is ~80-90%, and reasoning models (o1/o3-class models) only escalate this consumption as data from Epoch AI shows 2.5x more tokens consumed by reasoning models over regular models. (6), (10)</p><h2>The Data Center Level: Where it all aggregates</h2><p>IEA estimates global data center energy consumption to double from ~415 TWh (c.1.5% of global electricity) to 945 TWh as the base case between 2024 and 2030. For context, this is equivalent to the entire current annual electricity consumption of Japan. A Goldman Sachs report further adds that data centers will command 8% of the total US power consumption by 2030 against ~3% in 2022. This massive demand for power explains the approx. $720B grid spending estimates by 2030 by Goldman Sachs, and an estimated ~$736B capex announcements by the top 5 hyperscalers in 2025-26 alone as outlined above. (12), (13)</p><h2>Conclusion</h2><p>The wattage waterfall makes the constraints of physics (thermodynamics) structural requirements. This brings us to the three primary conclusions relevant for all investors trying to build a foundational understanding of the AI trade. </p><p>First, power must be generated. Second, that power must be effectively transmitted across the racks, and third, these AI workloads will generate significant heat which must be removed from the environment. </p><p>These three requirements not only mandate large upfront investments in the build-out for power, grid, and cooling infrastructure; but also force them as structural, not cyclical requirements irrespective of who wins the model race. </p><div><hr></div><p><strong>Now that we&#8217;ve explained the constraints of thermodynamics in Part 1, and the current energy &amp; cooling requirements for SOTA chips powering massive AI workloads, the next question is who builds it, who powers it, and who profits from it. </strong></p><p><strong>In Part 3, I'll map the exact companies whose business models exist to solve the physical constraints outlined above. The investment case for each one is anchored in physical laws that don&#8217;t change regardless of who wins the model race. Stay tuned!</strong></p><p>If you found this useful, <a href="https://www.rochlabs.com/?r=7bpf4y&amp;utm_campaign=subscribe-page-share-screen&amp;utm_medium=web">subscribe</a>. Part 3 drops next week.</p><h1>Citations</h1><ol><li><p>NVIDIA Corporation,<a href="https://www.nvidia.com/en-in/data-center/h100/"> "NVIDIA H100 GPU &#8212; Product Specifications,"</a> accessed February 2026.</p></li><li><p>NVIDIA Corporation,<a href="https://openzeka.com/en/wp-content/uploads/2022/04/ai-for-enterprise-dgx-h100-datasheet-nvidia-a4-2146027-r3-web.pdf"> "NVIDIA DGX H100 Datasheet,"</a> March 2022. Document number A4-2146027-R3.</p></li><li><p>NVIDIA Corporation,<a href="https://www.nvidia.com/en-in/data-center/gb200-nvl72/"> "DGX GB200 NVL72 Datasheet,"</a> NVIDIA Documentation, accessed March 2026.</p></li><li><p>Pilz, K., Rahman, R., Sanders, J. and Heim, L.,<a href="https://epoch.ai/data-insights/training-cluster-size"> "AI Training Cluster Sizes Increased by More Than 20x Since 2016,"</a> Epoch AI, October 23, 2024.</p></li><li><p>James Sanders, Luke Emberson and Yafah Edelman,<a href="https://epoch.ai/data-insights/grok-4-training-resources"> "What did it take to train Grok 4?"</a> Epoch AI, September 12, 2025.</p></li><li><p>James O'Donnell and Casey Crownhart,<a href="https://www.technologyreview.com/2025/05/20/1116327/ai-energy-usage-climate-footprint-big-tech/"> "We did the math on AI's energy footprint,"</a> MIT Technology Review, May 20, 2025.</p></li><li><p>U.S. Energy Information Administration,<a href="https://www.eia.gov/energyexplained/use-of-energy/electricity-use-in-homes.php"> "Electricity Use in Homes,"</a> updated December 18, 2023. Derived estimate: 50 GWh &#247; 10,500 kWh/household/year = ~4,762 households. Rounded to ~4,800 in text.</p></li><li><p>Sevilla, J. and Rold&#225;n, E.,<a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year"> "Training Compute of Frontier AI Models Grows by 4-5x Per Year,"</a> Epoch AI, May 28, 2024.</p></li><li><p>Sam Altman,<a href="https://blog.samaltman.com/the-gentle-singularity"> &#8220;The Gentle Singularity,&#8221;</a> blog.samaltman.com, June 11, 2025.</p></li><li><p>You, J.,<a href="https://epoch.ai/gradient-updates/how-much-energy-does-chatgpt-use"> "How much energy does ChatGPT use?"</a> Epoch AI, February 7, 2025.</p></li><li><p>Dastin, J.,<a href="https://techcrunch.com/2025/07/21/chatgpt-users-send-2-5-billion-prompts-a-day/"> &#8220;ChatGPT users send 2.5 billion prompts a day,&#8221;</a> TechCrunch, July 21, 2025.</p></li><li><p>International Energy Agency, &#8220;Energy and AI &#8212; Energy Demand from AI,&#8221; IEA, April 10, 2025.<a href="https://www.iea.org/reports/energy-and-ai/energy-demand-from-ai"> https://www.iea.org/reports/energy-and-ai/energy-demand-from-ai</a></p></li><li><p>Goldman Sachs Research, "<a href="https://www.goldmansachs.com/insights/articles/ai-to-drive-165-increase-in-data-center-power-demand-by-2030">AI to drive 165% increase in data center power demand by 2030</a>," Feb 4, 2025.</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Physics Behind the AI Trade]]></title><description><![CDATA[Why Thermodynamics, Not Narratives, Decides Where Value Accrues in the AI Trade]]></description><link>https://www.rochlabs.com/p/the-physics-behind-the-ai-trade</link><guid isPermaLink="false">https://www.rochlabs.com/p/the-physics-behind-the-ai-trade</guid><dc:creator><![CDATA[Rohit Chauhan]]></dc:creator><pubDate>Fri, 27 Feb 2026 15:56:01 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/388e93a2-2115-40bc-ba4a-b5ed50ef8478_2752x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;02849e7b-a1b8-43c5-8edf-740a6642fe98&quot;,&quot;duration&quot;:null}"></div><h1>A Physics first lens to playing the AI trade</h1><p>Physics is the baseline for the natural world, and also the absolute ceiling for any problem we solve (including artificial intelligence.) Everything above the &#8216;physics layer&#8217; are human choices that can theoretically be changed. Confusing human choices for physics laws is what most get wrong.</p><p>The analysis that follows is grounded in the constraints and opportunities physics imposes. Hard science, not media narratives, will decide how to find the winning opportunities in the AI trade.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>The physics that follows exists to serve the investment thesis, not academic completeness. The concepts most applicable are chosen to highlight the constraints, build the analytical foundation, and establish a new lens to evaluate the AI trade.</p><p>This framework also serves as a filter to fact-check every tall claim about AI disruption. physics is the reality check that prevents costly mistakes born of following the crowd rather than grounding analysis in reality.</p><h1>The Physics behind the AI trade</h1><p>At its core, the entire AI lifecycle from inception to end is governed by five primary physical fundamentals. Each manifests at various layers of the AI assembly line.</p><ol><li><p>The law of Thermodynamics (Landauer&#8217;s Limit and Heat Dissipation)</p></li><li><p>The law of speed of light</p></li><li><p>The principles of Quantum Mechanics</p></li><li><p>The laws of information theory (Shannon&#8217;s theorem)</p></li><li><p>The law of Wave physics (Diffraction Limit)</p></li></ol><p>Thermodynamics is already the dominant cost driver in AI infrastructure. The energy and cooling buildout it demands is leading capital allocation decisions across every stakeholder in the race to dominate AI. (1)</p><p>Data compiled by Goldman Sachs shows the five largest hyperscalers in the US will deploy a cumulative $736B in Capex for the AI buildout in 2025 and 2026 alone. Going further, McKinsey has projected a total investment of ~$5.2 trillion for AI data centres by 2030. (2)</p><p>Of that $5.2 trillion, McKinsey estimates $1.3 trillion flows directly into the energisers; the utilities, cooling system manufacturers, and electrical infrastructure providers whose entire business exists to solve the heat and power problem thermodynamics creates. This is the thermodynamics trade expressed in dollars.</p><p>The IEA projects the total demand for energy from data centres to double from ~415 TWh in 2024 to ~945 TWh by 2030, this is the equivalent to adding the entire electricity consumption of Japan in 2024 to the global grid in just six years. These figures establish the scale of what thermodynamics demands in capital terms, and the physics below explains why this demand is permanent. (3, 4)</p><h2>Thermodynamics imposes physical energy constraints on chip compute capabilities</h2><p>If you strip away AI to the precise point where physics shows up, it boils down to a single factor: transistors switching states. Billions of times per second, per chip. Those switching events are what execute the matrix multiplications underlying every model, and every switching event costs energy and generates heat.</p><p>Every compute operation is subject to the constraints imposed by the second law of thermodynamics. A constraint neither engineering nor software can solve. This is nature imposing its will on every transistor switch on every chip at every data centre ever built by humanity.</p><p>The physical law creates two distinct constraints on compute, one theoretical, and one practical. Understanding both builds the foundation for our thesis that the energy infrastructure trade isn&#8217;t cyclical capex but a permanent structural requirement.</p><h3>Constraint 1: Landauer&#8217;s Limit - The Theoretical Floor of Compute</h3><p>First introduced in 1961, the Landauer&#8217;s limit provides the baseline of the minimum quantity of heat dissipated in the environment to erase a single bit of information. At room temperature (~300K), the minimum heat released is 2.8 &#215; 10&#8315;&#178;&#185; joules per bit operation. This is thermodynamics enforcing absolute limits no material, architecture, optimization or engineering can go beyond. This is the theoretical floor of all compute, including AI.</p><p>Peer reviewed literature on this topic argues that modern chips release roughly a million times more energy per logic operation which exceeds the Landauer&#8217;s limit by many orders of magnitude. (5)</p><p>While this sets the physics enforced ceiling on compute per watt, a better lens to understand this is backed by solid research published by Ho, Erdil &amp; Besiroglu (2023) which ignores the Landauer floor and instead tackles the more important and immediate question, &#8220;how much more efficient can silicon transistors realistically get before hitting their own engineering limits?&#8221; (6)</p><p>The current state of the art CMOS architectures, the transistor-based chip designs running today&#8217;s AI GPUs are operating ~207x above the maximum efficiency Ho et al&#8217;s research outlines. In other words, there is capacity to improve chip efficiency to the order of 207x before we hit the practical limits of how much energy per logic operation these chips will consume.</p><p>While this may sound a lot, current AI compute demand is scaling 4-5x per year, and all the efficiency gains are consumed by demand growth before they can actually reduce energy infrastructure requirements. This explains the massive grid capex and hyperscaler expansion efforts in line with the estimates and data compiled by McKinsey and Goldman Sachs outlined above. (7)</p><h3>Constraint 2: Heat Dissipation - The practical constraint of compute</h3><p>Landauer&#8217;s principle is evidence of physics imposing its will on compute. Every switch event is generating heat as a consequence of executing logic operations within the constraints of physics.</p><p>Take the Nvidia H100 for instance, which draws ~700W per chip. The next-gen GB200 NVL72 rack draws nearly 120kW for 72 Blackwell GPUs, in comparison, a single H100 rack consumes 10-30kW. This is 4-12x increase in rack power density in a single generation of GPU offerings. This increase across generations isn&#8217;t a one-time thing either, mapping the energy consumption across four generations of chips starting with V100 (~300W in 2017), A100 (~400W in 2020), H100 (~700W in 2022), and the latest generation of GB200 (~1,000W in 2024) shows how each generation demands meaningfully more power and cooling than the last. (8)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YI-k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YI-k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 424w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 848w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 1272w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YI-k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png" width="1200" height="520" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:520,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:43248,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rochlabs.com/i/189370836?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YI-k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 424w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 848w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 1272w, https://substackcdn.com/image/fetch/$s_!YI-k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58707a95-a322-45fe-a9e5-189c4ef8b136_1200x520.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Note: Citations 9-12 links to the original source material for the table data above.</em></p><p>At these rates of power consumption, liquid cooling isn&#8217;t just optional infrastructure, it is mandatory, and scales permanently with every watt deployed.</p><p>To sum up, Landauer&#8217;s limit establishes the physical cost of compute which forces generation and consumption of electricity, and heat dissipation mandates cooling infrastructure as heat scales permanently with compute. These aren&#8217;t cyclical capex driven by a temporary infrastructure gap. They are structural requirements imposed by physics.</p><p><strong>Thermodynamics has already decided what the AI infrastructure stack must look like. The next question is who builds it, who powers it, and who profits from it. In Part 2, I'll map the exact companies whose business models exist to solve the physical constraints outlined above. The investment case for each one is anchored in physical laws that don&#8217;t change regardless of who wins the model race. Stay tuned!</strong></p><p>If you found this useful, <a href="https://www.rochlabs.com/?r=7bpf4y&amp;utm_campaign=subscribe-page-share-screen&amp;utm_medium=web">subscribe</a> below. Part 2 drops next week.</p><div><hr></div><h1>Citations</h1><ol><li><p>Goldman Sachs Research, "<a href="https://www.goldmansachs.com/insights/articles/how-ai-is-transforming-data-centers-and-ramping-up-power-demand">How AI Is Transforming Data Centers and Ramping Up Power Demand</a>," July 16, 2024.</p></li><li><p>Jesse Noffsinger et al., "<a href="https://www.mckinsey.com/industries/technology-media-and-telecommunications/our-insights/the-cost-of-compute-a-7-trillion-dollar-race-to-scale-data-centers">The cost of compute: A $7 trillion race to scale data centers</a>," McKinsey Quarterly, April 2025.</p></li><li><p>International Energy Agency, "<a href="https://www.iea.org/reports/energy-and-ai/energy-demand-from-ai">Energy and AI &#8212; Energy Demand from AI</a>," April 2025.</p></li><li><p>Enerdata via Statista, "<a href="https://www.statista.com/statistics/267081/electricity-consumption-in-selected-countries-worldwide/">Electricity consumption worldwide in 2024, by leading country</a>," July 2025.</p></li><li><p>Chattopadhyay, P., Misra, A., Pandit, T., and Paul, G.,<a href="https://arxiv.org/html/2506.10876v1"> "Landauer Principle and Thermodynamics of Computation,"</a>, June 2025.</p></li><li><p>Ho, A., Erdil, E., and Besiroglu, T.,<a href="https://arxiv.org/pdf/2312.08595"> "Limits to the Energy Efficiency of CMOS Microprocessors,"</a> <em>2023 IEEE International Conference on Rebooting Computing</em>, December 2023. . H100 today achieves ~1.4&#215;10&#185;&#178; FP16 FLOP/J. The estimated maximum CMOS efficiency ceiling is ~2.9&#215;10&#185;&#8308; FP16/J. That gap is approximately 207x.</p></li><li><p>Sevilla, J. and Rold&#225;n, E.,<a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year"> "Training Compute of Frontier AI Models Grows by 4-5x Per Year,"</a> Epoch AI, May 28, 2024.</p></li><li><p>NVIDIA Corporation,<a href="https://docs.nvidia.com/dgx/dgxgb200-user-guide/hardware.html"> "DGX GB200 User Guide &#8212; Hardware,"</a> NVIDIA Documentation, August 2025.</p></li><li><p>NVIDIA Corporation,<a href="https://images.nvidia.com/content/technologies/volta/pdf/volta-v100-datasheet-update-us-1165301-r5.pdf"> &#8220;NVIDIA V100 Tensor Core GPU Datasheet,&#8221;</a> January 2020.</p></li><li><p>NVIDIA Corporation,<a href="https://www.nvidia.com/content/dam/en-zz/Solutions/Data-Center/a100/pdf/nvidia-a100-datasheet-nvidia-us-2188504-web.pdf"> "NVIDIA A100 Tensor Core GPU Datasheet,"</a> May 2022.</p></li><li><p>NVIDIA Corporation,<a href="https://www.nvidia.com/en-in/data-center/h100/"> "NVIDIA H100 GPU &#8212; Product Specifications,"</a> accessed February 2026.</p></li><li><p>Lenovo, "ThinkSystem NVIDIA HGX B200 180GB 1000W GPU Product Guide," Lenovo Press, LP2226, January 13, 2026. <a href="https://lenovopress.lenovo.com/LP2226">https://lenovopress.lenovo.com/LP2226</a> &#8212; citing NVIDIA GPU specifications, Table 2.</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rochlabs.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading ROCH Labs! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>