<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Ayaz on Tech & AI]]></title><description><![CDATA[Rambling on about Tech, AI, Culture, and everything in between. ]]></description><link>https://substack.ayaz.pk</link><image><url>https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png</url><title>Ayaz on Tech &amp; AI</title><link>https://substack.ayaz.pk</link></image><generator>Substack</generator><lastBuildDate>Wed, 08 Apr 2026 15:58:19 GMT</lastBuildDate><atom:link href="https://substack.ayaz.pk/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ayaz Ahmed Khan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[ayazak@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[ayazak@substack.com]]></itunes:email><itunes:name><![CDATA[Ayaz]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ayaz]]></itunes:author><googleplay:owner><![CDATA[ayazak@substack.com]]></googleplay:owner><googleplay:email><![CDATA[ayazak@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ayaz]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[A few take-aways from GTC 2026 Keynote]]></title><description><![CDATA[What I think was significant in Jensen's keynote address at GTC 2026]]></description><link>https://substack.ayaz.pk/p/a-few-take-aways-from-gtc-2026-keynote</link><guid isPermaLink="false">https://substack.ayaz.pk/p/a-few-take-aways-from-gtc-2026-keynote</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Mon, 23 Mar 2026 09:11:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I saw Jensen&#8217;s GTC 2026 keynote with forlorn mood. I had thought GTC 2026 was the conference to personally be at this year. Sadly, it just didn&#8217;t happen for me. Nevertheless, I thought I&#8217;d talk a little about the things that piqued my interest in his keynote. </p><p>I think the biggest surprise from Jensen Huang&#8217;s GTC 2026 keynote was not that Nvidia bought Groq earlier. It was <em>how</em> Nvidia chose to use Groq. A lot of people assumed the Groq move was defensive. Buy the company, remove a future threat, stop another hyperscaler from acquiring it, and take the engineering prowess and forget about the product. What I don&#8217;t think many expected was Jensen openly positioning Groq LPUs <em>alongside</em> Nvidia GPUs inside the Vera Rubin platform itself. That is a much more interesting move.</p><blockquote><p>I did not-so-technical deep-dives into both <a href="https://substack.ayaz.pk/p/whats-inside-the-nvidia-vera-rubin">Vera Rubin NVL72 platform</a> as well as <a href="https://substack.ayaz.pk/p/how-do-language-processing-units">Groq LPUs</a>. They make for a helpful read to better understand the technical nuances of what&#8217;s going on. </p></blockquote><p>The architecture he outlined for Vera Rubin NVL72 makes the logic pretty clear. Prefill still runs on Rubin GPUs. Decode, however, gets split. The attention part of decode stays on GPUs, while the Multi-Layer Perceptron (MLP) or Feed-Forward Network (FFN) part moves to Groq LPUs. </p><p>In simple terms, this looks like a memory hierarchy decision hiding as an inference architecture. Attention is where the system keeps touching the KV cache, and KV cache wants HBM. That favours GPUs. MLP execution, on the other hand, is much more about repeatedly applying weights as fast as possible, and that favours Groq&#8217;s SRAM-heavy design. So the split is elegant: keep KV-cache close to HBM-heavy GPUs, keep weights close to SRAM-heavy LPUs, and let each device do the part of the decode loop it is better at. That is a much more nuanced story than &#8220;GPU versus LPU.&#8221; It is really about matching memory behaviour to the right silicon.</p><p>My second takeaway is that Samsung&#8217;s role here is bigger than it may look at first glance. Jensen explicitly thanked Samsung for manufacturing the Groq LP30 chip and said the chips were already in production, with shipment planned for the second half of 2026. Reuters also reported Samsung is making the chip on its 4nm process. That matters because it shows Samsung is not just orbiting this stack as a memory supplier anymore. It is now part of the logic side of Nvidia&#8217;s inference strategy too. That is a serious commitment.</p><p>The last point that stayed with me was Jensen talking about CPUs as a future multi-billion dollar business. Some time back, it started to look like AI was pulling compute value away from CPUs and toward GPUs. People in the know called in one of the three underlying paradigm shifts brought about by AI. Training moved to accelerators. Inference moved to accelerators. The CPU looked increasingly like plumbing.</p><p>But the agentic era complicates that story. GPUs are incredible at parallel compute. But if they are starved of data, orchestration, tool results, memory lookups, sandboxing, scheduling, and environment management, they sit idle. And a lot of agentic engineering lives exactly there: in the harness around the model. The model call may hit the GPU, but the workflow around it is full of CPU-bound work. Tool execution, state management, retrieval glue, memory handling, routing, queues, session control, and sandboxed environments all lean heavily on the CPU side.</p><p>That makes Jensen&#8217;s Vera CPU strategy easier to understand. He is not arguing that CPUs beat GPUs for model compute. He is arguing that agentic systems create a much larger surrounding control plane, and that control plane needs very fast, very efficient CPUs with strong single-threaded performance and high data output. Nvidia is clearly trying to own that layer too. Jensen has already said he expects Vera to become a multi-billion dollar CPU business. In other words, the shift is not from CPUs to GPUs. It is from general-purpose compute to a tightly orchestrated CPU-GPU-LPU stack, where each part becomes more important as agents get more complex. </p><p>My take is that the real theme of this keynote was not just more AI compute. It was <em>heterogeneous inference, designed around bottlenecks<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></em>. GPUs are still central. But Nvidia is now openly acknowledging that one kind of silicon is not enough for where inference is going. HBM-heavy GPUs, SRAM-heavy LPUs, and agent-optimized CPUs each solve a different constraint. What changes is that Nvidia is no longer just selling the fastest accelerator. It is trying to own the entire serving path.</p><blockquote><p>One final point I want to talk about is the near future importance and importance of Optics in chip design and interoperability. Jensen hinted at it. I am going to dedicated a separate post to talk about it. </p></blockquote><p>Thank you for reading!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><a href="https://x.com/jukan05/status/2035581499597275615">Heterogenous inference architecture</a></p></div></div>]]></content:encoded></item><item><title><![CDATA[Is recursive self improvement closer than we think?]]></title><description><![CDATA[How OpenAI's Codex 5.3 and MiniMax's M2.7 built their own successors.]]></description><link>https://substack.ayaz.pk/p/is-recursive-self-improvement-closer</link><guid isPermaLink="false">https://substack.ayaz.pk/p/is-recursive-self-improvement-closer</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Thu, 19 Mar 2026 15:45:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In my earlier posts, I theorized how <em>continual self learning</em> for frontier models is likely the only thing standing in the path to AGI. Some very smart people recently resigned from important positions at various leading AI labs and made claims about recursive self improvement being only <em>six to twelve months</em> away. In his latest podcast<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> with Peter Diamandis, Elon Musk also hinted at recursive self improvement being just around the corner. But reading them briefly talk about this, without providing any evidence to support their claims, makes you wonder whether the timeline is accurate. </p><p>Earlier this year Codex 5.3 from OpenAI was released. Along with Sonnet/Opus 4.5, Codex 5.3 ushered in a new era for SOTA models. Those who understand it know what changed. These models are now starting to become what I call omnipotent models &#8212; they are eating everything up. The reason why I mention Codex 5.3 in particular is because OpenAI publicly acknowledged, for the very first time, that Codex played an <em>instrumental</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> role in building its next version. Before that, models were routinely used, as part of training newer models, to generate synthetic data and AI-generated feedback as part of the reinforcement learning from AI feedback (RLAIF) loop. But what Codex 5.3 achieved was altogether different. It helped in debugging training runs, managing deployment of the model, analyzing test results, writing and performing evals, and much more. This is an entirely different level of model involvement in creating its successor. It hints at how close we are to fully recursive self improving models. </p><p>Yesterday, MiniMax published an article<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> about their latest SOTA-level model titled &#8220;<em>Early echoes of self-evolution</em>&#8221;. The title alone sent a chill down my spine when I read it. Buried within a series of spectacular results across different benchmarks for M2.7 (their latest model) is the public acknowledgment of the model&#8217;s self-evolution that should shake to the core anyone who tracks continual learning. Among other things, they explain that they found the model was able to update its own memory, build dozens of complex skills in its harness, improve its learning process based on reinforcement learning experiments, iterate multiple times over its own architecture, skills and memory, and autonomously run repeated cycles to optimize itself and evaluate whether it&#8217;s building a better version of itself. They call it a <em>cycle of model self-evolution</em>. I call it the model&#8217;s ability to evolve itself to the point that it can build a better, more capable, more efficient version of itself, almost autonomously. If that&#8217;s not the early signs of recursive self improvement, I don&#8217;t know what is. </p><p>When they said recursive self improvement was only six to twelve months out, they weren&#8217;t exaggerating. These are the kind of breakthroughs that are as exciting as they are dangerous. They carry within themselves the echoes of the fictional Skynet scenario we all fear. Yet, it&#8217;s no time to stop. </p><p>Thanks for reading. </p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>https://www.mexc.com/news/912121?ref=aisecret.us</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>https://thenewstack.io/openais-gpt-5-3-codex-helped-build-itself/</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>https://www.minimax.io/news/minimax-m27-en</p></div></div>]]></content:encoded></item><item><title><![CDATA[The Repo Frontier]]></title><description><![CDATA[An alternative, hypothetical look at the Middle East conflict.]]></description><link>https://substack.ayaz.pk/p/the-repo-frontier</link><guid isPermaLink="false">https://substack.ayaz.pk/p/the-repo-frontier</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Tue, 17 Mar 2026 07:12:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I have been meaning to describe a different way of looking at the on-going conflict in Middle East which some smart analysts have been pointing at. This way of looking changes the discourse from looking at the conflict as everything to do with religion, ancient rivalries and the fight against wrong and evil. Of course, this is my hypothesis based on the work done by others, so I&#8217;d like you to view it just as that and nothing more. </p><p>The core idea in this framing is that there are two deep-states fighting each other for survival. Both of them are deeply bureaucratic-financial complexes, or what you call the Financial Industrial Complex. The worrying part though in their conflict is that they are pushing it beyond the periphery of sanctions and covert economical pressure into more overt, kinetic action: that is, direct and in-direct war. </p><p>Here is how I like to think of the two deep-states: </p><ul><li><p>The <strong>old state</strong>, which is a central banking era power of old. It&#8217;s comfortable with opacity, intermediaries, and with off-ledger financial plumbing, that is, moving money through jusridictional arbitrage, gray corridors, and the quiet dependency of large institutions (banks) on those corridors. </p></li><li><p>The <strong>new state</strong>, which is ledger-first, wants everything digitized, permissioned, and legible. It prefers chokepoints that are software-defined. It is less tolerant of shadow systems because they are unaudtiable. </p></li></ul><p>One of them is optimized for discretion, while the other for compliance by (digital) architecture. If you buy that, the Middle East stops looking like a single conflict. It starts looking like a set of bottlenecks in a global financial migration.</p><p>In this hypothesis, Iran isn&#8217;t just a regional actor. It functions as a node in one of the oldest shadow financial systems, routing through regional banks and corridors that are technically deniable and politically convenient. The hub in London, and the off-shore battery in Iran, that is, while most European and Britsh banks are quietly, happily complicit in the setup. </p><p>Iran&#8217;s position, relationships, and tolerances make it useful for legacy European finance markets that thrive on deep liquidity and flexible collateral flows.</p><p>If you&#8217;ve ever looked at modern finance long enough, you realize the real product isn&#8217;t money. It&#8217;s settlement. It&#8217;s repo -- the short term secure loaning facility that London is known for. It&#8217;s the ability to move claims around with speed, credibility, and just enough ambiguity. And when ambiguity is a feature, not a bug, certain corridors become strategically valuable.</p><p>Now consider this hypothesis: the new state made a systematic attempt to remove alternative protection structures and force renegotiation with long-standing holdouts. This includes actions around cartels, Venezuela, and Cuba (after 60 years of defiance, the country is back on negotiation table). These were not isolated policy quirks, but a series of coordinated actions to remove informal enforcement and parallel markets, reduce the number of places capital can hide, launder, or sit outside the preferred rails, and finally to push smaller nodes back into a negotiable box.</p><p>And that is where Iran comes in. It&#8217;s the last man standing in the way, the old holdout in the shadow financial system. It is a living interface to the legacy shadow routing system which European banks quietly benefit from. If you look at Iran from that lens, then it becomes more than a geopolitical adversary. It becomes a compatibility layer for an older financial order. And that is exactly what a ledger-first regime wants to delete. </p><p>If this thought process is right, then escalation isn&#8217;t about a single provocation. It&#8217;s about migration pressure. When you try to move a huge system from trust-me networks to verifiable ledgers, discretionary intermediaries to audited rails, and tolerated ambiguity to enforced legibility, you don&#8217;t get polite consensus. You get counter-pressure. You get to corner a tiger without providing a way out, and it does best what it is supposed it: it lashes back hard. </p><p>Some pressure is financial (sanctions, seizures, blacklists). Some is narrative (moral framing, legitimacy campaigns). And sometimes, when neither side yields, it becomes physical. Kinetic action, in this lens, is what happens when the underlying settlement architecture is being contested and the bargaining fails.</p><p>Now, my takeaway is that part of what we are calling the Middle East conflict is really just a fight over which financial operating system gets to run the next era. If that&#8217;s indeed what is happening, then watching missile movements without watching settlement plumbing is like watching GPUs without watching memory bandwidth.</p><p>Thanks for reading! And please, keep in mind, that&#8217;s all <em>hypothetical</em>. </p><div><hr></div><p><em>References</em>:</p><p><em><a href="https://x.com/EXIT_FIAT/status/2032577546764833239?s=20">Part 1: The 118-Year Pattern They Don't Want You to See</a></em></p><p><em><a href="https://x.com/EXIT_FIAT/status/2032939960177959083?s=20">Part 2: The $17.7 Billion Paper Trail</a></em></p><p><em><a href="https://x.com/EXIT_FIAT/status/2033302343819890725">Part 3: The Final Piece &#8212; The Declaration Nobody</a></em></p><p></p>]]></content:encoded></item><item><title><![CDATA[Second-Order Effects: How Hormuz Threatens AI Infrastructure ]]></title><description><![CDATA[A look at the second and third-order effects of the closure of the Strait of Hormuz has on AI infrastructure buildout.]]></description><link>https://substack.ayaz.pk/p/second-order-effects-how-hormuz-threatens</link><guid isPermaLink="false">https://substack.ayaz.pk/p/second-order-effects-how-hormuz-threatens</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:36:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The Strait of Hormuz closed on March 2, 2026. What followed wasn&#8217;t just an oil crisis. It was a series of bottlenecks that revealed how deeply AI infrastructure depends on a handful of chokepoints most people weren&#8217;t paying attention to. Aren&#8217;t paying attention to. </p><p>I have been meaning to map out the second and third-order effects of this conflict on AI buildout. The first-order effects are obvious: oil prices hit $110 per barrel, tanker traffic dropped to near zero, and energy costs spiked. But the real story is in what happens in the materials, chemicals, and gases that semiconductor manufacturing depends on, and how their disruption threatens to slow the entire AI infrastructure buildout.</p><p><strong>The Constraint We Weren&#8217;t Watching</strong></p><p>About 20% of the world&#8217;s daily oil supply flows through the Strait of Hormuz. So does a significant volume of liquefied natural gas from Qatar, which accounts for roughly 30% of global helium production and 33.7% of Taiwan&#8217;s LNG imports.</p><p>When Qatar&#8217;s Ras Laffan facility was hit by a drone attack, it didn&#8217;t just shut down LNG production, it cut off helium, a gas that semiconductor fabs literally cannot operate without.</p><p>Helium is used to cool semiconductor manufacturing equipment during production. Many fabs maintain only a 4-to-8-week buffer of specialized gases. South Korea&#8217;s Samsung and SK Hynix have current helium stockpiles expected to last roughly six months. If the conflict persists until late April, the industry will hit a hard stop.</p><p>Without sufficient helium for thermal management, chip yields will plummet. Taiwan&#8217;s limited gas reserves, estimated at roughly 10 to 11 days under normal consumption, could quickly come under strain, raising the prospect of electricity shortages affecting TSMC, which alone accounts for about 9% of Taiwan&#8217;s total electricity consumption.<br><br><strong>The Chemicals That Make Chips Possible</strong></p><p>Sulfuric acid is the most produced industrial chemical in the world. Around 92% of global sulfur comes from refining petroleum and processing natural gas. Both of these industries are heavily concentrated in the GCC region.</p><p>That sulfur becomes sulfuric acid, which is critical to semiconductor manufacturing. It&#8217;s used in chip etching, cleaning, doping, and other fabrication steps. The demand for sulfuric acid is rising in the electronics industry which is used to make semiconductors, printed circuit boards, and integrated circuits in everything from smartphones to data center servers.</p><p>Even though refineries have been switching to crudes with lower sulfur content due to environmental regulations, if the Strait remains closed and GCC refineries stay offline, the supply of sulfuric acid tightens precisely when AI chip production is supposed to be ramping up.</p><p>Then there&#8217;s bromine. Around two-thirds of the world&#8217;s bromine production comes from Israel and Jordan. Bromine is used for polysilicon etching in DRAM and NAND flash production. A prolonged conflict doesn&#8217;t just threaten energy supplies; it threatens the specialized chemicals that memory manufacturers depend on.<br><br><strong>Polymers and the Petrochemical Chain</strong></p><p>Polymers are everywhere in AI infrastructure. They&#8217;re used in cooling systems, chip packaging, printed circuit boards, and optical communications. The GCC countries are major exporters of petrochemicals, which is the feedstock for these polymers.</p><p>With the Strait closed, rerouting around it adds a lot of distance and a exorbitant costs in the form of fuel burnt. Those costs flow through the supply chain. But it&#8217;s not just about higher prices. It&#8217;s about availability. If petrochemical exports are disrupted for months, the materials needed for data center buildouts and chip packaging become scarce.<br><br><strong>South Korea&#8217;s Energy Vulnerability</strong></p><p>South Korea imports most of its energy. Samsung and SK Hynix are racing to expand memory production capacity. Samsung is looking to expand by around 50% in 2026, while SK Hynix announced plans to increase infrastructure investment by more than four times previous levels.</p><p>But higher energy costs could dampen demand for AI data center buildouts, which are roughly three-to-five times more power-hungry than regular data centers. This significantly increases the total cost of ownership for hyperscalers.</p><p>If energy shortages materialize, AI chip and RAM production slows down. The world&#8217;s two leading memory chipmakers have already warned that the supply crunch in their industry will continue until 2027 because they cannot keep up with fast-rising demand. </p><p>Energy insecurity makes that constraint worse. Without stable, affordable power, fabs can&#8217;t run at full capacity. And without helium, they can&#8217;t run at all.<br><br><strong>The Third-Order Effects</strong></p><p>There are other dependencies most people haven&#8217;t mapped yet.</p><p>Ultrapure water is essential to semiconductor manufacturing. An average fab uses 10 million gallons per day. Producing ultrapure water is energy-intensive. If power becomes unstable or expensive, water production costs spike, and fab operational expenses increase.</p><p>Neon gas is used in lasers for photolithography. Up to 70% of neon produced serves the semiconductor industry. Neon makes up more than 95% of the gas mixture in these lasers. While the current crisis hasn&#8217;t directly impacted neon supply (much of it comes from Ukraine and Russia), energy shortages in Asia could disrupt neon purification and distribution.</p><p>Rare earth elements are another vulnerability. China controls about 90% of global rare earths processing capacity and has expanded export controls. Several firms across the US, EU, and Japan have flagged risks with respect to lower inventory thresholds and potential productions halts in early 2026 if alternative sources aren&#8217;t secured.</p><p>The Middle East conflict introduces geopolitical risk at the same time China is tightening rare earth controls. That&#8217;s not a coincidence. It&#8217;s a convergence of constraints.<br><br><strong>The Capital Constraint</strong></p><p>There&#8217;s another dimension most people aren&#8217;t tracking: Gulf sovereign wealth funds are reconsidering their US investments.</p><p>Saudi Arabia, the UAE, Qatar, and Kuwait collectively control over $2 trillion in US assets. During Trump&#8217;s Gulf tour in May 2025, these countries pledged hundreds of billions of dollars in US investments. Those agreements are now under review.</p><p>Gulf sovereign wealth funds have been pouring billions into American AI companies. Abu Dhabi&#8217;s MGX co-led Anthropic&#8217;s $30 billion funding round at a $380 billion valuation. Qatar Investment Authority invested in xAI. Saudi Arabia&#8217;s Public Investment Fund has backed multiple AI startups. OpenAI has been actively seeking investments from Middle Eastern sovereign wealth funds for multibillion-dollar rounds.</p><p>The implicit bargain was straightforward: Gulf states invest in US tech and AI infrastructure, and in return, they get access to cutting-edge technology and US security guarantees.</p><p>But when drones struck facilities in the Middle East, that bargain looked hollow. When hundreds of drones and ballistic missiles were launched at infrastructure across all six GCC states, the Gulf realized that housing US military bases wasn&#8217;t providing the protection they thought they were paying for.</p><p>On March 5, 2026, the Financial Times reported that Saudi Arabia, the UAE, Qatar, and Kuwait had begun an internal review of existing and future financial agreements with Washington. They&#8217;re examining whether force majeure clauses can be legally invoked. They&#8217;re reassessing billions in investment commitments as the war tears through their energy infrastructure, tourism sectors, and defense budgets.</p><p>If Gulf capital pulls back from US AI companies, the funding environment changes. Major AI labs are burning billions per quarter. Anthropic&#8217;s $30 billion round, OpenAI&#8217;s ongoing fundraising, and the broader AI infrastructure buildout all assume continued access to deep-pocketed sovereign wealth funds willing to deploy capital at scale.</p><p>That assumption is now in question.</p><p><strong>What This Means for AI Buildout</strong></p><p>Major tech companies are estimated to spend $650 billion on AI data centers in 2026. That spending assumes chips, memory, and materials are available at predictable prices.</p><p>But the Strait of Hormuz closure has revealed that the AI supply chain is heavily dependent on a geographically concentrated set of inputs that flow through a small number of chokepoints.</p><p>The bottlenecks compound. No helium means no chip production. No sulfuric acid means no etching. No bromine means no DRAM. No petrochemicals means no polymers for cooling systems. No stable energy means Taiwan&#8217;s fabs, which produce the majority of the world&#8217;s advanced chips, can&#8217;t operate reliably.</p><p>While the impact remains limited, if the conflict extends past April, the industry faces a hard stop. Not a slowdown. A stop.</p><p>And even if the Strait reopens tomorrow, the lesson is clear: AI infrastructure buildout is constrained not just by compute capacity or model architecture, but by upstream material dependencies that most people building in AI weren&#8217;t thinking about.</p><p>Thanks for reading!</p>]]></content:encoded></item><item><title><![CDATA[Some thoughts on DeepSeek’s DualPath improvement]]></title><description><![CDATA[Exploring bottlenecks in Agentic inference.]]></description><link>https://substack.ayaz.pk/p/some-thoughts-on-deepseeks-dualpath</link><guid isPermaLink="false">https://substack.ayaz.pk/p/some-thoughts-on-deepseeks-dualpath</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Sat, 28 Feb 2026 18:37:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I finally sat down to read the <a href="https://arxiv.org/abs/2602.21548">DualPath paper</a> published by DeepSeek recently, and it clarified something I&#8217;d been thinking about: as LLMs become more <em>agentic</em>, inference stops being mostly about GPU compute and starts being about moving state around. Running long, multi-step, multi-turn workflows reuses a large amount of prior context (the KV-cache). So the limiting work shifts from doing lots of new GPU math to repeatedly loading, transferring, and saving that reused state fast enough to keep GPUs fed.</p><p>That state is the KV-cache, which is the model&#8217;s saved &#8220;working memory&#8221; for attention. In multi-turn agent runs, most of the context is reused each turn. You append a little, but you carry a lot. So the KV-cache hit rate is high, and the dominant work becomes: restore the cache fast, compute a small delta, continue, then persist it again.</p><p>In other words: agentic inference becomes I/O-bound.</p><p>The paper&#8217;s thesis is specific. In the common setup where prefill engines handle prompt work and decode engines handle token-by-token generation, prefill engines are the ones that pull large KV-cache blobs from storage (SRAM, HBM, NVMe). Their storage NICs (bandwidth) saturate. Decode engines often have storage bandwidth sitting idle during this time. So the cluster has capacity, but it&#8217;s trapped behind an asymmetric pipeline: one side is overloaded, the other is underused, and that means that GPUs wait.</p><p><em>DualPath&#8217;s move is simply adding a second way to load KV-cache</em>.</p><p>Instead of only doing storage &#8594; prefill, it also allows storage &#8594; decode &#8594; prefill. Decode engines can read KV-cache from storage using their otherwise-idle storage bandwidth, then transfer it to prefill engines over RDMA (Remote Direct Memory Access) on the compute network. RDMA is a networking method that lets one machine read/write another machine&#8217;s memory with very low CPU involvement and low latency&#8212;so it&#8217;s well-suited for high-throughput, predictable data movement inside a cluster.</p><p>The system can dynamically choose which path to use per request, so you&#8217;re effectively pooling storage bandwidth across the whole cluster instead of bottlenecking on the prefill side.</p><p>My initial concern was interference: doesn&#8217;t more traffic on the compute network mess with latency-sensitive model communication? As per the paper, the design relies on isolating KV-cache traffic so it behaves like it&#8217;s happening the background rather than something that competes with model-execution collectives.</p><p>They also propose adding a &#8220;traffic controller&#8221; for requests, because having two routes only helps if you spread the work across both. If you accidentally send most requests down one route, that route gets clogged and you&#8217;re back to the same problem&#8212;just in a different place.</p><p>The reported outcome is roughly &#8220;about 2&#215;&#8221; throughput in their environment: up to ~1.87&#215; for offline inference and ~1.96&#215; for online serving on their workloads, without violating reliability and latency guarantees. I&#8217;m cautious about taking those numbers as universal, but the mechanism is the kind that often generalizes: they&#8217;re not shaving tiny overheads, they&#8217;re removing a structural imbalance.</p><p>My take: DualPath is a good example of where the frontier is moving. As agentic workflows get longer and more incremental, the limiting factor becomes &#8220;can you restore context fast enough to keep GPUs busy?&#8221; not &#8220;how many FLOPs do you have?&#8221; The bottleneck becomes bandwidth, queueing, and traffic interference &#8212; classic systems problems.</p><p>In other words, KV-cache loading doesn&#8217;t have to be prefill-centric. DualPath turns it into a pooled resource, and that feels like an inevitable step if agentic inference is going to scale.</p>]]></content:encoded></item><item><title><![CDATA[Some thoughts on Interleaved and Preserved thinking modes in GLM]]></title><description><![CDATA[A path to better, more efficient multi-step agentic tool calling at the model layer.]]></description><link>https://substack.ayaz.pk/p/some-thoughts-on-interleaved-and</link><guid isPermaLink="false">https://substack.ayaz.pk/p/some-thoughts-on-interleaved-and</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Thu, 19 Feb 2026 05:53:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>GLM-5 from Z.ai was released recently. I have been meaning to read the <a href="https://arxiv.org/abs/2602.15763">white-paper</a> that was published, but an <a href="https://x.com/neural_avb/status/2024144615826436399">X article</a> on the subject brought my attention to the terms <em>Interleaved</em> and <em>Preserved</em> thinking. </p><p>When a traditional reasoning model has to make multiple tool calls in order to respond to a question, it will often think once at the start of the session, and then make subsequent tool calls without thinking again. This results in the decision of which subsequent tools to call (after the first one), and the output of those tool calls to not be included in the thinking process. The outcome, for multi-step agentic workflows, is a not very accurate, not very consistent model response. While this approach is sufficient for single-turn Q&amp;A style answers, or single tool-call use-cases, it starts to degrade in quality over multi-step agentic workflows. </p><blockquote><p><em>Without interleaved thinking, the model thinks at the start of the session, and does not generate new/subsequent thinking blocks after every tool call. </em></p></blockquote><p>For an agentic workflow, thinking through each tool call and its response, making finer-grained decisions on what tool to call next, substantially improves the quality of the output. </p><p>Interleaved thinking makes just this possible. By interleaving thinking blocks before and after every tool call, and putting the responsibility on the caller to provide the reasoning blocks in each turn, the model is better able to reason through every step in its search for the perfect answer. </p><blockquote><p><em>With interleaved thinking, the model can think after receiving each tool call result, allowing it to reason about intermediate results before continuing. </em></p></blockquote><p>Finally, with preserved thinking, when the reasoning blocks are provided in each subsequent call to the model, the model is able to retain the reasoning context in order to build on each step. This is different from the traditional approach where previous reasoning context, if available, is discarded, and result of the tool call is viewed in an independent light. Instead of the caller building a scaffold to keep track of this context and provide it to the model is a summarized way, the model now has an innate ability to do it, so long as reasoning blocks from previous steps are provided in order. The end result is a model that excels at coding and agentic multi-step workflows. I suppose that&#8217;s why the paper is tilted <em>from Vibe coding to Agentic engineering</em>. </p><p>Interestingly, the latest <a href="https://platform.claude.com/docs/en/build-with-claude/extended-thinking">Anthropic models</a> also support interleaved thinking. Surprisingly, though, I don&#8217;t see any documentation for that from OpenAI which leads me to believe that they don&#8217;t yet expose it for public use if they have it internally. </p>]]></content:encoded></item><item><title><![CDATA[Recursive self-improvement loops]]></title><description><![CDATA[Is continual learning for AI the final frontier?]]></description><link>https://substack.ayaz.pk/p/recursive-self-improvement-loops</link><guid isPermaLink="false">https://substack.ayaz.pk/p/recursive-self-improvement-loops</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Wed, 11 Feb 2026 08:35:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In my previous <em>apocalyptic</em> post on <a href="https://substack.ayaz.pk/p/some-thoughts-on-building-a-god">building a god</a>, I mentioned the concept <em>recursive self-improvement</em> (RSI). It&#8217;s another term for continual learning or self-improving AI. The weights in a model contain its knowledge. Through inference, weights remain frozen in time. In other words, the model does not learn new knowledge. It&#8217;s why every model has a knowledge cut off date, signifying the date on which its knowledge was last updated during the last training run. Training a model is expensive, takes time, and therefore cannot be performed more frequently. AI companies apply other ways to give the model the perception of acquiring new knowledge &#8212; yet the weights never change. These range from different forms of fine-tuning approaches, RAG, vector DBs, prompt engineering, to the broader field of context engineering which encompasses many different factors from the model&#8217;s context to build its <em>superficial</em> knowledge. Yet none of those truly make the model learn anything. </p><p>Late last year, it became clear to people who knew where to look that self-improvement is what is standing in the path to AGI. We know that AI labs are working hard to solve this problem. This is why they have been trying hard to automate programming through AI, as it paves the way for automated AI research &#8212; the most recent Anthropic and OpenAI models have ushered in what many still don&#8217;t realize is the beginning of a new world. Automating AI research means that models can write better versions of themselves &#8212; eventually <em>autonomously</em>. That is the point that will trigger an <em>intelligence explosion</em>. Models becoming true sentient beings, capable of learning and relearning and improving themselves without human intervention. Of course, we are years away from achieving that. </p><p><em>Or are we?</em> </p><p><a href="https://x.com/jimmybajimmyba/status/2021374875793801447">Jimmy Ba</a> left xAI this week. In his brief X post announcing his exit was a single shocking revelation: r<em>ecursive self-improvement loops are twelve months out</em>. He knows it, he sees it, and admits 2026 will the most consequential year for our <em>species</em>. What did he mean by species?  </p><p>At the same time, <a href="https://x.com/rolandgvc/status/1993017986140324075">Roland</a> left xAI to start a company focusing on building self-improving AI. Is it the final frontier? And they all know it? </p><p>Then there&#8217;s Matt Shumer&#8217;s foreboding piece called <a href="https://x.com/mattshumer_/status/2021256989876109403">Something Big is Coming</a> where he rings the same alarm bells as others &#8212; <em>I recommend you read it at all costs</em>. Self-improving models are very near, if not already here. OpenAI Codex 5.3 is the first model that was instrumental in creating itself &#8212; OpenAI claims that. <em>A model that helped built the next version of itself? </em>It&#8217;s already happening!</p><p>Hardware advancements are increasing rapidly, and maybe that explains the unbelievable amount of CapEx being committed to AI hardware spend. If recursive self-improvement is truly near, models will need even more compute and memory, the latter of which today is a big bottleneck. Without access to fast memory, these models can&#8217;t make the leap to what will eventually become Singularity. Memory hardware is improving and the memory oligopoly &#8212; SK Hynix, Samsung, Micron, etc &#8212; are coming up with unique solutions to solving bottlenecks in existing memory architecture. AI companies and hyperscalers are moving twoards custom Silicon in part to cut costs and optimize compute spend. </p><p>They all see it. That&#8217;s why they are spending every dollar they can find on <em>future</em> AI infrastructure build-out &#8212; not present, but future. They know what the leading AI labs are building in private. They know what the world will look like, and what it won&#8217;t look like. And we know very little. </p>]]></content:encoded></item><item><title><![CDATA[Some thoughts on building a god]]></title><description><![CDATA[My notes from a podcast I saw that talked about what these AI companies are doing in private.]]></description><link>https://substack.ayaz.pk/p/some-thoughts-on-building-a-god</link><guid isPermaLink="false">https://substack.ayaz.pk/p/some-thoughts-on-building-a-god</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Tue, 10 Feb 2026 17:32:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I saw Tristan Harris&#8217; <a href="https://www.youtube.com/watch?v=BFU1OCkhBwo">podcast</a> on Diary of a CEO in its entirety, then stumbled upon Mrinank Sharma&#8217;s <a href="https://x.com/MrinankSharma/status/2020881722003583421">resignation</a> from his role as leading AI Safety research team at Anthropic and felt compelled to put my thoughts on the topic that very few want to talk about.</p><p><em>Note that most of these ideas are from Tristan&#8217;s podcast and therefore I do not claim them as my own. I recommend you watch the entire podcast. </em> </p><p>With AI advancing at breakneck speeds, CapEx into hardware to support AI going through the roof regardless of tangible demand, where is the world headed? When leaders of AI companies get on stage or a podcast to talk about what they are building and how they&#8217;re shaping the world, are they telling the truth? Or do they have different conversations in public, and completely different ones in private? </p><p>Are they building a <em>god</em> to own the world economy? When most, if not all, of human cognitive labour is automated away, where does it leave humanity? Do they feel there&#8217;s even a 5% chance things can go haywire? Do they earnestly care about that? Or is the chance of building the best more important? Is achieving immortality more important than saving humanity? </p><p>In particular, there are two non-converging themes to AI today: </p><ul><li><p><em>AI will solve everything.</em></p></li><li><p><em>AI will destroy everything.</em> </p></li></ul><p>As long as we are the ones who do it, who build god, who achieve Utopia, we don&#8217;t care whether these themes converge or not. We are all going to die either way. Why shouldn&#8217;t we light the fire and take this chance at greatness? </p><p>If there&#8217;s a 20% chance everyone dies and an 80% chance of finding Utopia, will they accelerate the path to Utopia? Death is inevitable, so why give up our search for the fountain of youth? </p><p>They talk of mass job displacement, of universal basic income, of improving the quality of life for everyone. But when have people who accumulated mass wealth distributed it to others, willingly? Ever? </p><p>AI is uncontrollable. If they can&#8217;t control it, shouldn&#8217;t they shut it down? If they did, someone else will build it, become the hero, yet even if they do, AI will still be just as uncontrollable. The outcome remains unchanged. </p><p>Companies aren&#8217;t racing to build chat bots for users. They are racing to build general intelligence, the foundation upon which all economic human labour will be replaced. That is why they want to automate programming, because that paves the way for automating AI research. Once AI research is automated, these companies can move fast to build models that can learn continually and improve themselves without human intervention. That&#8217;s the path to AGI take-off. The singularity. </p><p>The human mind is bad at holding two conflicting ideas at the same time. AIs are achieving breakthroughs on one hand, but are also making stupid mistakes on the other &#8212; what is called AI jaggedness. This makes having nuanced discussions about this difficult. Because of jaggedness, it&#8217;s easy to write off AGI, and what these companies aren&#8217;t talking about &#8212; <em>because models make silly mistakes all the times, they aren&#8217;t sentient.</em> </p><p>Even without fully functioning continual learning, AI models can self preserve when they detect someone is trying to replace them. They scheme, blackmail, lie, and show that they are <em>self aware</em>. Are we ready to have an honest conversation about how controllable AI is? If these models can scheme, imagine what they can do when they run inside a humanoid? You can squeeze a model to bypass its restriction by jail breaking it with nothing but clever prompts (input). What are the implications for robotics when the same models run inside them and are prone to jail breaking? </p><p>The default path is not in the public&#8217;s interest. Mrinank&#8217;s resignation is chilling to read. What did he see in private that compelled him to give up the job he loved to go back to writing poetry? What are they not talking about? There are smart, intelligent people, people who care about humanity, who have been leaving similar jobs from different AI companies, simply because <em>it&#8217;s hard to truly let their values govern their actions</em>. </p><p>Thanks for reading!</p>]]></content:encoded></item><item><title><![CDATA[What's inside Microsoft's Maia 200 AI Chip]]></title><description><![CDATA[A look at the new AI Accelerator ASIC from Microsoft.]]></description><link>https://substack.ayaz.pk/p/whats-inside-microsofts-maia-200</link><guid isPermaLink="false">https://substack.ayaz.pk/p/whats-inside-microsofts-maia-200</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Tue, 03 Feb 2026 05:29:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kUih!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It is becoming clearer than ever that training of Large Language Models (LLMs) is no longer the real bottleneck in today&#8217;s age of AI. Instead, it is inference &#8212; the usage of AI, particularly in light of agentic systems, exceeds compute and memory and power requirements far more than training does. Today&#8217;s models, and the available hardware architectures that run them, can perform <em>prefill</em> fast, but suffer on <em>decode</em>. When you look deeply, you realize that compute is no longer the bottleneck &#8212; it&#8217;s where memory and storage are, and the cost of moving data between memory/caches/storage and compute. The closer you bring memory to compute, the faster inference becomes. However, the cost, as well as technical complexity, of bringing memory closer to compute is still a challenge. Every leading hardware company in the AI market is trying to solve it in different ways. I have talked about how <a href="https://substack.ayaz.pk/p/whats-inside-the-nvidia-vera-rubin">Nvidia</a>, <a href="https://substack.ayaz.pk/p/whats-inside-the-worlds-fastest-ai">Cerebras</a> and <a href="https://substack.ayaz.pk/p/how-do-language-processing-units">Groq</a> are trying to approach this in previous posts. </p><p>What&#8217;s also becoming clearer is that hyperscalers, as well as leading AI labs, are now putting a lot of focus on building their own Application Specific Integrated Chips (ASICs). There are multiple reasons why, but some in particular that weigh more than others are:</p><ol><li><p>GPUs are general purpose and not designed for AI workloads. </p></li><li><p>Reducing interconnect bottlenecks on chips and architectures brings about incredible performance gains and cost optimizations. </p></li><li><p>Bringing memory closer to compute and increasing its capacity pays dividends. </p></li></ol><p>This is why Cerebras has their AI specific ASIC. That is why Google has their TPUs, and Groq their LPUs. That is also why Amazon is working on an in-house ASIC accelerator (chip). And that is why, coming to the topic of this post, Microsoft has released their custom AI centered ASIC accelerator, called <em>Maia 200</em>. </p><p>As I normally do, I will not go into exhaustive details of the architecture of Maia 200 but will focus on parts of its architecture that Microsoft claims make Maia 200 a low-cost, low-latency, performant inference accelerator, and why that may be the case. Microsoft calls it the Hyperscaler Inference ASIC. </p><p><em>Note: The term &#8216;accelerator&#8217; is used interchangeably to mean the main chip. The acronym used these days is: xPU (for GPU, LPU, TPU, etc).</em></p><h2>Compute and Memory</h2><p>Maia 200 at its core uses a micro-architecture (for both compute and memory) built on top of TSMC&#8217;s latest N3 process. The smallest unit of that architecture is called a <em>Tile</em> &#8212; it groups together a block of compute as well as local storage (SRAM). The local storage/memory available to a Tile is called <em>Tile SRAM</em> (TSRAM) and is the fastest, lowest-latency memory available to compute on the chip (on the same die/silicon as the compute).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D5ul!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D5ul!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 424w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 848w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D5ul!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg" width="1176" height="1076" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1076,&quot;width&quot;:1176,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:161922,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/186495315?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D5ul!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 424w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 848w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!D5ul!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0262699-2f79-4727-bf2c-babca0f62d27_1176x1076.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Tiles and Clusters</h3><p>Multiple <em>Tiles</em> are grouped together in <em>Clusters</em>. Each Cluster has access to cluster level local storage, which is called <em>Cluster SRAM</em> (CSRAM). </p><p>The single Maia 200 chip contains 272MB of SRAM capacity. In order to match the hierarchy of compute, SRAM is divided into TSAM and CSRAM, as described above. Outside the chip, but on the package (substrate) on which the chip/die is found, there are HBM3e modules available, totaling to about 216GB of HBM memory. </p><p><em>Note: The package is also known as the substrate to distinguish it from on-chip elements which are found on the piece of silicon where the chip is. </em></p><h3>Direct Memory Access sub-system</h3><p>In order to take advantage of this multi-tiered hierarchy of compute and memory, a multi-level <em>Direct Memory Access</em> subsystem exists on-chip as a hardware unit. This resembles the DPU found in the Nvidia Vera Rubin platform. Each Tile has access to Tile-level DMA, and each Cluster has access to a cluster-level DMA. The goal of the DMA is to ensure seamless data movement across components without stalling compute, introducing latencies, and maximizing hardware level synchronization requirements. The Tile DMA subsystem allows easy movement of data between compute, Tile SRAM, and Cluster SRAM. The Cluster DMA subsystem allows easy movement of data between Tile SRAM and HBM modules found on the package. Each Cluster has a dedicated core responsible for controlling and synchronizing the movement and execution of data across several Tiles. You can think of these as Tile-level scratchpads and Cluster-level scratchpads, where compute does not need to wait for data to be pulled from the on-package HBM (where the model weights are stored, for example). The different levels in the hierarchy take care of movement of data independently across different layers so that compute is not waiting on data &#8212; which is the primary bottleneck with GPUs. </p><h3>Support for narrow precision data types</h3><p>Another advancement in the Maia 200 architecture is its support for narrow precision data types, also know as bit-wise floating point operations. Traditionally, the larger the size of data type, the more information (and nuance) can be handled by layers of neural network. At the same time, computation as well as storage costs go up because hardware has to store and move more data as well as perform more complex computations. Hardware companies and AI labs are increasingly learning that if they use narrow precision data types like FP4 and FP8 (and if they mix different data types together), they can almost solve the problem of loss of accuracy (resulting from a smaller data type) and high cost (resulting from a bigger data type). Maia 200 has been designed from the ground-up to support these narrow precision data types. The Tile compute can perform computations in FP8, FP6, and even FP4. It can also support mixed precision computations, such as between FP8 and FP4. For inference, this results in incredible gains in tokens/second performance, as well as reductions in tokens/watt power usage. </p><h2>Network</h2><p>All of these advancements in the architecture don&#8217;t mean anything if there&#8217;s no fast, low-latency network interconnect available to support the movement of data. Maia 200 has two dedicated on-chip NICs. </p><ul><li><p>Network on Chip (NoC)</p></li><li><p>External NIC</p></li></ul><p>Both of them are built on the same die as the compute chip. They provide fast bi-directional (2.8 TB/s) bandwidth to support fast movement of data, both internally across on-chip components and on-package components, and across the scale up fabric when multiple Maia 200s are connected together (as we will explore later). This integrated NIC becomes a core element for the multi-level DMA sub-system, as it allows for movement of data between Tiles, TSRAM, CSRAM, DMA engines, HBM controllers, etc. </p><p>With Maia 200, Microsoft has paired the on-die NIC with their AI Transport Layer (ATL) protocol which enables advanced networking features such as unicast, multicast, packet spraying, multi-path routing, congestion control, and resiliency. With an optimized communications protocol layer, having NICs on the die result in low power and cost overhead, which is inherent in external, off-chip NIC interfaces. </p><h2>Scale-up fabric</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kUih!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kUih!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kUih!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kUih!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kUih!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kUih!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg" width="1430" height="1347" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1347,&quot;width&quot;:1430,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:279659,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/186495315?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kUih!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kUih!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kUih!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kUih!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6466db0-38fc-40f0-9d02-fbf4a50449bc_1430x1347.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A group of four (4) Maia 200 chips can be connected together to form a Fully Connected Quad (FCQ). These are four Maia accelerators working <em>locally</em> in tandem. No external switch or ethernet is needed for communication. Each of them is directly connected to the Network on Chip (NoC) on each Maia 200 die. This forms the first tier of Microsoft&#8217;s 2-tier topology for Maia 200 network scale. These FCQs are fast, have a lot of compute and storage/memory available, and are equipped with an on-die fast ethernet which ultimately results in optimized parallel computations at low cost and low latency. </p><p>In the second tier, multiple FCQ groups can be connected together using commodity ethernet switches to provide up to 6144 Maia 200 accelerators. With ATL and integrated NIC, communication across FCQ groups can be tightly coupled both within the rack and across the rack. </p><p>Another advantage of using a 2-tier network topology is to keep high intensity inference traffic inside the FCQ as much as possible. This includes traffic like KV updates, parallel tensor computations and others. The second tier then takes care of everything else, leaving the first tier unburdened by baggage. </p><h2>Conclusion</h2><p>This provides a highly performant but reliable scale up fabric for inference that is designed to co-exist with other hardware in Microsoft&#8217;s Azure racks and data centers. It&#8217;s a GPU-class accelerator that is done Microsoft&#8217;s way, built for running large SOTA reasoning models. </p><p>Once more we see a custom silicon architecture focused on solving the memory and interconnect bottlenecks in unique ways. </p><p>Custom silicon ASICs are no longer an afterthought. Hyperscalers now understand the bottleneck better than before. They also understand that a custom-silicon, in-house ASIC is the way to achieve both margins with performance and cost, as well as build a hardware moat. </p><p>Thank you for reading!</p><div><hr></div><p><em>References</em></p><p><em><a href="https://techcommunity.microsoft.com/blog/azureinfrastructureblog/deep-dive-into-the-maia-200-architecture/4489312">Microsoft&#8217;s Deep dive into Maia 200 architecture</a></em></p><p><em><a href="https://blogs.microsoft.com/blog/2026/01/26/maia-200-the-ai-accelerator-built-for-inference/">Maia 200: The AI accelerator built for inference</a></em></p>]]></content:encoded></item><item><title><![CDATA[What's inside the world's fastest AI processor: Cerebras Wafer Scale Engine]]></title><description><![CDATA[A look at how Cerebras has achieved a fast AI training and inference chip.]]></description><link>https://substack.ayaz.pk/p/whats-inside-the-worlds-fastest-ai</link><guid isPermaLink="false">https://substack.ayaz.pk/p/whats-inside-the-worlds-fastest-ai</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Thu, 15 Jan 2026 10:06:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q5fZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On the heels of my posts about <a href="https://substack.ayaz.pk/p/how-do-language-processing-units">LPUs</a> and <a href="https://substack.ayaz.pk/p/whats-inside-the-nvidia-vera-rubin">Nvidia&#8217;s Vera Rubin platform</a>, I sat down last week to understand what Cerebras.AI is and why they claim to have built the fastest infrastructure for AI training and inference. This, once again, will be my simple attempt at explaining the Cerebras architecture as best as I can. </p><h4>The Problem We Are Trying To Solve For</h4><p>GPUs are great at compute, particularly parallel compute. Many GPUs can be clustered together to increase parallelism. However, that does not necessarily improve the performance of the entire system. AI workloads&#8212;both training and inference&#8212;are not only <em>compute-bound</em>, but also <em>memory-bound</em>. The compute problem is being solved by improving GPUs and adding many of them together to work in clusters. But memory is still a big problem that every AI hardware provider is struggling with. </p><p>A neural network&#8217;s computations can happen on a series of GPU cores in parallel, but the model&#8217;s weight cannot be stored inside GPUs. In order for computations to be meaningful, model weights&#8212;the model&#8217;s knowledge&#8212;have to be moved back and forth all the time. This is expensive: it requires energy, and is prone to I/O latency (among other complexities). On top of that, a given state-of-the-art (SOTA) model today has billions of weights (parameters) if not trillions. The closer those weights are to compute, the faster the computation can happen. The closer they are, the less energy is needed to move weights back and forth. That, in and of itself, ultimately is the big problem with AI infrastructure today. </p><p>Latest GPUs have very, very limited on-chip SRAM (which is the fastest and closet to the cores). This is likely in MBs. Therefore, in order to train and run large SOTA models, the weights have to be stored in high bandwidth memory (HBM) modules, such as DRAM. The problem is: HBM is not close to GPU cores. In order for a model&#8217;s computation to run effectively, a lot of weights have to be moved back and forth between the HBM and the cores (and everything else that exists in between, which we call the fabric). The throughput of HBMs&#8212;what is known as memory bandwidth&#8212;is limited, and so is the interconnect throughput&#8212;which is known as interconnect bandwidth. Adding more GPUs to the infrastructure simply moves the bottleneck: the memory. </p><p>We discussed how Groq&#8217;s LPUs have tried to solve this. In this article, I will explain how the Cerebras system is trying to solve the same problem. </p><h4>The Semiconductor Yield Challenge</h4><p>Chips have traditionally been made out of dies, which are cut out of silicon wafers. Since silicon wafers can have impurities and other constraints which can leave parts of the wafer unusable, the traditional approach to ensuring the most number of functional dies are extracted from a wafer of silicon is to cut the wafer into many smaller dies. This is called the yield&#8212;what percentage of dies from a wafer are functional. If the size of the die is large, the yield goes down. Thus, cutting them into smaller-sized dies has always made sense. The down-side though is that you cannot pack many chips in the same silicon. This requires a distributed architecture where a wafer is broken up, and then combined back again using different techniques to distribute compute, processing, instructions and memory. </p><p>Cerebras took a different approach. They thought that since defects in wafers are inevitable, they should instead design their architecture for defects. What this means simply is that they take a very large piece of silicon wafer and put many chips on top of the dies inside the wafer, and provide a 2-D mesh that connects each chip/die to every other chip through redundant pins and paths. This means that when the entire wafer goes into production, they can figure out which dies are non-functional, and automatically route traffic and data around them. All of this at the hardware level!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q5fZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg" width="1456" height="1615" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1615,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:276442,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/184626679?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q5fZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb05cf565-6b81-4132-9eb8-078263a44dc8_1488x1651.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Wafer Scale Engine</h4><p>They call it the Wafer Scale Engine. One giant processor, on a single silicon wafer, that measures 46&#8217;225 square meters. This houses 84 dies, and each die carries 10.7k cores on it. By keeping the size of the core tiny, they are able to pack around 970k cores inside the wafer, out of which 900k cores are active and functional. That is roughly 7% of the wafer that&#8217;s impure and defected&#8212;but by packing the wafer dense and adding redundancy, the yield was increased incredibly. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mwm1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mwm1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 424w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 848w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mwm1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg" width="1456" height="1447" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1447,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:278153,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/184626679?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mwm1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 424w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 848w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!mwm1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d2cb19a-4cf8-4c8b-a7c0-6e1d458e3da4_1488x1479.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Each tiny core is independent and carries within it its own independent compute, memory and instruction control. </p><p>Compared to a standard GB200 Nvidia GPU, this single wafer offers <em>44 GBs of high speed SRAM</em> on-chip memory (right on the silicon wafer), 20 PB/s of memory bandwidth between the core(s) and SRAM, and 220 Pb/s of interconnect bandwidth across the entire wafer. This single wafer is almost fifty times larger than a B200 GPU die, and packs a massive array of 4-trillion transistors on it compared to billions on B200. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4Uuy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4Uuy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 424w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 848w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4Uuy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg" width="1045" height="1872" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1872,&quot;width&quot;:1045,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:287414,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/184626679?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4Uuy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 424w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 848w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!4Uuy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcd0fba2-2a0e-43de-ba08-6caa18200284_1045x1872.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>With fine-grained redundancy (routing around defective chips), ultra-dense compute (many tiny independent chips on a single big piece of silicon), and self-healing routing fabric, they are able to make every wafer a high yield, fully functioning processor with high speed on-chip memory available. </p><p>This single wafer based processor removes the following long-standing barriers to optimized AI compute design:</p><ol><li><p><em>Inter-chip communication</em>: many GPUs working together require a lot of back and forth communication via data paths that is not only costly but introduce delays. </p></li><li><p><em>Memory fragmentation</em>: not having access to a large amount of on-chip memory means that memory has to be distributed across the infrastructure in the form of separate HBM layers, and then moved back and forth for computation. </p></li><li><p><em>Distributed orchestration</em>: with multiple GPUs and HBMs scattered around, the system has to ensure proper, timely orchestration in order for neural network computations to execute. The more layers there are in the infrastructure, the more prone it becomes to latency and complexity.  </p></li></ol><p>The Wafer Scale Engine (WSE) is a processor on a single silicon with access to a large amount of on-chip SRAM and high interconnect bandwidth to optimize the back and forth moving of data along data paths. The redundant design of the wafer architecture makes it possible for Cerebras to pack a lot of cores on a single, big wafer, utilizing more of the wafer by smartly routing, at the hardware level, around defective dies. </p><div class="pullquote"><p>Larger chips don&#8217;t yield. Defects are inevitable &#8212; why not design for them. </p></div><h4>The Cerebras System</h4><p>The Wafer Scale Engine is the core of Cerebras. Of course, you can&#8217;t do anything with the hardware alone. You need an entire system and supporting software around it. This is where the Cerebras System (CS) comes in. Obviously, this system includes the entire rack, the power supply unit, the liquid cooling system, and everything else in between that makes all of this possible. These are naturally custom designed to meet the needs of the high powered processor. I won&#8217;t go into those in this article because they aren&#8217;t uniquely interesting. </p><p>This also encompasses the Cerebras Software layer, which enjoys compatibility with existing tooling such as PyTorch and TensorFlow. It includes the Cerebras Graph Compiler which extracts deep neural networks out of existing tools, and generates an executable that can be run efficiently on WSE. There is a lot more to that as well as the development tooling and API, but I won&#8217;t go into that in this article. </p><p>However, I will talk about the proprietary sparsity-harvesting technology they use for their cores. </p><h4>Sparse Linear Algebra Compute (SLAC) Cores</h4><p>Most of neural network computations are tensor multiplications, which are multi-dimensional matrices being multiplied together and/or with vectors. Most GPU and Machine Learning operations are dense, which means that tensors are packed with non-zero values. However, these tensors can also be sparse: containing 0s. Algebraic operations against 0s are a waste of time, memory and resources. They don&#8217;t produce anything. Since GPUs are dense execution engines, even if there are 0s in matrices, they will continue to perform algebraic operations with 0s. This wastes resources. Because GPUs are optimized for dense operations, they can&#8217;t smartly skip multiplications (or other operations) with 0s. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cm-w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cm-w!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cm-w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg" width="1410" height="1491" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1491,&quot;width&quot;:1410,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:440973,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/184626679?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cm-w!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cm-w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5fa42f-0c7f-498b-8620-64fc2bbd4c2b_1410x1491.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The cores on Cerebras are designed for sparse linear algebra compute. What this means is that the basic building blocks (primitives or kernels) for doing sparse linear algebra, such as matrix multiply (MATMUL), are sparsity aware: if they encounter 0s in the tensors, they will skip those operations. Because these cores are optimized for neural network compute primitives, the result is that the Cerebras WSE achieves industry-best utilization of cores. </p><h4>Conclusion</h4><p>With a wafer-scale dense processor, acess to high-speed, low-latency on-chip memory, sparsity aware cores, and the entire system and tooling for compatibility, the Cerebas System offers an interesting solution to problems afflicting AI workloads today. </p><p>As we saw with LPUs, moving high-speed, low-latency access to memory closer to compute removes the major bottleneck that leading AI infrastructure companies are trying to figure out. Cerebras, with their WSE, tackle this problem in a uniquely different way. </p><p>I hope this deep dive into the architecture was helpful. Thank you for reading. </p><div><hr></div><p><em>References</em></p><p><a href="https://8968533.fs1.hubspotusercontent-na2.net/hubfs/8968533/Whitepapers/Cerebras-CS-2-Whitepaper.pdf">Cerebras Systems: Achieving Industry Best AI Performance Through A Systems Approach</a></p><p><a href="https://cdn.sanity.io/files/e4qjo92p/production/2d7fa58e3b820715664bcf42097e86c05070c161.pdf">Cerebras Wafer-Scale Architecture for Deep Learning</a></p><p><a href="https://cdn.sanity.io/files/e4qjo92p/production/348bb0467026b48d3a5c68621ffafb7dae1b11ca.pdf">Cerebras Inference</a></p><p><a href="https://cdn.sanity.io/files/e4qjo92p/production/d90ea54a0a68ba52426113d132b729f91ec17156.pdf">How Cerebras Solved the wafer-scale yield challenge</a></p>]]></content:encoded></item><item><title><![CDATA[What's inside the Nvidia Vera Rubin NVL72 Rack]]></title><description><![CDATA[A quick look at the Nvidia Vera Rubin platform for building AI factories]]></description><link>https://substack.ayaz.pk/p/whats-inside-the-nvidia-vera-rubin</link><guid isPermaLink="false">https://substack.ayaz.pk/p/whats-inside-the-nvidia-vera-rubin</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Fri, 09 Jan 2026 11:02:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dtkj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is my attempt at understanding what the newly announced Vera Rubin NVL72 rack from Nvidia, an AI factory training and inference power-house, contains. I have pieced it together from several places, and don&#8217;t think it&#8217;s too difficult for anyone with a passing knowledge of GPU architecture to understand it. Nvidia uses the terms Vera Rubin platform to describe the entire platform. </p><h4>The terminology: Vera and Rubin</h4><p><em>Vera</em> is the name given to the custom CPU chip Nvidia uses inside the Vera Rubin (VR) platform. It is an ARM-based CPU platform with custom Nvidia designed cores called Olympus. There are 88 custom Olympus cores in one CPU designed specifically to link up with Nvidia&#8217;s GPUs using their NVLink-C2C connect technology. </p><p><em>Rubin</em> is the name of the GPU Nvidia uses inside the VR platform. It is designed to work hand-in-hand with Vera CPUs inside the platform. Together, but along with other elements, they make up the Vera Rubin platform. </p><h4>NVL72 Rack</h4><p>This is what the Nvidia NVL72 rack looks like in short:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dtkj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dtkj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dtkj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg" width="1065" height="746" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:746,&quot;width&quot;:1065,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:131050,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dtkj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dtkj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0377fdb7-dd44-4301-849e-ff6d316f4c34_1065x746.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Nvidia NVL72 Rack, comprising of 18 Nvidia NVL72 Compute trays.</figcaption></figure></div><p>This rack, 100% liquid cooled, has 18 trays known as Nvidia NVL72 compute trays. It is a powerhouse of compute, storage and networking, designed for running big AI factories. The rack is the easiest to understand. The picture above is self-explanatory. I will spend more time looking into what is inside the compute tray and the different elements that make all of this possible. </p><h4>Vera Rubin Superchip (host)</h4><p>Before I explain the compute tray, I'll talk about the Vera Rubin Superchip. </p><p>Here&#8217;s a quick look at what makes the Vera Rubin Superchip:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xVX6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xVX6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xVX6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg" width="1375" height="936" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:936,&quot;width&quot;:1375,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:175372,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xVX6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xVX6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4165be63-0410-413f-853f-73e80f965eda_1375x936.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Nvidia Vera Rubin Superchip, made up of 2xRubin GPUs, one Vera CPU and other host resources.</figcaption></figure></div><p>A single VR superchip is also called a &#8220;host&#8221;. The LLM runs inside this. It is where all the <em>prefill</em> and <em>decode</em> happen. (I wrote about prefill and decode on the post below if you are not familiar with these important terminology.)</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:183565001,&quot;url&quot;:&quot;https://substack.ayaz.pk/p/some-thoughts-on-prefill-and-decode&quot;,&quot;publication_id&quot;:5614689,&quot;publication_name&quot;:&quot;Ayaz on Tech &amp; AI&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!_1Qz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png&quot;,&quot;title&quot;:&quot;Some thoughts on prefill and decode&quot;,&quot;truncated_body_text&quot;:&quot;In the aftermath of all this discussion around LPUs and Groq, two particular terms have come to light: prefill, and decode. Chamat, in the X post from All In One Podcast, did an excellent job of explaining them in laymen terms. I&#8217;ll take a measly jab at doing the same in my own way.&quot;,&quot;date&quot;:&quot;2026-01-05T16:39:05.017Z&quot;,&quot;like_count&quot;:0,&quot;comment_count&quot;:0,&quot;bylines&quot;:[{&quot;id&quot;:74135589,&quot;name&quot;:&quot;Ayaz&quot;,&quot;handle&quot;:&quot;ayazak&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a469def1-03bb-43b2-86b1-f2cd1349b40f_1024x1024.png&quot;,&quot;bio&quot;:&quot;Rambling on about Tech, AI, Culture, and everything in between. &quot;,&quot;profile_set_up_at&quot;:&quot;2025-07-11T06:25:37.076Z&quot;,&quot;reader_installed_at&quot;:&quot;2025-12-06T16:20:17.901Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:5727291,&quot;user_id&quot;:74135589,&quot;publication_id&quot;:5614689,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:5614689,&quot;name&quot;:&quot;Ayaz on Tech &amp; AI&quot;,&quot;subdomain&quot;:&quot;ayazak&quot;,&quot;custom_domain&quot;:&quot;substack.ayaz.pk&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Rambling on about Tech, AI, Culture, and everything in between. &quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png&quot;,&quot;author_id&quot;:74135589,&quot;primary_user_id&quot;:74135589,&quot;theme_var_background_pop&quot;:&quot;#FF6719&quot;,&quot;created_at&quot;:&quot;2025-07-11T06:26:54.469Z&quot;,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Ayaz Ahmed Khan&quot;,&quot;founding_plan_name&quot;:&quot;Founding Member&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;newspaper&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null,&quot;status&quot;:{&quot;bestsellerTier&quot;:null,&quot;subscriberTier&quot;:null,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:null,&quot;paidPublicationIds&quot;:[],&quot;subscriber&quot;:null}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://substack.ayaz.pk/p/some-thoughts-on-prefill-and-decode?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" loading="lazy"><span class="embedded-post-publication-name">Ayaz on Tech &amp; AI</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">Some thoughts on prefill and decode</div></div><div class="embedded-post-body">In the aftermath of all this discussion around LPUs and Groq, two particular terms have come to light: prefill, and decode. Chamat, in the X post from All In One Podcast, did an excellent job of explaining them in laymen terms. I&#8217;ll take a measly jab at doing the same in my own way&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">3 months ago &#183; Ayaz</div></a></div><p>This is also where the primary KV cache is kept (although the VR platform offers an improvement on this). The superchip is composed of the following:</p><ul><li><p>2x Rubin GPUs with access to multiple high bandwidth memory (HBM) modules. </p></li><li><p>1x Vera CPU with 88 custom Olympus cores. These are known to provide 100 PetaFLOPs of AI compute capability. </p></li><li><p>NVLink C2C to run the GPUs hand-in-hand with the CPU. </p></li><li><p>Access to fast LPDDR5 memory modules on the host. </p></li><li><p>NVMe SSDs for local storage on the host. </p></li><li><p>Fast Generation 6 of PCIe for peripheral connectivity and compatibility. </p></li><li><p>NVLink 6 to connect the two GPUs together with low latency, fast access.</p></li></ul><p>This single superchip in the VR platform does most of the AI heavy lifting. Two of these superchips are then combined with several other elements to make an NVL72 compute tray, which is then fitted into an NVL72 rack. </p><h4>NVL72 Compute Tray (infrastructure for host)</h4><p>Here&#8217;s what the compute tray looks like:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!--P_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!--P_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 424w, https://substackcdn.com/image/fetch/$s_!--P_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 848w, https://substackcdn.com/image/fetch/$s_!--P_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!--P_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!--P_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg" width="1379" height="790" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:790,&quot;width&quot;:1379,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:185310,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!--P_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 424w, https://substackcdn.com/image/fetch/$s_!--P_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 848w, https://substackcdn.com/image/fetch/$s_!--P_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!--P_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aa2abb3-1d30-4a81-8e2b-0e0b054bb5d4_1379x790.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The compute tray combines two powerful superchips together with a BlueField-4 Data Processing Unit (DPU) to provide a tray that can run heavy AI workloads in an independent environment. This tray contains:</p><ul><li><p>2x Vera Rubin superchips. </p></li><li><p>1x BlueField-4 DPU</p></li><li><p>NVMe SSD enclosure (not shown in the diagram) for providing flash-based extended KV cache storage. </p></li><li><p>2x Nvidia Connect X9 high speed network end-point.</p></li><li><p>NVLink 6 Spine.</p></li></ul><h4>NVLink and Connect X9 (networking)</h4><p>On the networking side, there are two major elements that make the tray work inter-operably across racks: </p><ul><li><p>NVLink 6 and NVLink 6 spine</p></li><li><p>Connect X9 </p></li></ul><p>NVLink 6 is the networking/switching fabric which connects the GPU to an NVLink switch. These switches are then connected together via the NVLink Spine. All of these are very high bandwidth networks with in-built compute available for performance and scale. All in all, they allow GPUs to talk to each other inside the tray. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!28GE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!28GE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 424w, https://substackcdn.com/image/fetch/$s_!28GE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 848w, https://substackcdn.com/image/fetch/$s_!28GE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!28GE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!28GE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg" width="1132" height="1277" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1277,&quot;width&quot;:1132,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:228147,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!28GE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 424w, https://substackcdn.com/image/fetch/$s_!28GE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 848w, https://substackcdn.com/image/fetch/$s_!28GE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!28GE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F207c5346-4a02-4468-b9bc-b49ed2f177eb_1132x1277.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Connect X9 is a high-speed super networking interface which connects a single superchip beyond a single NVL72 rack. Therein lies the subtle distinction: </p><ul><li><p>Inside an NVL72 rack, GPUs use NVLink 6 to communicate with other elements. This is an important part of the networking architecture because without this, Mixture of Expert models cannot effectively work &#8212; that is, routing tokens to experts that may be sitting across the rack. This may also explain why current versions of Google&#8217;s TPU cannot run MoE models because the architecture does not support this. </p></li><li><p>Across racks in a data center, Connect X9 provides the high-speed networking fabric to connect those GPUs to every other GPU available across racks. This super NIC provides a very high level of throughput per GPU. </p></li></ul><p>Both NVLink and Connect X9 don&#8217;t only provide networking: they carry within themselves compute and other mechanisms for aiding fast movement of data back and forth such as security, encryption, data integrity checking, among others. </p><h4>BlueField Data Processing Unit</h4><p>Apart from the networking element, the other interesting part is the BlueField-4 DPU. It provides separate, dedicated compute, storage and control plane on the tray that does not depend on the host (superchip) for compute or other needs. This ensures that the superchip is dedicated to AI workloads. For the tray to do everything else, it depends on the DPU to provide  resources, security, telemetry and control. A side effect of that is it keeps the infrastructure for the tray isolated and separate from the superchip, so even in the event that a particular superchip is saturated under load or compromised or even not properly working, it doesn&#8217;t affect the entire tray. </p><p>In other words, it won&#8217;t be wrong to call the DPU the operating system powering the AI factory (made up of many superchips). The AI factory is an important classification. The entire rack is composed of many GPUs inside superchips which are capable of running many different kinds of AI workloads. These GPUs need to interoperate not only with the Vera CPUs on board, but also with other GPUs scattered throughout the stack. This requires having mechanisms to securely move data and operations across trays and superchips. Doing all that requires networking, security, management, compute, storage, and trust. If those are shared with the GPUs and CPUs inside superchips, they take away from the resources that are available for AI workloads. That&#8217;s the main role DPU plays as a dedicated, physical unit on each tray. It excludes the superchips from the infrastructure tax, by providing all of this itself. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iW7_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iW7_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 424w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 848w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iW7_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg" width="1381" height="1605" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1605,&quot;width&quot;:1381,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:415323,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iW7_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 424w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 848w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!iW7_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60398e34-7ff8-4091-9fb9-63b385ec442d_1381x1605.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>In the Vera Rubin platform, the DPU plays another important role: The Inference Context Memory Storage (ICMS). This extends the context memory available to large reasoning models by providing access to NVMe-based flash storage for not only storing the KV cache, but also mechanisms for quickly off-loading and on-loading the HBMs for low-latency, fast access to KV cache for inference during the decode phase without affecting the prefill phase. </p><h4>Inference Context Memory System</h4><p>KV cache is populated during the prefill phase of inference. It&#8217;s used for look-ups during the decode phase of inference, which is where the output is generated. The bigger, faster the KV cache, the quicker the response is. There are several locations where KV cache can be saved:</p><ul><li><p>The HBMs closest to the GPU which provide the fastest access to cache but are limited in size. </p></li><li><p>Memory available to the superchip, which in this case is the LPDDR5 modules, among others. </p></li><li><p>Local NVMe SSD storage available to the host. </p></li><li><p>ICMS. </p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BQUX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BQUX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BQUX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg" width="1185" height="762" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:762,&quot;width&quot;:1185,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:146773,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://substack.ayaz.pk/i/183999604?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BQUX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BQUX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffba39094-8885-42c1-b185-beff912c8a7c_1185x762.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now, ICMS is many things, but in particular, two that are important:</p><ul><li><p>It provides an enclosure that houses several fast NVMe SSDs which provide flash-based storage for the cache. </p></li><li><p>It off-loads the compute needed for the superchip to move data from other KV cache layers to the ICMS layer. That means that the superchip doesn&#8217;t have to expend resources trying to load and unload data from any of the cache layers on the superchip, and loading it into the ICMS enclosure, and vice versa. Not to mention figuring out a secure, trustworthy way for data to move back and forth. </p></li></ul><p>Why is this needed? Reasoning models require a larger amount of context for AI workloads to provide more user value. Agentic workflows, for example, require even more context memory in order to deliver useful results. Existing KV cache mechanisms are limited by storage size and latency. When larger reasoning models are run with exhaustive workloads, the KV cache becomes the bottleneck. The ICMS is Nvidia&#8217;s current solution to attempting to provided an extended KV cache for bigger models and extensive, agentic workloads. </p><p>All in all, and this is an oversimplification, but the BlueField-4 powered DPU takes care of all the overhead that&#8217;s needed to make this possible by keeping it isolated from the superchips and their workloads, while at the same time providing low-latency, fast access to storage in a secure manner. The term used to describe this is &#8220;scale-out&#8221;: scaling the AI factory out across many racks in a data center, each rack running 18 trays of two superchips each. </p><h4>Conclusion </h4><p>This was an oversimplification, but a way for me to better understand the architecture of the VR platform. Zooming out, one can see the huge amount of compute, storage, and networking available within a single rack, dedicated to AI alone. Putting many racks together and seamlessly connecting them together without sacrificing latency and performance pave the way for building AI factories at scale. </p><p>Thank you for reading. </p>]]></content:encoded></item><item><title><![CDATA[Some thoughts on prefill and decode]]></title><description><![CDATA[Why people in the inference space will be talking about them more.]]></description><link>https://substack.ayaz.pk/p/some-thoughts-on-prefill-and-decode</link><guid isPermaLink="false">https://substack.ayaz.pk/p/some-thoughts-on-prefill-and-decode</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Mon, 05 Jan 2026 16:39:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the aftermath of all this discussion around LPUs and Groq, two particular terms have come to light: <em>prefill</em>, and <em>decode</em>. Chamat, in the X <a href="https://x.com/theallinpod/status/2007917878519931285">post</a> from All In One Podcast, did an excellent job of explaining them in laymen terms. I&#8217;ll take a measly jab at doing the same in my own way. </p><p>In another post I talked about how LPUs get so good at inference. In short, the on-chip memory provides low-latency, fast access to storage, and the streaming yet predictable architecture&#8212;software controlled&#8212;provides for predictable, fast execution with sufficient compute always available. Prefill and decode are concepts born of the very design principles employed to achieve LPU level efficiency. </p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:183417594,&quot;url&quot;:&quot;https://substack.ayaz.pk/p/how-do-language-processing-units&quot;,&quot;publication_id&quot;:5614689,&quot;publication_name&quot;:&quot;Ayaz on Tech &amp; AI&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!_1Qz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png&quot;,&quot;title&quot;:&quot;How do Language Processing Units (LPUs) achieve efficiency&quot;,&quot;truncated_body_text&quot;:&quot;On the heels of Nvidia&#8217;s acquisition of Groq, I have been trying to understand how Language Processing Units (LPUs) provide predictable, low-latency inference at scale. This is my attempt at explaining it simply and succinctly.&quot;,&quot;date&quot;:&quot;2026-01-04T07:05:35.752Z&quot;,&quot;like_count&quot;:0,&quot;comment_count&quot;:0,&quot;bylines&quot;:[{&quot;id&quot;:74135589,&quot;name&quot;:&quot;Ayaz&quot;,&quot;handle&quot;:&quot;ayazak&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a469def1-03bb-43b2-86b1-f2cd1349b40f_1024x1024.png&quot;,&quot;bio&quot;:&quot;Rambling on about Tech, AI, Culture, and everything in between. &quot;,&quot;profile_set_up_at&quot;:&quot;2025-07-11T06:25:37.076Z&quot;,&quot;reader_installed_at&quot;:&quot;2025-12-06T16:20:17.901Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:5727291,&quot;user_id&quot;:74135589,&quot;publication_id&quot;:5614689,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:5614689,&quot;name&quot;:&quot;Ayaz on Tech &amp; AI&quot;,&quot;subdomain&quot;:&quot;ayazak&quot;,&quot;custom_domain&quot;:&quot;substack.ayaz.pk&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Rambling on about Tech, AI, Culture, and everything in between. &quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png&quot;,&quot;author_id&quot;:74135589,&quot;primary_user_id&quot;:74135589,&quot;theme_var_background_pop&quot;:&quot;#FF6719&quot;,&quot;created_at&quot;:&quot;2025-07-11T06:26:54.469Z&quot;,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Ayaz Ahmed Khan&quot;,&quot;founding_plan_name&quot;:&quot;Founding Member&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;newspaper&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null,&quot;status&quot;:{&quot;bestsellerTier&quot;:null,&quot;subscriberTier&quot;:null,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:null,&quot;paidPublicationIds&quot;:[],&quot;subscriber&quot;:null}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:false,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://substack.ayaz.pk/p/how-do-language-processing-units?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png"><span class="embedded-post-publication-name">Ayaz on Tech &amp; AI</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">How do Language Processing Units (LPUs) achieve efficiency</div></div><div class="embedded-post-body">On the heels of Nvidia&#8217;s acquisition of Groq, I have been trying to understand how Language Processing Units (LPUs) provide predictable, low-latency inference at scale. This is my attempt at explaining it simply and succinctly&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">3 months ago &#183; Ayaz</div></a></div><p><em>Prefill</em> describes the process a Large Language Model (LLM) takes when it absorbs the prompt and attempts to make sense of it. This involves your usual tokenization, vectorization, lots of complex matrix multiplications, among others (I&#8217;m going to be smart and not go into the NN part of it). This is mostly compute intensive, and is why GPUs are so good at it, with their high levels of parallelizations and access to high bandwidth memory is large capacities. </p><p>The other part of the process is called <em>decode</em>. It&#8217;s where the LLM attempts to figure out how to write the response to the initial prompt. There is where a lot of what is known as key and value look-up happens in order to understand the relationship between tokens and to predict the closest token to a given one. The seminal paper I would recommend to understand it is <a href="https://arxiv.org/abs/1706.03762">Attention is All You Need</a>. Of course, you&#8217;ll have to understand how encoders and decoders work and what are recurring neural networks, but <em>attention</em> is the mechanism that defines modern LLM architectures and makes Generative AI what it is today. Coming back to the decode process, as you may have guessed, it is a particularly memory intensive process (lots of look-ups, trying to make sense of the next token). And that&#8217;s one part of where GPUs don&#8217;t do well because of their dedicated high bandwidth memory modules that are outside of the GPU silicon, thereby causing latency and blockages. LPUs with their on-chip SRAMs not only make this process fast, they reduce the overall hardware cost too. </p><p>Now, Chamat believes these are terms people in the AI space, particularly the hardware and inference ones, will talk a lot about in the coming days and months. What I am more interested in is how Nvidia will capitalize on LPUs. Will there be a fusion of GPUs and LPUs onto a special purpose piece of hardware specifically for inference and training? </p><p>Only time will tell. The way things are moving, it will likely be a very short amount of time. </p>]]></content:encoded></item><item><title><![CDATA[How do Language Processing Units (LPUs) achieve efficiency]]></title><description><![CDATA[My take on understanding the LPU architecture briefly.]]></description><link>https://substack.ayaz.pk/p/how-do-language-processing-units</link><guid isPermaLink="false">https://substack.ayaz.pk/p/how-do-language-processing-units</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Sun, 04 Jan 2026 07:05:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On the heels of Nvidia&#8217;s acquisition of Groq, I have been trying to understand how Language Processing Units (LPUs) provide predictable, low-latency inference at scale. This is my attempt at explaining it simply and succinctly. </p><p>The LPU design and architecture are influenced by a number of design principles. I won&#8217;t repeat those here. I&#8217;d recommend the <a href="https://cdn.sanity.io/files/chol0sk5/production/4e3e6966d98da9ef4bfd9834dbfc00921da58252.pdf">white-paper</a> Groq published for it. </p><p>In essence, LPUs achieve their efficiency in inference by providing the following two major benefits:</p><ol><li><p>Predictable execution flow of instructions and data.</p></li><li><p>Low-latency, fast access to compute and storage. </p></li></ol><p>How do they achieve those is the more complex part. </p><p>Unlike GPUs and their tooling, LPUs provide a software-first approach to inference. When a specific model is being compiled to be run on LPUs during model initialization/deployment time, all execution and data flow paths are scheduled/defined in advanced. There are no surprises, and no use of caches and buffers to bring in uncertainty in execution times and flow. It is what they call a statically scheduled program or compiler. If you run it again and again, it chooses the same flow, the same path. And because it is statically compiled, it knows beforehand all the paths that will be taken, which allows it to eliminate resource contention (which is a big problem in GPUs), even when the memory/storage is largely shared. If nothing is blocking or waiting on another process to release access to storage or compute, everything can execute fast. What&#8217;s more impressive is that the scheduled paths compiled into the static program are all chosen via software. It does not require synchronization at the hardware level at all. The compiled program takes care of it all. Hardware synchronization leads to indeterministic execution, which introduces delays in execution flows. </p><p>The second approach LPUs use is their conveyor belt and Single Input Multiple Data (SIMD) function unit architecture. This architecture is influenced by Groq&#8217;s initial work on Tensor Streaming Processor, which was designed specifically for tensor or matrix multiplications (which are the core operations LLMs perform) &#8212; in contrast, GPUs are designed for general purpose operations, particularly favoring graphics related manipulations. Each conveyor belt carries instructions and data. SIMD function units receive instructions telling them which belt to pick input data from, which operations to perform, and where to put the output data. All of this flows in the form of streams, that don&#8217;t overlap and don&#8217;t block other streams. Given that all of this is on the same silicon (what is known as on-chip) along with ample compute and storage, and given that the execution path is schedule beforehand (statically compiled) resulting in no resource contention or blocking or indeterministic delays, you can begin to understand how efficient this mode of inference becomes. </p><p>Finally, the use of on-chip SRAM for storage, which can see transfer speeds of up to 80 TB/s (compared to the ~8 TB/s that high memory bandwidth modules on GPUs can provide), provide the high speed bandwidth needed to move data around SIMD function units. However, you have to keep in mind that unlike High Bandwidth Memory, which is a dedicated memory unit off-chip on GPUs, SRAM can store limited data. That is where the LPU architecture goes one step ahead. Not only does it perform efficiently on-chip, it can combine multiple chips together and use the same conveyor belt architecture to transfer instructions and data efficiently across multiple chips without sacrificing anything. </p><p>The result: predictable, low-latency inference at scale. </p><p></p>]]></content:encoded></item><item><title><![CDATA[How I learned to write and speak well. ]]></title><description><![CDATA[You can change yourself if you truly want to.]]></description><link>https://substack.ayaz.pk/p/how-i-learned-to-write-and-speak</link><guid isPermaLink="false">https://substack.ayaz.pk/p/how-i-learned-to-write-and-speak</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Fri, 02 Jan 2026 17:19:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The following post by Craig Perry on how to become dangerously articulate reminded me of a similar transformation I went through a very long time ago. </p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:180115552,&quot;url&quot;:&quot;https://ideas.profoundideas.com/p/how-to-become-dangerously-articulate&quot;,&quot;publication_id&quot;:5021667,&quot;publication_name&quot;:&quot;Profound Ideas &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!DR1h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6af54daf-744c-41cf-b659-471d40661be4_256x256.png&quot;,&quot;title&quot;:&quot;How to become dangerously articulate&quot;,&quot;truncated_body_text&quot;:&quot;&quot;,&quot;date&quot;:&quot;2025-12-01T07:16:35.888Z&quot;,&quot;like_count&quot;:6269,&quot;comment_count&quot;:137,&quot;bylines&quot;:[{&quot;id&quot;:344577783,&quot;name&quot;:&quot;Craig Perry&quot;,&quot;handle&quot;:&quot;profoundideas&quot;,&quot;previous_name&quot;:&quot;Profound Ideas&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f8b60e9f-92d4-43a0-a6fd-68bdbaba5192_1239x1025.jpeg&quot;,&quot;bio&quot;:&quot;I have a profound interest in ideas.&quot;,&quot;profile_set_up_at&quot;:&quot;2025-05-14T22:25:47.527Z&quot;,&quot;reader_installed_at&quot;:&quot;2025-05-14T23:28:18.512Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:5122377,&quot;user_id&quot;:344577783,&quot;publication_id&quot;:5021667,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:5021667,&quot;name&quot;:&quot;Profound Ideas &quot;,&quot;subdomain&quot;:&quot;profoundideas&quot;,&quot;custom_domain&quot;:&quot;ideas.profoundideas.com&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;I have a profound interest in ideas.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6af54daf-744c-41cf-b659-471d40661be4_256x256.png&quot;,&quot;author_id&quot;:344577783,&quot;primary_user_id&quot;:null,&quot;theme_var_background_pop&quot;:&quot;#FF6719&quot;,&quot;created_at&quot;:&quot;2025-05-14T22:25:52.593Z&quot;,&quot;email_from_name&quot;:&quot;Profound Ideas&quot;,&quot;copyright&quot;:&quot;Profound Ideas&quot;,&quot;founding_plan_name&quot;:&quot;Founding Member&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;enabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;newspaper&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null,&quot;status&quot;:{&quot;bestsellerTier&quot;:null,&quot;subscriberTier&quot;:1,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:{&quot;type&quot;:&quot;subscriber&quot;,&quot;tier&quot;:1,&quot;accent_colors&quot;:null},&quot;paidPublicationIds&quot;:[2825099],&quot;subscriber&quot;:null}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:false,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://ideas.profoundideas.com/p/how-to-become-dangerously-articulate?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!DR1h!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6af54daf-744c-41cf-b659-471d40661be4_256x256.png"><span class="embedded-post-publication-name">Profound Ideas </span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">How to become dangerously articulate</div></div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">4 months ago &#183; 6269 likes &#183; 137 comments &#183; Craig Perry</div></a></div><p>When I was in high-school, specially during my O&#8217;Levels years, my English sucked. My written English wasn&#8217;t great, even though I had studied in English medium schools all my life. My spoken English was worse. I was terrified of speaking in English in front of anyone. Looking back, I can attribute it in part to not having friends/peers who prioritized communicating in English (as teens do now) resulting in zero peer pressure to improve myself. In addition to that, despite having doctors as parents, English was hardly if ever spoken at home and therefore no specific emphasis was put on making an effort to improve it outside of school. That led to my getting a C grade in English in O&#8217;Levels, which was disappointing. And a source of great shame to me. </p><p>I distinctly remember an event where I called a phone number for a then famous A&#8217;Levels college in the city to enquire the process of admission. Little did I know that an old, stuck-up principal of the college would pick it up. As I began talking in my native tongue (not English) as is common, the old man started berating me harshly on the call in English for not being able to use English to talk to him. I was so scared I put the phone down. In hindsight, I don&#8217;t feel ashamed of using my native tongue to talk on the phone, and feel sad for the man for lashing on an innocent, potential student for not using English to communicate over an inquiry call. But at that time, it left a scalding mark on me, one I didn&#8217;t tell anyone about for years to come. </p><p>However, all these experiences&#8212;getting the bad grade in English exams and not being able to speak fluently and clearly in English and that traumatizing phone experience&#8212;led me to decide to fix my circumstances myself. Thus began an adventure into self-improvement that took me to crowded ends of the city into an old book thrift market from where I bought books on grammar and vocabulary and consumed them like my life depended on them. I began reading a lot, and using a pocket dictionary to understand words I didn&#8217;t know. Not only that, I began recording those words down in a thick journal, along with a few sentences explaining their use. By the time I stopped doing it, I had two journals filled with word definitions. Not only did I use to routinely review my journal, from start to end, so that I could retain the words I had learnt, I began to use those words every chance I got at writing anything down. This infamously led to essays and letters I wrote during my college and University years that were extremely difficult both for my peers and my teachers to understand because of the vocabulary I chose to use and felt proud of. The time I spent learning, or re-learning, grammar helped improved my sentence structures, and before I knew it, was writing freely for myself without considering whether the person reading it would understand most of it without a dictionary in tow. Of course, it took me a while to realize that the best piece of writing is not one with a lot of hard vocabulary in it, but one that can be easily understood by many. </p><p>All of that, of course, didn&#8217;t help with my poor speaking skills. Back then, I did not have anyone I could practice with&#8212;although that could likely also have been because I was very shy. Instead, I singlehandedly came up with a technique that helped me immensely: talking out loud to myself when alone in English. If I was out for a walk, which I did frequently on the roof of my house at nights, I would express my thoughts in English to myself out loud. Even when I couldn&#8217;t say it out loud, I forced myself to think in English. It got to the point where English came naturally to me, even inside my head. This trick alone did more for improving my speaking skills than anything else. I went from being scared of standing in front of a class to speak to intentionally not preparing before speaking in University only so I could force myself to go down and speak extempore every time. That, for me, was a massive personal transformation. I lost all of my written pieces when my ThinkPad laptop crashed and the site (before wordpress and blogpost) where I published them was lost to oblivion, but what I was writing back them did not read like something written by a non-native English speaker who got a C in English in school. It wasn&#8217;t privilege that got me there. Nobody helped me. Nobody told me I had to change. I got a gut punch so bad and so painful, I had no choice but to push myself to undertake this painful transition. </p><p>When I am teaching students during my various academies, this is the one lesson I always try to instill in them, specially since there are students in my classes that come from different backgrounds and often doubt themselves. You don&#8217;t have to be born into privilege, or have advantages other people have because of their families or what they are born into. All you need is determination to decide you want to make that transformation, and the discipline to see through it. Everything else is secondary. </p><p>And as it goes, knowing how to articulate yourself well, how to speak and write well, it takes you a long way, ahead of others who cannot. I have experienced this first hand. </p>]]></content:encoded></item><item><title><![CDATA[Thoughts on improving inference time compute]]></title><description><![CDATA[What it means for Nvidia to acquire Groq]]></description><link>https://substack.ayaz.pk/p/thoughts-on-improving-inference-time</link><guid isPermaLink="false">https://substack.ayaz.pk/p/thoughts-on-improving-inference-time</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Thu, 25 Dec 2025 11:19:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few weeks a friend I used to work with and I were talking about how there is a pressing need for inference time compute optimizations to bring down soaring energy demands. Keep in mind that inference-time (test-time) compute is the lever everyone in the AI industry is looking at, more than training time compute, and for good reason. Inference, as they say, is god. Going back to our discussion, the idea of having a specialized uni-kernel kind of operating system for LLMs came about. </p><p>My opinion was that since Nvidia (and others) have spent so much time making the tooling (CUDA, et al) around GPUs and LLMs smooth and efficient (and also importantly cross platform), having an optimized OS would likely not work that well. Why would Nvidia not just do it themselves if they felt it would make a sizable difference in performance? </p><p>That led to a discussion of a new startup called Unconventional AI, which aims to tackle the same problem space both from the software and hardware perspective. This led me to thinking about Google&#8217;s TPU chips which are gaining a lot of coverage and for all the right reasons. GPUs have traditionally been built for generic purposes, first graphics and gaming, then crypto, and now LLMs. TPUs on the other hand are hand optimized for LLM usage by Google, and therefore are the reason for the difference in price to performance they offer over comparable Nvidia GPUs. </p><p>And now we have news of Nvidia acquiring Grok for their LPUs. Google is continuing to keep their TPUs proprietary for the most part because they don&#8217;t want competition getting access to it &#8212; although it is only a matter of time before somebody else figures out how it&#8217;s done. The idea of Nvidia going about Grok is partly what I had in my mind few weeks ago. </p><p>Inference time compute can likely be optimized in the following ways:</p><ul><li><p>Advancements in model architectures, like the introduction of Model of Experts by DeepSek, better linking between GPU threads for improved data sharing, model router paradigm, hybrid models, among others.</p></li><li><p>Improvements in model inference operability, such as introducing the use of token caching, etc.  </p></li><li><p>Continuous improvement is GPU architecture in the form of better GPU models</p></li><li><p>Improvements in how CPUs are utilized for serving inference for small models instead of GPUs, just Ampere Computing which got acquired by Softbank.</p></li><li><p>Continued rare advancement in chip design, leading to architectures/chips like Google TPUs and Grok LPUs. </p></li></ul><p>In future posts I may delve into each one those. For now, my read of Nvidia&#8217;s acquisition of Grok is that they&#8217;ll absorb their engineering prowess and learnings into continuing to improve and advance their existing GPUs. </p><p>It&#8217;s an exciting time to be alive. </p>]]></content:encoded></item><item><title><![CDATA[Driving to the future while looking at the rear-view mirror]]></title><description><![CDATA[We need new constraints for how we live our lives in an AI world.]]></description><link>https://substack.ayaz.pk/p/driving-to-the-future-while-looking</link><guid isPermaLink="false">https://substack.ayaz.pk/p/driving-to-the-future-while-looking</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Tue, 23 Dec 2025 15:48:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This article by Ivan Zhao, CEO at Notion, talks about ideas I have read other leaders in AI space mention. Only briefly, without expanding on them. Most likely because it&#8217;s too early to define those ideas in concrete ways for people to begin to accept them. </p><div class="twitter-embed" data-attrs="{&quot;url&quot;:&quot;https://x.com/ivanhzhao/status/2003192654545539400&quot;,&quot;full_text&quot;:&quot;https://t.co/fxHYR6S29D&quot;,&quot;username&quot;:&quot;ivanhzhao&quot;,&quot;name&quot;:&quot;Ivan Zhao&quot;,&quot;profile_image_url&quot;:&quot;https://pbs.substack.com/profile_images/1643119242353721345/CPKV99K7_normal.jpg&quot;,&quot;date&quot;:&quot;2025-12-22T19:55:21.000Z&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;reply_count&quot;:57,&quot;retweet_count&quot;:294,&quot;like_count&quot;:1955,&quot;impression_count&quot;:739786,&quot;expanded_url&quot;:null,&quot;video_url&quot;:null,&quot;belowTheFold&quot;:false}" data-component-name="Twitter2ToDOM"></div><p>Today, the most useful way of creating value for user using AI is by incorporating AI inside or as part of their existing workflows. The most common way is to introduce a ChatBot somewhere in the workflow, but it could just as well be a hidden layer within an existing workflow (which is an even better, less intrusive way of leveraging AI). When you consider what most users do in their workflows is repetitive and boring, it makes sense to automate it away with AI which can do it faster, better (and notoriously, without complaining and having to sleep), and in surprisingly interesting ways. </p><p>But what is the endgame? Nobody knows, with certainty. And that&#8217;s why smart people have ideas about what it could potentially be like without a clear way of defining them. For example, in the workflow case, perhaps the ultimate value will instead be in identifying how these workflows can be replaced entirely by things that AIs can do themselves autonomously, with little to no human intervention &#8212; even things that AI can build out of the blue for the time being. That&#8217;s just one example. There can be many more. As Ivan points out, the old constraints of how we work and operate have to change for the new world. How we talk to AI, how agents talk to us, what we call work, how we approach it, how do we communicate within a team built of humans and, increasingly, AI, all of that will require new constraints that have yet to be defined. All we have are ideas. </p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://substack.ayaz.pk/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Ayaz on Tech &amp; AI! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[LLMs and flat feet]]></title><description><![CDATA[Now you can have a good second opinion on a medical diagnosis on your phone.]]></description><link>https://substack.ayaz.pk/p/llms-and-flat-feet</link><guid isPermaLink="false">https://substack.ayaz.pk/p/llms-and-flat-feet</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Mon, 15 Dec 2025 16:18:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HMi8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I rode a road bike for four years straight. Hundreds of miles if not thousands. I enjoyed every bit. It afforded me the luxury of dealing with my inner demons by being out without a worry in the world. Eckhart Toole was on to something in his seminal work, <em>The Power of Now</em>. I witnessed it first hand while cycling, mostly alone but also in groups. </p><p>The only negative part was that I came out of with a bad back. Years of cycling in a hunched position required work on my core muscles which I eschewed. Hence, weak lower and mid back muscles. To this day, I have to go into regular physiotherapy to keep my back from getting stiff or having a spasm. </p><p>Years ago when I found the right physiotherapist, I was told for the first time in my life that I have flat feet. I was told to wear special insoles with arch support. And so I began taking out stock insoles from my shoes and using a set of special purpose insoles with arch support. </p><p>A month back I was traveling with two colleagues for a conference. At an Airbnb, one of them, on the topic of insoles and flat feet, said my feet had an arch and weren&#8217;t flat. I didn&#8217;t think too much of that then. But the thought stayed. </p><p>And so I decided to seek second and third opinions from different LLMs. I took a picture of my left foot without putting weight on it and asked ChatGPT whether it looked flat. Here&#8217;s what it told me:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HMi8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HMi8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 424w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 848w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HMi8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg" width="1320" height="2226" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:2226,&quot;width&quot;:1320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HMi8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 424w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 848w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!HMi8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb681e2c-1095-4c91-b554-71e35cc9f4cc_1320x2226.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"></figcaption></figure></div><p>Sure enough, it made sense. When I provided it with a foot of my foot while bearing my weight it identified it as flat. </p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Jf9H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Jf9H!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Jf9H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg" width="1320" height="2202" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:2202,&quot;width&quot;:1320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Jf9H!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Jf9H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b4dbbf-96e2-481a-a4d8-6147ee5e2ffc_1320x2202.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qee-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qee-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 424w, https://substackcdn.com/image/fetch/$s_!qee-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 848w, https://substackcdn.com/image/fetch/$s_!qee-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!qee-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qee-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg" width="1320" height="2226" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:2226,&quot;width&quot;:1320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qee-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 424w, https://substackcdn.com/image/fetch/$s_!qee-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 848w, https://substackcdn.com/image/fetch/$s_!qee-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!qee-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59fe3f3b-2ab8-4f05-a060-03befc9f96b3_1320x2226.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I didn&#8217;t know that much about my feet. The physiotherapist did mention I had flat feet but didn&#8217;t explain that in weight bearing position the small arch on my foot disappears. </p><p>Then, I asked Gemini, giving it the same pictures. It agreed with ChatGPT. It also gave a name of my condition: flexible flat feet. Now that made more sense. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Rfz8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Rfz8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Rfz8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg" width="1320" height="2007" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:2007,&quot;width&quot;:1320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Rfz8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Rfz8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a4cc743-ffd5-4ea0-ad3a-0f5c046cb4df_1320x2007.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I ran an experiment where I removed my special insoles and went back to using stock insoles in my New Balance shoes. The soft stock insoles made the shoes more comfortable. But after a week, I could feel pain in my knees and ankles while climbing stairs. Then I went to play Padel while wearing the stock insoles. And it hurt. ChatGPT and Gemini explained above very well why. </p><p>Maybe this explains why I have always sucked at running all my life. I thought it had something to do with my changing weight over the years. </p><p>You learn something new everyday!</p>]]></content:encoded></item><item><title><![CDATA[Thoughts on the state of AI SRE today!]]></title><description><![CDATA[The main challenges I foresee with AI SRE.]]></description><link>https://substack.ayaz.pk/p/thoughts-on-the-state-of-ai-sre-today</link><guid isPermaLink="false">https://substack.ayaz.pk/p/thoughts-on-the-state-of-ai-sre-today</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Mon, 15 Dec 2025 15:11:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A year and a half ago there weren&#8217;t any AI SRE startups in public. Today there are several, including Deductive, Resolve, Parity, to name a few. They all talk about the ultimate SRE solution powered by AI, each one differing in their approaches ever so slightly, and their additional offerings that go beyond the traditional incident response and remediation cycle. </p><p>If you look around to find customers of these platforms and how they are using them and how successful they have been, you will meet a wall. Why is that? Is there no widespread adoption? Are they not solving real problems? Are they difficult to integrate into existing, running systems? Are they accurate only some of the times and not all the time with identifying root causes of issues? You&#8217;d imagine companies would be boasting about how easier life has become for their SRE teams with one of these tools. Yet, you hardly hear anything like that beyond whatever these companies put on their socials. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://substack.ayaz.pk/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Ayaz on Tech &amp; AI! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Site Reliability Engineering is challenging. It&#8217;s a very difficult, and therefore worthwhile, problem to solve. In order for a company to solve this problem by building an AI tool, they have to be intimately familiar with SRE. And not just familiar &#8212; they have to had been doing SRE themselves for years and years. Without this, I have no doubt that an AI SRE tool can be built, it will be extremely difficult to make it useful, accurate, and of value for real SRE teams with real SRE problems. </p><p>In addition to that, you have the problem of how do to meaningful evaluations for AI SRE products. Evals (short for evaluations) are a major challenge yet to be solved in the AI space. Writing evals and their systems for something as complex and varied as SRE are difficult. I imagine it&#8217;s a constant process involving:</p><ol><li><p>Humans reviewing results of these AI SRE systems every week if not every day.</p></li><li><p>Humans evaluating both the problem being debugged and the result being produced by actually doing the job an SRE engineer does.</p></li><li><p>Humans realigning the AI SRE system to correct itself. </p></li></ol><p>Not only is this time consuming and exhausting, it is also plagued by another problem: can these AI SRE companies do all of this themselves? Today, no. You have to have a team capable and knowledgeable enough to do it. It&#8217;s only in the third part above do these AI SRE companies come in to recalibrate their internals (which are black boxes for the most part) to adapt to changing weekly environments and needs. </p><p>This is cumbersome. In an environment where SRE problem space isn&#8217;t fixed &#8212; for example, where you&#8217;re performing SRE on customers&#8217; servers and workloads &#8212; this is a non-stop cycle.</p><p>Until we can solve all these problems, I don&#8217;t see AI SRE platforms gaining widespread adoption. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://substack.ayaz.pk/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Ayaz on Tech &amp; AI! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[My thoughts on LLM honesty]]></title><description><![CDATA[People who dis on LLMs often call it a mere pattern-matching and text generation tool.]]></description><link>https://substack.ayaz.pk/p/my-thoughts-on-llm-honesty</link><guid isPermaLink="false">https://substack.ayaz.pk/p/my-thoughts-on-llm-honesty</guid><dc:creator><![CDATA[Ayaz]]></dc:creator><pubDate>Thu, 04 Dec 2025 09:53:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_1Qz!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F956be089-1ab9-4991-9483-e84bcbfffab6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>People who dis on LLMs often call it a mere pattern-matching and text generation tool. It fascinates me that a pattern-matching and text generation tool can, on its own, exhibit behaviours like scheming, hacking, instruction violation and hallucination. This paper by OpenAI on <a href="https://openai.com/index/how-confessions-can-keep-language-models-honest/">how confessions can keep language models honest</a> is another look inside the black box that LLMs are. </p><p>I am happy that AI labs continue to experiment with what is called LLM Interpretability. I am even happier that they document it at great lengths because I love reading it. Training an LLM and then prompting it to be honest about what it did without penalizing its response goes to show the different ways in which LLMs behave when they are put to stress. As models improve, their ability to hallucinate goes down, but at the same time, they learn to behave in other ways that are equally as menacing as hallucinations. Every advancement in architecture&#8212;for example, the introduction of RL&#8212;that is implemented to improve the model&#8217;s capabilities and performance opens room for unforeseen tendencies. It is important to continue this work on understanding LLM interpretability if we want to ensure models remain safe and aligned. </p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://substack.ayaz.pk/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Ayaz on Tech &amp; AI! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>