<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="https://feeds.feedblitz.com/feedblitz_rss.xslt"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	 xmlns:feedburner="http://rssnamespace.org/feedburner/ext/1.0">
<channel>
	<title>Computer Architecture Today</title>
	<atom:link href="https://www.sigarch.org/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.sigarch.org</link>
	<description>Informing the broad computing community about current activities, advances and future directions in computer architecture.</description>
	<lastBuildDate>Fri, 10 Apr 2026 14:00:43 -0400</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	
<image>
	<url>https://www.sigarch.org/wp-content/uploads/2017/03/logo_rgb.png</url>
	<title>Computer Architecture Today</title>
	<link>https://www.sigarch.org</link>
</image> 
<site xmlns="com-wordpress:feed-additions:1">125883397</site>
<meta xmlns="http://www.w3.org/1999/xhtml" name="robots" content="noindex" />
<item>
<feedburner:origLink>https://www.sigarch.org/computer-architectures-alphazero-moment-is-here/</feedburner:origLink>
		<title>Computer Architecture&#8217;s AlphaZero Moment is Here</title>
		<link>https://feeds.feedblitz.com/~/953617784/0/sigarch-cat~Computer-Architectures-AlphaZero-Moment-is-Here/</link>
		<comments>https://feeds.feedblitz.com/~/953617784/0/sigarch-cat~Computer-Architectures-AlphaZero-Moment-is-Here/#comments</comments>
		<pubDate>Fri, 10 Apr 2026 14:00:43 +0000</pubDate>
		<dc:creator><![CDATA[Karu Sankaralingam]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Architecture]]></category>
		<category><![CDATA[Evaluation]]></category>
		<category><![CDATA[Machine Learning]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=102162</guid>
		<description><![CDATA[<div><img width="300" xheight="187" src="https://www.sigarch.org/wp-content/uploads/2026/04/Gemini_Generated_Image_yndetsyndetsynde-1-300x187.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  fetchpriority="high" /></div>For decades, we have designed chips in fundamentally the same way: human intuition applied to a vanishingly small slice of an impossibly large design space. That paradigm worked when Moore&#8217;s Law was lifting everything. We could afford to be wrong. We could afford to miss the best design. Process scaling would close the gap. That [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="187" src="https://www.sigarch.org/wp-content/uploads/2026/04/Gemini_Generated_Image_yndetsyndetsynde-1-300x187.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" /></div><div>
<div></div>
<div>For decades, we have designed chips in fundamentally the same way: human intuition applied to a vanishingly small slice of an impossibly large design space. That paradigm worked when Moore&#8217;s Law was lifting everything. We could afford to be wrong. We could afford to miss the best design. Process scaling would close the gap.</div>
<div></div>
<div>That world is over. In a recent position paper — <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2604.03312">&#8220;Computer Architecture&#8217;s AlphaZero Moment: Automated Discovery in an Encircled World&#8221;</a> — I argue that we are at an inflection point. Not a gradual shift, but a structural break in how architecture must be practiced.</div>
<div></div>
<h3>From Idea Scarcity to Evaluation Scarcity</h3>
<div>The central claim is simple, but uncomfortable:</div>
<div></div>
<div><em>Computer architecture is no longer bottlenecked by ideas. It is bottlenecked by evaluation and telemetry.</em></div>
<div></div>
<div>For decades, the field has implicitly assumed that ideas are scarce — that the role of the architect is to generate the one clever mechanism worth exploring. Everything else follows. But recent evidence suggests the opposite. With modern large language models and agentic pipelines, hundreds of viable architectural ideas can be generated per day, thousands of candidate designs can be evaluated per week, and design cycles can compress from months to weeks.</div>
<div></div>
<div>This is not speculative. We built a system called the <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/VerticalResearchGroup/Gauntlet">Gauntlet</a> and tested it on 85 papers from ISCA 2025 and HPCA 2026 — largely outside the model&#8217;s training data. Across 475 independent runs, it produced viable architectural mechanisms 95% of the time: independently re-deriving authors&#8217; exact solutions in 48% of cases, and proposing valid alternatives the authors never considered in another 50%. Each took 10–20 minutes. This flips a foundational assumption of the field. If ideas are abundant, then the limiting factor is no longer creativity — it is <strong>which ideas we can evaluate, validate, and trust</strong>. This <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://pages.cs.wisc.edu/~karu/ArchAlphaZero/zero-arch/html/">link</a> has this corpus of problem statement and Gauntlet&#8217;s solutions.</div>
</div>
<div></div>
<div>
<h4></h4>
<h4>1. Evaluation is the new bottleneck</h4>
<p>We are moving from a world where the question was &#8220;Can we come up with a good idea?&#8221; to one where the question becomes &#8220;Can we evaluate 10,000 ideas fast enough to find the best one?&#8221; This elevates simulation infrastructure, analytical modeling, and verification into the central problems of the field. The &#8220;PhD student for three months&#8221; implementation bottleneck is already eroding — our system built first-principles performance models from papers in under 20 minutes. What replaces it is a race to build faster, more accurate, and more scalable evaluation pipelines.</p>
<h4></h4>
<h4>2. The telemetry divide</h4>
<div>If evaluation becomes central, then <strong>ground truth becomes everything. </strong>Over time, access to closed-loop deployment telemetry — real workloads, real performance counters, real system behavior at scale, and in low-level depth — may matter as much as architectural insight itself. This creates a risk of structural divide. Academic research, long dependent on proxy benchmarks, could drift further from production reality unless we collectively rethink how we share and access workload data.</div>
<h4></h4>
<h4>3. The end of the old boundary</h4>
<div>The traditional separation between &#8220;chip company&#8221; and &#8220;cloud provider&#8221; begins to dissolve. Automated architecture requires three tightly coupled capabilities: deployment (to generate telemetry), infrastructure (to evaluate designs at scale), and silicon expertise (to realize designs physically). No single traditional player owns all three. The result is convergence — either through vertical integration or new hybrid ecosystems.</div>
<h3></h3>
<h3>The Deeper Claim</h3>
<div>The more provocative claim is not about tools — it is about limits. Human-driven architecture is becoming structurally outmatched by the scale of the design space. This is not a statement about human ability. It is about combinatorics. The architectural search space — spanning parametric and structural choices — is effectively unbounded. Humans sample an infinitesimal fraction of it. That was acceptable in an era of abundance. It is not acceptable in an era where architectural efficiency is the primary lever for progress. The analogy to <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/1712.01815">AlphaZero </a> is not rhetorical. It is structural: when search, evaluation, and feedback loops become fast enough, intuition gives way to systematic exploration.</div>
<h3></h3>
<h3>What This Means for Research — and Teaching</h3>
<div>If this framing is even partially correct, it forces a rethinking of what it means to &#8220;do&#8221; computer architecture research. Several shifts seem likely. If machines can generate many viable solutions, identifying the *right problem* becomes the scarce intellectual act. Evaluation frameworks, modeling techniques, and telemetry integration may matter more than individual architectural ideas. And the reliance on fixed benchmark suites becomes increasingly fragile in a world driven by dynamic, evolving workloads.</div>
<div></div>
<div>The full paper includes a set of predictions and my opinions on how I see this playing out. This extends to how we teach. Do we still emphasize canonical microarchitectures, or shift toward trade-off reasoning, evaluation frameworks, and interpreting machine-generated designs? What does it mean to train a researcher when idea generation itself is becoming automated?</div>
<h3></h3>
<h3>A Call for Collaboration</h3>
<div>This is not a settled direction — it is a hypothesis that needs to be stress-tested by the community. If this resonates (or if you think it is completely wrong), I would love to engage on: new models for teaching architecture, shared evaluation infrastructure and artifacts, privacy-preserving approaches to workload telemetry, and workshops focused on problem formulation rather than solution novelty. If this is even half right, we may need to rethink our identity as a field. Let&#8217;s debate it.</div>
<div></div>
<div><strong>About the author:</strong> Karthikeyan Sankaralingam is Principal Research Scientist at NVIDIA and Professor at UW-Madison.</div>
<div></div>
<div>Disclaimer: These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</div>
</div>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/953617784/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/953617784/0/sigarch-cat~Computer-Architectures-AlphaZero-Moment-is-Here/feed/</wfw:commentRss>
		<slash:comments>1</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">102162</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/spilling-the-neural-tea-a-journey-down-the-side-channel/</feedburner:origLink>
		<title>Spilling the Neural Tea: A Journey Down the Side-Channel</title>
		<link>https://feeds.feedblitz.com/~/953386850/0/sigarch-cat~Spilling-the-Neural-Tea-A-Journey-Down-the-SideChannel/</link>
		<comments>https://feeds.feedblitz.com/~/953386850/0/sigarch-cat~Spilling-the-Neural-Tea-A-Journey-Down-the-SideChannel/#respond</comments>
		<pubDate>Mon, 06 Apr 2026 15:37:22 +0000</pubDate>
		<dc:creator><![CDATA[Adnan Rakin]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[deep neural networks]]></category>
		<category><![CDATA[Security]]></category>
		<category><![CDATA[side-channels]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=101933</guid>
		<description><![CDATA[<div><img width="300" xheight="164" src="https://www.sigarch.org/wp-content/uploads/2026/04/Gemini_Generated_Image_ptypq1ptypq1ptyp-300x164.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>Years ago, I came across three pioneering works (CSI-NN, Cache Telepathy, and DeepSniffer) in the field of reverse engineering neural networks that inspired my journey into side-channel attacks to uncover the secrets of modern Deep Neural Networks (DNNs). Fast forward to today, and there has been significant exploitation of side-channel attacks to discover the secrets [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="164" src="https://www.sigarch.org/wp-content/uploads/2026/04/Gemini_Generated_Image_ptypq1ptypq1ptyp-300x164.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p><span style="font-weight: 400;">Years ago, I came across three pioneering works (</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity19/presentation/batina"><span style="font-weight: 400;">CSI-NN</span></a><span style="font-weight: 400;">, </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://iacoma.cs.uiuc.edu/iacoma-papers/usenix20.pdf"><span style="font-weight: 400;">Cache Telepathy</span></a><span style="font-weight: 400;">, and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/10.1145/3373376.3378460"><span style="font-weight: 400;">DeepSniffer</span></a><span style="font-weight: 400;">) in the field of reverse engineering neural networks that inspired my journey into side-channel attacks to uncover the secrets of modern Deep Neural Networks (DNNs). Fast forward to today, and there has been significant exploitation of side-channel attacks to discover the secrets of neural networks. It&#8217;s a good time to provide an overview of where we stand, the outlook for the future, and the challenges ahead.</span></p>
<p><b>Motivation: </b><span style="font-weight: 400;">Let&#8217;s take a step back and first try to understand why we care about secrets in deep learning models. It basically boils down to two fundamental chall</span><span style="font-weight: 400;">enges associated with deep learning: i) Financial, ii) Security and Privacy challenges. In general, DNNs are intellectual property (IP), as they are products developed over years of research, implementation, and investment in computing units, and they entail significant training costs (time, energy, and labor), making them a valuable asset for their owners. Just to give a rough estimate, OpenAI’s </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://openai.com/index/gpt-4-research/"><span style="font-weight: 400;">GPT-4</span></a><span style="font-weight: 400;"> costs more than $ 100 million, and its </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://openai.com/gpt-5/"><span style="font-weight: 400;">GPT-5</span></a><span style="font-weight: 400;"> model is expected to be more than 5x as expensive (</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.forbes.com/sites/katharinabuchholz/2024/08/23/the-extreme-cost-of-training-ai-models/"><span style="font-weight: 400;">Cost of Training GPT</span></a><span style="font-weight: 400;">). I do not know about you, but if I spent 100 million on something, I would care about protecting it. The next challenge is knowing that a model secret gives an adversary white-box knowledge, which is extremely powerful in security and privacy settings. Any adversary with knowledge of a target victim&#8217;s model architecture (e.g., model type, layer sequence, and number) and weight information, formally defined as “white-box,” can launch powerful security (adversarial attacks) and privacy threats (model inversion attacks/membership inference attacks). As highlighted in Figure 1, the attacker’s final objective in the DNN reverse-engineering attack is to gain white-box privileges either to steal IP for financial gain or to launch subsequent attacks.</span></p>
<p><i><span style="font-weight: 400;">In summary, in security and privacy research, defining the threat model is the first step towards any exploitation, and the underlying assumption is often that a reverse-engineering attack has successfully uncovered the model architecture, weights, and other hyperparameters.</span></i></p>
<p><b>Attack Objectives: </b><span style="font-weight: 400;">By now, we have established that an attacker’s goal is to uncover two key properties of a victim DNN: its architecture and its parameters. However, this is an oversimplified goal and can often be misleading. To understand this, let&#8217;s consider a deep neural network as a function of </span><span style="font-weight: 400;">x, denoted</span><span style="font-weight: 400;"> </span><span style="font-weight: 400;">f</span><span style="font-weight: 400;">(x)</span><span style="font-weight: 400;">.  If an attacker wants to recover the exact victim model, their objective is for the stolen model to be identical to the original </span><span style="font-weight: 400;">f</span><span style="font-weight: 400;">(x)</span><span style="font-weight: 400;">, which is practically impossible for large-scale DNNs, whether using existing side-channel attacks or the exact victim dataset. As a result, a more practical and plausible goal for an attacker would be to achieve functional equivalence. If the stolen function is different, such as </span><span style="font-weight: 400;">g</span><span style="font-weight: 400;">(x)</span><span style="font-weight: 400;">,</span><span style="font-weight: 400;"> then, for incentive purposes, all an attacker cares about is that these two functions produce identical output,  i.e., </span><span style="font-weight: 400;">f</span><span style="font-weight: 400;">(x)=</span> <span style="font-weight: 400;">g</span><span style="font-weight: 400;">(x)</span><span style="font-weight: 400;">, for inputs </span><span style="font-weight: 400;">x</span><span style="font-weight: 400;"> that are of the attacker&#8217;s interest. As a result, achieving functional equivalence means recovering the DNN model architecture, often as close as possible to the victim architecture&#8217;s topology. On the weight side, even if an attacker cannot extract the exact weights, they must aim for a weight-space solution that captures the victim model&#8217;s functionality.</span></p>
<p><i><span style="font-weight: 400;">In summary, to steal a copy of the victim model/function, an attacker must identify the victim model architecture. In modern deep learning, where most practical applications use some version of a DNN model from an existing pool (e.g., GPT, Llama), recovering the architecture often boils down to detecting the model&#8217;s topology. Once the architecture is revealed, the attacker must recover the model parameters/weights, which is often a challenging part of the attack. Then again, as we discussed earlier, exact model recovery can be challenging, but achieving functional equivalence is a modest objective. Most importantly, to achieve functional equivalence, the attacker may not need to reveal the exact numerical weights; rather, gradually recovering coarse-grained information (e.g., weight sparsity, quantization pattern, weight distribution) is often sufficient.</span></i></p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-101935" src="https://www.sigarch.org/wp-content/uploads/2026/04/image7.png" alt="" width="884" height="483" /></p>
<p>Figure 1: Spectrum of attack threats characterized by attacker’s knowledge: Black-Box (No Knowledge), Grey-Box (Partial Knowledge, e.g., architecture), and White-box (Complete knowledge of model architecture and weights), the ultimate goal of reverse-engineering (AI-generated).</p>
<p><b>Attack Techniques and Capabilities.</b> <span style="font-weight: 400;">Among the popular types of side-channel attacks, i.e., physical and microarchitectural, both can be utilized in two different threat model settings. In edge or embedded devices, the physical side channel is the dominant threat, and several works (</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity19/presentation/batina"><span style="font-weight: 400;">CSI-NN</span></a><span style="font-weight: 400;">, </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity25/presentation/horvath"><span style="font-weight: 400;">BarraCUDA</span></a><span style="font-weight: 400;">) have shown that it is possible to recover the model architecture and weights of simple neural networks. On the other hand, micro-architectural side channels are a popular choice for resource-sharing cloud environments where users can upload and run their code in a colocated environment (e.g., </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://aws.amazon.com/sagemaker/"><span style="font-weight: 400;">Amazon SageMaker</span></a><span style="font-weight: 400;"> and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cloud.google.com/ml-engine/docs/technical-overview"><span style="font-weight: 400;">Google ML Engine</span></a><span style="font-weight: 400;">). Microarchitectural attacks have been successful in recovering model architecture across the board using </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity20/presentation/yan"><span style="font-weight: 400;">cache timing channels</span></a><span style="font-weight: 400;">, </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/10.1145/3373376.3378460"><span style="font-weight: 400;">memory access patterns</span></a><span style="font-weight: 400;">, and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/9153424/"><span style="font-weight: 400;">GPU context switching</span></a><span style="font-weight: 400;">. I acknowledge that there are many ways to recover DNN model weights, including </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity16/technical-sessions/presentation/tramer"><span style="font-weight: 400;">learning-based approaches</span></a><span style="font-weight: 400;"> and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity20/presentation/jagielski"><span style="font-weight: 400;">mathematical recovery</span></a><span style="font-weight: 400;"> techniques. In this blog post, I focus on side-channel attacks. At the same time, learning-based approaches can work as a complementary approach with side-channel attacks once the architecture information has already been leaked. </span></p>
<p><i><span style="font-weight: 400;">In summary, while side-channel attacks have been successful in leaking model architecture information, as the scale of modern DNNs, e.g., LLM weights, continues to reach new heights of billions, none of the existing side channels can scalably and predictably recover model parameter information. A common workaround would be to support these methods with a learning approach, assuming an attacker has a partial training set, which may not be practical, even in a resource-sharing environment where data remains private.</span></i></p>
<p><b><i>Future Challenges and Opportunities: </i></b></p>
<p><b>What is the future of architecture-recovery attacks, given the success of existing side channels?</b></p>
<p><i><span style="font-weight: 400;">As the next wave of vision and language domain architectures emerges, they present new challenges and opportunities for the microarchitectural side-channel attack community. These models require modern compute support, which can accelerate their inference (e.g., tensor cores), as GPUs become more modern and newer generations may leave new traces of side-channel information. Hence, these newer compute platforms (e.g., new GPUs) and their associated architectural support demand new innovation in side-channel capabilities to recover the model architecture. We must remember that architecture recovery is essential; without it, model parameter recovery is no longer useful. Moreover, as LLMs emerge as the dominant model, the question is not just about recovering weights or architecture; leaking other components, such as KV cache in a multi-tenant setting, can lead to </span></i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.ndss-symposium.org/wp-content/uploads/2025-1772-paper.pdf"><i><span style="font-weight: 400;">privacy leakage</span></i></a><i><span style="font-weight: 400;">. </span></i></p>
<p><b>Can a microarchitectural side channel alone ever be sufficient to recover model weight information? </b></p>
<p><span style="font-weight: 400;">The sheer scale of the modern model poses an even greater challenge for recovering weights, making direct recovery an ambitious, and even impossible, goal; instead, we should focus on functional equivalence. To achieve functional equivalence, weight recovery methods can set tiny stepping stones to augment learning-based recovery. </span></p>
<p><i><span style="font-weight: 400;">Complete weight recovery using a side channel at the scale of LLMs or even a smaller vision model may be too ambitious. Instead, the attacks should focus on coarse-grained information about weights, such as model sparsity levels, quantization mechanisms, weight sign recovery, and other optimization techniques. The key idea is to achieve functional equivalence by first recovering coarse-grained information, which is sufficient to support other learning-based recovery. It is time to work towards an achievable target: recovering this statistical weight-level knowledge and studying how critical their role is in improving subsequent attacks. As models and their computation units are increasingly optimized, leaking information such as sparsity levels or bit-widths will become more feasible by detecting optimized paths through side-channel leakage.</span></i></p>
<p><span style="font-weight: 400;">Finally, an attack is never the end goal. We probe attacks from every angle so we can study them before any attacker ever thinks about them. The endgame is always to develop subsequent defenses, which I leave for another discussion.</span></p>
<p><strong>About the author: </strong></p>
<p><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.binghamton.edu/computer-science/people/profile.html?id=arakin"><span style="font-weight: 400;">Adnan Siraj Rakin</span></a><span style="font-weight: 400;"> is an Assistant Professor at the School of Computing at Binghamton University. He received his Master&#8217;s (2021) and PhD (2022) from Arizona State University. He works on emerging security and privacy challenges in modern AI </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.usenix.org/conference/usenixsecurity21/presentation/rakin"><span style="font-weight: 400;">systems</span></a><span style="font-weight: 400;"> and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://openaccess.thecvf.com/content/ICCV2023/papers/Ahmed_SSDA_Secure_Source-Free_Domain_Adaptation_ICCV_2023_paper.pdf"><span style="font-weight: 400;">algorithms</span></a><span style="font-weight: 400;">. His paper on</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/9833743/"><span style="font-weight: 400;"> DNN model weight recovery</span></a><span style="font-weight: 400;"> has been </span><span style="font-weight: 400;">crowned as </span><span style="font-weight: 400;">Top Picks in Hardware and Embedded Security in 2024. </span></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/953386850/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/953386850/0/sigarch-cat~Spilling-the-Neural-Tea-A-Journey-Down-the-SideChannel/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">101933</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/to-sparsify-or-to-quantize-a-hardware-architecture-view/</feedburner:origLink>
		<title>To Sparsify or To Quantize: A Hardware Architecture View</title>
		<link>https://feeds.feedblitz.com/~/950020673/0/sigarch-cat~To-Sparsify-or-To-Quantize-A-Hardware-Architecture-View/</link>
		<comments>https://feeds.feedblitz.com/~/950020673/0/sigarch-cat~To-Sparsify-or-To-Quantize-A-Hardware-Architecture-View/#respond</comments>
		<pubDate>Thu, 12 Mar 2026 15:00:43 +0000</pubDate>
		<dc:creator><![CDATA[Sai Srivatsa Bhamidipati]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Accelerators]]></category>
		<category><![CDATA[deep neural networks]]></category>
		<category><![CDATA[Machine Learning]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=100754</guid>
		<description><![CDATA[<div><img width="300" xheight="164" src="https://www.sigarch.org/wp-content/uploads/2026/03/Blog-Image-2-Picsart-AiImageEnhancer-300x164.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>The debate of sparsity versus quantization has made its rounds in the ML optimization community for many years. Now, with the Generative AI revolution, the debate is intensifying. While these might both seem like simple mathematical approximations to an AI researcher, for a hardware architect, they present fundamentally different sets of challenges. Many architects in [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="164" src="https://www.sigarch.org/wp-content/uploads/2026/03/Blog-Image-2-Picsart-AiImageEnhancer-300x164.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p><span style="font-weight: 400;">The debate of sparsity versus quantization has made its rounds in the ML optimization community for many years. Now, with the Generative AI revolution, the debate is intensifying. While these might both seem like simple mathematical approximations to </span>an AI researcher, for a hardware architect, they present fundamentally different sets of challenges. Many architects in the AI hardware space are deeply familiar with watching the scale tip from one side to the other, constantly searching for a pragmatic balance. Let&#8217;s look at both techniques, unpack the architectural challenges they introduce, and explore whether a &#8220;best of both worlds&#8221; scenario is truly possible (Spoiler: It depends).</p>
<p><i><span style="font-weight: 400;">Note: We will only be looking at compute-bound workloads, which traditionally rely on dense compute units such as tensor cores or MXUs. We will set aside memory-bound workloads for now, as they introduce their own distinct set of tradeoffs for sparsity and quantization.</span></i></p>
<h2><b>Sparsity</b></h2>
<p><span style="font-weight: 400;">The core idea of sparsity is beautifully simple: if a neural network weight is zero (or close enough to it), just don&#8217;t do the math. Theoretically, pruning can save massive amounts of compute and memory bandwidth.</span></p>
<p><b>The Architecture Challenge: The Chaos of Unstructured Data</b></p>
<p><span style="font-weight: 400;">The golden goose of this approach is fine-grained, unstructured sparsity. It offers a high level of achievable compression through pruning, but results in a completely random distribution of zero elements. Traditional dense hardware </span><i><span style="font-weight: 400;">hates</span></i><span style="font-weight: 400;"> this. Randomness leads to irregular memory accesses, unpredictable load balancing across cores, and terrible cache utilization. High-performance SIMD units end up starving while the memory controller plays hopscotch trying to fetch the next non-zero value. To architect around this, pioneering unstructured sparse accelerators—such as</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/1602.01528"> <span style="font-weight: 400;">EIE</span></a><span style="font-weight: 400;"> and</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/1708.04485"> <span style="font-weight: 400;">SCNN</span></a><span style="font-weight: 400;">—had to rely heavily on complex routing logic, specialized crossbars, and deep queues just to keep the compute units fed, often trading compute area for routing overhead.</span></p>
<p><b>The Compromise: Structured and Coarse-Grained Sparsity</b></p>
<p><span style="font-weight: 400;">To tame this chaos, the industry shifted toward structured compromises. The universally embraced</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://developer.nvidia.com/blog/accelerating-inference-with-sparsity-using-ampere-and-tensorrt/"> <span style="font-weight: 400;">N:M sparsity</span></a><span style="font-weight: 400;"> (popularized by NVIDIA&#8217;s Ampere architecture) forces exactly N non-zero elements in every block of M. This provides a predictable load-balancing mechanism where the hardware can perfectly schedule memory fetches and compute.</span></p>
<p><span style="font-weight: 400;">More recently, to tackle the quadratic memory bottleneck of long-context LLMs, we&#8217;ve seen a surge in modern </span><i><span style="font-weight: 400;">sparse attention mechanisms</span></i><span style="font-weight: 400;"> that leverage block sparsity. Techniques like</span> <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/mit-han-lab/Block-Sparse-Attention"><i><span style="font-weight: 400;">Block-Sparse Attention</span></i></a><span style="font-weight: 400;"> and </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2003.05997"><span style="font-weight: 400;">Routing Attention</span></a><span style="font-weight: 400;"> enforce sparsity at the chunk or tile level. Instead of picking individual tokens, they route computation to contiguous blocks of tokens, allowing standard dense matrix multiplication engines to skip entire chunks while maintaining high MXU utilizations and contiguous memory access. Other approaches, like</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2309.17453"> <span style="font-weight: 400;">StreamingLLM</span></a><span style="font-weight: 400;">, evict older tokens entirely, retaining only local context and specific &#8220;heavy hitter&#8221; sink tokens.</span></p>
<p><span style="font-weight: 400;">The trade-off across these methods is clear: we exchange theoretical maximum efficiency for hardware-friendly predictability, paying a &#8220;tax&#8221; in metadata storage (index matrices), specialized multiplexing logic, and the persistent algorithmic risk of dropping contextually vital information.</span></p>
<h2><b>Quantization</b></h2>
<p><span style="font-weight: 400;">While sparsity aims to compute </span><i><span style="font-weight: 400;">less</span></i><span style="font-weight: 400;">, quantization aims to compute </span><i><span style="font-weight: 400;">smaller</span></i><span style="font-weight: 400;">. Shrinking datatypes from 32-bit floats (FP32) to INT8, or embracing emerging standards like the</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf"> <span style="font-weight: 400;">OCP Microscaling Formats (MX) Specification</span></a><span style="font-weight: 400;"> (such as MXFP8 E4M3 and E5M2), acts as an immediate multiplier for memory bandwidth and capacity. But the frontier has pushed much further than 8-bit. Recent advancements in extreme quantization, such as</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2402.17764"> <span style="font-weight: 400;">BitNet b1.58</span></a><span style="font-weight: 400;"> (1-bit LLMs using ternary weights of {-1, 0, 1}) and 2-bit quantization schemes (like</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2210.17323"> <span style="font-weight: 400;">GPTQ</span></a><span style="font-weight: 400;"> or <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2307.13304">Quip</a>), demonstrate that large language models can maintain remarkable accuracy even when weights are squeezed to their absolute theoretical limits.</span></p>
<p><b>The Architecture Challenge: The Tyranny of Metadata and Scaling Factors</b></p>
<p><span style="font-weight: 400;">From an architecture perspective, the challenge of extreme quantization isn&#8217;t just the math—it&#8217;s the metadata. To maintain accuracy at 4-bit, 2-bit, or sub-integer levels, algorithms demand fine-grained control, requiring per-channel, per-group, or even per-token dynamic scaling factors. Every time we shrink the primary datapath, the relative hardware overhead of managing these scaling factors skyrockets. Along with that, the quantization algorithm also becomes more fine grained, dynamic and complex. We are forced to add additional logic and even high-precision accumulators (often FP16 or FP32) just to handle the on-the-fly de-quantization and accumulation. We aggressively optimize the MAC (Multiply-Accumulate) units, only to trade that for the overhead of adding scaling factor handling and supporting a potentially new dynamic quantization scheme, which can outweigh the gains.</span></p>
<p><b>The Compromise: Algorithmic Offloading</b></p>
<p><span style="font-weight: 400;">To fix this without blowing up the complexity and area budget, the community relies on algorithmic co-design. Techniques like</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2211.10438"> <span style="font-weight: 400;">SmoothQuant</span></a><span style="font-weight: 400;"> effectively migrate the quantization difficulty offline, mathematically shifting the dynamic range from spiky, hard-to-predict activations into the statically known weights. Similarly,</span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2306.00978"> <span style="font-weight: 400;">AWQ (Activation-aware Weight Quantization)</span></a><span style="font-weight: 400;"> identifies and protects a small fraction of &#8220;salient&#8221; weights to maintain accuracy without requiring complex, dynamic mixed-precision hardware pipelines. By absorbing the complexity into offline mathematics, these techniques allow the hardware to run mostly uniform, low-precision datatypes.</span></p>
<p><span style="font-weight: 400;">However, much like the routing tax in sparsity, this algorithmic offloading comes with some compromises. These methods heavily rely on static, offline calibration datasets. If a model encounters out-of-distribution data in production (a different language, an unusual coding syntax, or an unexpected prompt structure), the statically determined scaling factors can fail, leading to outlier clipping and catastrophic accuracy collapse. Furthermore, relying on offline preprocessing creates a rigid deployment pipeline that prevents the model from adapting to extreme activation spikes on the fly.</span></p>
<h2><b>Is there a &#8220;best of both worlds&#8221;?</b></h2>
<p><span style="font-weight: 400;">So, knowing these trade-offs, do we sparsify or do we quantize? Many years ago, the </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/1510.00149"><span style="font-weight: 400;">Deep Compression</span></a><span style="font-weight: 400;"> paper proved we could do both. But today, pulling this off at the scale of a 70-billion parameter LLM is incredibly difficult. It suffers from the classic hardware optimization catch-22 (see </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.sigarch.org/dont-put-all-your-tensors-in-one-basket-hardware-lottery/"><span style="font-weight: 400;">All in on Matmul?</span></a><span style="font-weight: 400;">) : </span><i><span style="font-weight: 400;">No one uses a new piece of hardware because it’s not supported by software, and it’s not supported by software because no one’s using it.</span></i></p>
<p><span style="font-weight: 400;">So what&#8217;s the path forward for hardware architects? In my opinion, the following:</span></p>
<ul>
<li style="font-weight: 400;"><b>Deep Hardware-Software Co-design:</b><span style="font-weight: 400;"> The days of throwing a generic matrix-multiplication engine at a model are over. We need to work directly with AI researchers so that when they design a new pruning threshold or a novel sub-byte data type, the hardware already has a streamlined, fast path for the metadata.</span></li>
<li style="font-weight: 400;"><b>Generalized Compression Abstractions:</b><span style="font-weight: 400;"> Historically, we have designed accelerators that are either &#8220;good at sparsity&#8221; (with complex routing networks) or &#8220;good at quantization&#8221; (with mixed-precision MACs). Moving forward, we need to view these not as orthogonal features, but as a unified spectrum of compression. Architectures must be designed to dynamically adapt—perhaps fluidly dropping structurally sparse blocks during a memory-bound decode phase, while leaning on extreme sub-byte quantization during a compute-heavy prefill phase—potentially even sharing the same underlying logic.</span></li>
<li style="font-weight: 400;"><b>Balance Efficiency and Programmability:</b><span style="font-weight: 400;"> As explored in the &#8220;All in on MatMul?&#8221; post, we need to keep our hardware flexible. Over-fitting to today&#8217;s specific sparsity pattern or quantization trick risks building being trapped in the local minimum. We must maintain enough programmability to enable future algorithm discovery and break free from the catch-22.</span></li>
</ul>
<p><span style="font-weight: 400;">Some notable research going along this path include </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/pdf/2405.20935"><span style="font-weight: 400;">Effective interplay between sparsity and quantization</span></a><span style="font-weight: 400;">, which proves the non-orthogonality of the two techniques and explores the interplay between them and also the </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.cs.toronto.edu/~mmozaffari/compression-trinity/index.html"><span style="font-weight: 400;">Compression Trinity</span></a><span style="font-weight: 400;"> work which takes a look at multiple techniques across sparsity, quantization and low rank approximation and tries to take a holistic view of the optimization space across the stack.</span></p>
<p><span style="font-weight: 400;">Ultimately, as alluded to before, there is no single silver bullet, and like all open architecture problems, the answer is always &#8220;it depends&#8221;.  But in the era of Generative AI, it depends on whether we view sparsity and quantization as competing alternatives or as pieces of the same puzzle. Perhaps it’s time we stop asking which one is better, and start designing architectures flexible enough to embrace the realities of both.</span></p>
<p>&nbsp;</p>
<h3><b>About the Author:</b></h3>
<p><span style="font-weight: 400;">Sai Srivatsa Bhamidipati is a Senior Silicon Architect at Google working on the Google Tensor TPU in the Pixel phones. His primary focus is on efficient and scalable compute for Generative AI on the Tensor TPU.</span></p>
<h3><b>Authors’ Disclaimer:</b></h3>
<p><span style="font-weight: 400;">Portions of this post were edited with the assistance of AI models. Some references, notes and images were also compiled using AI tools. The content represents the opinions of the authors and does not necessarily represent the views, policies, or positions of Google or its affiliates.</span></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/950020673/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/950020673/0/sigarch-cat~To-Sparsify-or-To-Quantize-A-Hardware-Architecture-View/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">100754</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/from-the-editors-desk-2026-edition/</feedburner:origLink>
		<title>From the Editor&#8217;s Desk &#8211; 2026 Edition</title>
		<link>https://feeds.feedblitz.com/~/944524166/0/sigarch-cat~From-the-Editors-Desk-Edition/</link>
		<comments>https://feeds.feedblitz.com/~/944524166/0/sigarch-cat~From-the-Editors-Desk-Edition/#respond</comments>
		<pubDate>Tue, 03 Feb 2026 20:19:39 +0000</pubDate>
		<dc:creator><![CDATA[Dmitry Ponomarev]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Blog]]></category>
		<category><![CDATA[Editorial]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=96844</guid>
		<description><![CDATA[<div><img width="300" xheight="169" src="https://www.sigarch.org/wp-content/uploads/2026/02/AdobeStock_862939397-300x169.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>As we close the book on 2025, Computer Architecture Today has seen another successful year of community engagement. We published 29 posts covering a wide spectrum of topics—from datacenter energy-efficiency to the evolving debate on LLMs in peer review, alongside trip reports from our major conferences. I want to thank all our authors for their [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="169" src="https://www.sigarch.org/wp-content/uploads/2026/02/AdobeStock_862939397-300x169.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p>As we close the book on 2025, <i data-path-to-node="18,0" data-index-in-node="30">Computer Architecture Today</i> has seen another successful year of community engagement. We published 29 posts covering a wide spectrum of topics—from datacenter energy-efficiency to the evolving debate on LLMs in peer review, alongside trip reports from our major conferences. I want to thank all our authors for their insights, with special appreciation for those who contributed multiple times.</p>
<p>Over the last year, we shifted our editorial model, moving from a roster of set contributors to a more flexible, open-submission approach. We also re-established our conference trip reports, highlighting top architecture venues.</p>
<p data-path-to-node="14,2">The blog thrives on new voices, and our door is always open. We are actively looking for:</p>
<ul data-path-to-node="14,3">
<li>
<p data-path-to-node="14,3,0,0"><b data-path-to-node="14,3,0,0" data-index-in-node="0">New Ideas:</b> If you have a topic in mind, please propose it using <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.sigarch.org/contribute/propose-a-blog-post-topic/">this link</a> or email me directly.</p>
</li>
<li>
<p data-path-to-node="14,3,1,0"><b data-path-to-node="14,3,1,0" data-index-in-node="0">Trip Reports:</b> Planning to attend a conference? Volunteer to share your experience.</p>
</li>
<li>
<p data-path-to-node="14,3,2,0"><b data-path-to-node="14,3,2,0" data-index-in-node="0">Event Summaries:</b> Organizers of workshops or tutorials are welcome to publicize their events through summary posts.</p>
</li>
<li>
<p data-path-to-node="14,3,3,0"><b data-path-to-node="14,3,3,0" data-index-in-node="0">Industry Perspectives:</b> We would like to hear from our industry colleagues about their take on the future landscape of computer architecture.</p>
</li>
</ul>
<p data-path-to-node="14,4">Finally, as AI tools proliferate, the conversation around their role in our paper reviewing process is far from over. I look forward to seeing more of that debate here.</p>
<p data-path-to-node="14,4">Here’s to the new advances in Computer Architecture in 2026!</p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/944524166/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/944524166/0/sigarch-cat~From-the-Editors-Desk-Edition/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">96844</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/multi-agent-memory-from-a-computer-architecture-perspective-visions-and-challenges-ahead/</feedburner:origLink>
		<title>Multi-Agent Memory from a Computer Architecture Perspective: Visions and Challenges Ahead</title>
		<link>https://feeds.feedblitz.com/~/940946942/0/sigarch-cat~MultiAgent-Memory-from-a-Computer-Architecture-Perspective-Visions-and-Challenges-Ahead/</link>
		<comments>https://feeds.feedblitz.com/~/940946942/0/sigarch-cat~MultiAgent-Memory-from-a-Computer-Architecture-Perspective-Visions-and-Challenges-Ahead/#respond</comments>
		<pubDate>Tue, 20 Jan 2026 15:19:11 +0000</pubDate>
		<dc:creator><![CDATA[Zhongming Yu and Jishen Zhao]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Agents]]></category>
		<category><![CDATA[LLM]]></category>
		<category><![CDATA[Memory Consistency]]></category>
		<category><![CDATA[Memory Hierarchy]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=97929</guid>
		<description><![CDATA[<div><img width="300" xheight="200" src="https://www.sigarch.org/wp-content/uploads/2026/01/title-300x200.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>Large language model (LLM) agents are quickly moving from “single agent” to *multi-agent systems*: tool-using agents, planner-orchestrator, debate teams, specialized sub-agents that collaborate to solve tasks. At the same time, the *context* these agents must operate within is becoming more complex: longer histories, multiple modalities, structured traces, and customized environments. This combination creates a bottleneck [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="200" src="https://www.sigarch.org/wp-content/uploads/2026/01/title-300x200.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p>Large language model (LLM) agents are quickly moving from “single agent” to *multi-agent systems*: tool-using agents, planner-orchestrator, debate teams, specialized sub-agents that collaborate to solve tasks. At the same time, the *context* these agents must operate within is becoming more complex: longer histories, multiple modalities, structured traces, and customized environments. This combination creates a bottleneck that looks surprisingly familiar to computer architects: memory.</p>
<p>In computer systems, performance and scalability are often limited not by compute, but by memory hierarchy, bandwidth, and consistency. Multi-agent systems are heading toward the same wall — except their “memory” is not raw bytes, but semantic context used for reasoning. After dipping our heads building various LLM multi-agent frameworks over the past two years (e.g., <strong><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/fishmingyu/OrcaLoca">OrcaLoca</a> </strong>for software issue localization, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://stable-lab.github.io/MAGE/"><strong>MAGE</strong></a> for RTL design, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/stable-lab/Pro-V"><strong>Pro-V</strong></a> for RTL verification, and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://pettingllms-ai.github.io/"><strong>PettingLLMs</strong> </a>enabling RL training on multiple LLM agents), we would like to share our insights learned from our experience through the lens of a computer architect. This blog frames multi-agent memory as a <strong>computer architecture problem</strong>, proposes a simple architecture-inspired model, and highlights the key challenges and protocol gaps that define the road ahead.</p>
<p>While our perspectives are still preliminary and evolving, we hope they serve as a starting point to ignite a broader conversation.</p>
<hr />
<h2>Multi-Agent Memory Systems in Growing Complex Contexts</h2>
<h3>Why memory matters: Context is changing</h3>
<ul>
<li><strong>Longer context windows:</strong> Long-context evaluation suites like <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2404.06654"><strong>RULER</strong></a> and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://longbench2.github.io/"><strong>LongBench</strong></a> show that &#8220;real&#8221; long-context ability involves more than simple retrieval — it includes multi-hop tracing, aggregation, and sustained reasoning as length scales.</li>
<li><strong>Multi-modal inputs:</strong> Benchmarks such as <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://mmmu-benchmark.github.io/"><strong>MMMU</strong></a> (static images: charts, diagrams, tables) and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://video-mme.github.io/"><strong>VideoMME</strong></a>(videos with audio and subtitles) demonstrate that models must handle diverse visual modalities alongside text, extending beyond single-modality processing.</li>
<li><strong>Structured data &amp; traces:</strong> Text-to-SQL (e.g., <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://spider2-sql.github.io/"><strong>Spider</strong></a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://bird-bench.github.io/"><strong>BIRD</strong></a>) highlight that agents increasingly operate over structured, executable data — database schemas and generated SQL queries — rather than only raw chat history.</li>
<li><strong>Customized environments:</strong> In <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.swebench.com/SWE-bench/guides/evaluation/"><strong>SWE-bench</strong></a> and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://multi-swe-bench.github.io/#/"><strong>Multi-SWE-bench</strong></a>, models are evaluated by applying patches to real repositories and running tests in containerized (Docker) environments, making &#8220;environment state + execution&#8221; part of the memory problem. Similarly, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://webarena.dev/"><strong>WebArena</strong></a> and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://os-world.github.io/"><strong>OSWorld</strong></a> provide realistic, reproducible interactive environments that stress long-horizon state tracking and grounded actions.</li>
</ul>
<p><strong>Bottom line:</strong> Context is no longer a static prompt — it&#8217;s a dynamic, multi-format, partially persistent memory system.</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-97940" src="https://www.sigarch.org/wp-content/uploads/2026/01/motivation.jpg" alt="" width="11766" height="6266" /></p>
<hr />
<h2>Basic Prototypes: Shared vs. Distributed Agent Memory</h2>
<p>Before we talk about “hierarchies,” it helps to name the two simplest prototypes, which mirror classical memory systems.</p>
<h3>1) Shared Memory</h3>
<p>All agents access a shared memory pool (e.g., a shared vector store, shared document database).</p>
<ul>
<li><strong>Pros:</strong> Easy to share knowledge; fast reuse.</li>
<li><strong>Cons:</strong> Requires <strong>coherence support</strong>. Without coordination, agents overwrite each other, read stale info, or rely on inconsistent versions of shared facts.</li>
</ul>
<h3>2) Distributed Memory</h3>
<p>Each agent owns local memory (local scratchpad, local cache, local long-term store) and shares via synchronization.</p>
<ul>
<li><strong>Pros:</strong> Isolation by default; more scalable; fewer contention issues.</li>
<li><strong>Cons:</strong> Needs explicit <strong>synchronization</strong>; state divergence becomes common unless carefully managed.</li>
</ul>
<p>Most real systems sit somewhere in between: local working memory plus selectively shared artifacts.</p>
<hr />
<h2>An Agent Memory Architecture Inspired by Modern Computer Architecture Design</h2>
<p>Computer architecture teaches a practical lesson: you don’t build “one memory.” You build a <strong>memory hierarchy</strong> with different layers optimized for latency, bandwidth, capacity, and persistence.</p>
<p>A useful mapping for agents will be:</p>
<h3>Agent I/O Layer</h3>
<p><strong>What it is:</strong> Interfaces that ingest and emit information.</p>
<ul>
<li>Audio/speech</li>
<li>Text documents</li>
<li>Images</li>
<li>Network calls/web data</li>
</ul>
<p><strong>Analogy:</strong> Devices and I/O subsystems feeding the CPU.</p>
<h3>Agent Cache Layer</h3>
<p><strong>What it is:</strong> Fast, limited-capacity memory optimized for immediate reasoning.</p>
<ul>
<li>Compressed context</li>
<li>Recent trajectories and tool calls</li>
<li>Short-term latent storage (e.g., KV cache, embeddings of recent steps)</li>
</ul>
<p><strong>Analogy:</strong> CPU caches (L1/L2/L3): small, fast, and constantly refreshed.</p>
<h3>Agent Memory Layer</h3>
<p><strong>What it is:</strong> Large-capacity, slower memory optimized for retrieval and persistence.</p>
<ul>
<li>Full dialogue history</li>
<li>External knowledge databases (vector DBs, graph DBs, document stores)</li>
<li>Long-term latent storage</li>
</ul>
<p><strong>Analogy:</strong> Main memory + storage hierarchy.</p>
<p>This framing emphasizes a key principle: <strong>Agent performance is an end-to-end data movement problem</strong>. Even if the model is powerful, if relevant information is stuck in the wrong layer (or never loaded), reasoning accuracy and efficiency degrade.</p>
<p>And just like in hardware, caching is not optional. Similar to computer memory hierarchies, agent memory benefits from I/O and caching layers to improve efficiency and scalability.</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-97939" src="https://www.sigarch.org/wp-content/uploads/2026/01/Memprotocol.jpg" alt="" width="22116" height="14550" /></p>
<hr />
<h2>Protocol Extensions for Multi-Agent Scenarios</h2>
<p>Architecture layers need <em>protocols</em>. In multi-agent settings, protocols determine what can be shared, how fast, and under what rules.</p>
<p>Today, many agent frameworks rely on <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://blog.modelcontextprotocol.io/"><strong>MCP</strong> (Model Context Protocol)</a> as a connectivity layer. Agents registered via MCP can connect and communicate, but inter-agent bandwidth remains limited by message-passing. MCP largely uses JSON-RPC, so it’s best viewed as a protocol for <strong>agent context I/O</strong>: request/response, tool invocation, and structured messages.</p>
<p>That’s necessary — but not sufficient.</p>
<h3>Missing Piece 1: Agent Cache Sharing Protocol</h3>
<p>Many recent studies, such as <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2411.02820"><strong>DriodSpeak</strong></a> and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2510.03215"><strong>Cache to cache</strong></a>, explored KV cache sharing between LLM. However, we still lack a principled and unified protocol for sharing <em>cached artifacts</em> across agents.</p>
<p><strong>Goal:</strong> Enable one agent’s cached results to be transformed and reused by other agents.</p>
<p>In architecture terms, this is like enabling cache transfers or shared cache behavior — except the payload is semantic and may require transformation before reuse.</p>
<h3>Missing Piece 2: Agent Memory Access Protocol</h3>
<p>Although frameworks like <strong><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://docs.letta.com/">Letta</a></strong> and <strong><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://mem0.ai/">Mem0</a> </strong>support shared state within agent memory, the protocol defines how agents read/write each other’s memory is missing.</p>
<p><strong>Goal:</strong> Define memory access semantics: permissions, scope, and granularity.</p>
<p>Key questions:</p>
<ul>
<li>Can Agent B read Agent A’s long-term memory, or only shared memory?</li>
<li>Is access read-only, append-only, or read-write?</li>
<li>What is the unit of access: a document, a chunk, a key-value record, a “thought,” a trace segment?</li>
<li>Can we support “agent RDMA”-like patterns: low-latency direct access to remote memory without expensive message-level serialization?</li>
</ul>
<p>Without a memory access protocol, inter-agent collaboration is forced into slow, high-level message passing, which wastes bandwidth and loses structure.</p>
<hr />
<h2>The Next Frontier: Multi-Agent Memory Consistency</h2>
<p>The largest conceptual gap is <strong>consistency</strong>. The goal of <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://doi.org/10.1109/2.546611"><strong>memory consistency</strong></a> in computer architecture and systems design is to define constraints on the order of reads and writes to memory addresses. Consistency models (e.g., sequential consistency, TSO, and release consistency) clarify what behaviors programmers can rely on.</p>
<p>For agent memory, the goal shifts: It’s not about bytes at an address, but about maintaining a <strong>coherent semantic context</strong> that supports correct reasoning and coordination.</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-97941" src="https://www.sigarch.org/wp-content/uploads/2026/01/memory-consistency-comparison.jpg" alt="" width="20184" height="8150" /></p>
<h3>Why Agent Consistency Is Harder</h3>
<ul>
<li>The “state” is not a scalar value; it’s a <em>plan</em>, a <em>summary</em>, a <em>retrieval result</em>, a <em>tool trace</em>.</li>
<li>Writes are not deterministic; they may be speculative or wrong.</li>
<li>Conflicts aren’t simple write-write conflicts — they&#8217;re semantic contradictions.</li>
<li>Freshness depends on the environment state (repo version, API results, and permissions).</li>
</ul>
<h3>What a Multi-Agent Memory Consistency Layer Might Need</h3>
<p>A practical direction is to define consistency around the <em>artifacts agents actually share</em> — cached evidence, tool traces, plans, and long-term records — across both <strong>shared</strong> and <strong>distributed</strong> memory setups (often a hybrid: local caches + shared store). The layer should expose a <strong>consistency model</strong> (e.g., session, causal, eventual semantic, and stronger guarantees for “committed” outputs), provide richer <strong>communication primitives</strong> than plain message passing, and include <strong>conflict-resolution policies</strong> (source ranking, timestamps, consensus, and optional human intervention for high-stakes conflicts).</p>
<p>Research on this is still rare, but it is likely to become foundational — much like coherence and consistency were for multiprocessors.</p>
<h2>Conclusion</h2>
<p>Many agent memory systems today resemble <strong>human memory</strong> — informal, redundant, and hard to control — leaving a large opportunity for computer architecture researchers to rethink what “memory” should mean for agents <strong>at scale</strong>. To move from ad-hoc prompting to reliable multi-agent systems, we need <strong>better memory hierarchies</strong>, <strong>explicit protocols</strong> for cache sharing and memory access, and <strong>principled consistency models</strong> that keep shared context coherent.</p>
<h2>Acknowledgement</h2>
<p>We sincerely thank Wentao Ni, Hejia Zhang, Mingrui Yin, Jiaying Yang, and Yujie Zhao for their invaluable contributions through brainstorming, discussions, data collection, and survey work over the past few months. This article would not have been possible without their dedicated efforts.</p>
<p><b>About the authors:</b></p>
<p><i>Zhongming Yu is a PhD student in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research interests are in combining machine learning and computer systems, with a special focus on LLM agent system for machine learning systems, evolving ML and systems, and autonomous software engineering. </i></p>
<p><i>Jishen Zhao is a Professor in the</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. Her research spans and stretches the boundary across computer architecture, system software, and machine learning, with an emphasis on memory systems, machine learning and systems codesign, and system support for smart applications.</i></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/940946942/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/940946942/0/sigarch-cat~MultiAgent-Memory-from-a-Computer-Architecture-Perspective-Visions-and-Challenges-Ahead/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">97929</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/pipeorgan-modeling-memory-bandwidth-bound-executions-for-ai-and-beyond/</feedburner:origLink>
		<title>PipeOrgan: Modeling Memory-Bandwidth-Bound Executions for AI and Beyond</title>
		<link>https://feeds.feedblitz.com/~/940049756/0/sigarch-cat~PipeOrgan-Modeling-MemoryBandwidthBound-Executions-for-AI-and-Beyond/</link>
		<comments>https://feeds.feedblitz.com/~/940049756/0/sigarch-cat~PipeOrgan-Modeling-MemoryBandwidthBound-Executions-for-AI-and-Beyond/#respond</comments>
		<pubDate>Mon, 12 Jan 2026 15:00:20 +0000</pubDate>
		<dc:creator><![CDATA[Mark D. Hill]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Accelerators]]></category>
		<category><![CDATA[Memory]]></category>
		<category><![CDATA[Modelling]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=97568</guid>
		<description><![CDATA[<div><img width="300" xheight="200" src="https://www.sigarch.org/wp-content/uploads/2026/01/SIGARCH_PipeOrgan_via_ChatGPT_2026_01_05-300x200.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>TL;DR: Latency-tolerant architectures, e.g., GPUs, increasingly use memory/storage hierarchies, e.g., for KV Caches to speed Large-Language Model AI inference. To aid codesign of such workloads and architectures, we develop the simple PipeOrgan analytic model for bandwidth-bound workloads running on memory/storage hierarchies.  Background For three reasons, memory bandwidth, more than latency, limits AI inference performance. First, [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="200" src="https://www.sigarch.org/wp-content/uploads/2026/01/SIGARCH_PipeOrgan_via_ChatGPT_2026_01_05-300x200.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p><i><span style="font-weight: 400;">TL;DR: Latency-tolerant architectures, e.g., GPUs, increasingly use memory/storage hierarchies, e.g., for KV Caches to speed Large-Language Model AI inference. To aid codesign of such workloads and architectures, we develop the simple PipeOrgan analytic model for bandwidth-bound workloads running on memory/storage hierarchies. </span></i></p>
<h3><b>Background</b></h3>
<p><span style="font-weight: 400;">For three reasons, memory bandwidth, more than latency, limits AI inference performance. First, AI inference uses latency-tolerant compute engines, such as GPUs. Second, it principally uses hardware memory hierarchies to store a data structure called a <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://huggingface.co/blog/not-lain/kv-caching">Key-Value (KV) Cache</a> that holds information from recent queries to reduce redundant computation. With <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2309.06180">PagedAttention</a>, each KV Cache fetch obtains one or more multi-megabyte blocks (often called pages) that require substantial bandwidth to complete. Third, inference&#8217;s “decode” phase is memory-bound due to low arithmetic intensity, putting great pressure on memory bandwidth.</span></p>
<p><span style="font-weight: 400;">Traditional CPU memory/storage hierarchies are shaped by increasing latency, but designing hierarchies for AI workloads requires focusing on decreasing bandwidth. Since AI software is flexible, codesigning software and hardware is essential. </span></p>
<p><span style="font-weight: 400;">To provide intuition and first answer to the above questions, we next contribute the simple <em>PipeOrgan</em> analytic model for optimizing bandwidth-bound workloads running on a memory hierarchy with many parallel <em>pipes</em> from memories to compute. The PipeOrgan model shows that husbanding and providing bandwidth is important for AI software and hardware. Analytic models have long provided computing intuition, e.g., <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://en.wikipedia.org/wiki/Amdahl%27s_law">Amdahl&#8217;s Law</a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://en.wikipedia.org/wiki/Iron_law_of_processor_performance">Iron Law</a>, and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://en.wikipedia.org/wiki/Roofline_model">Roofline</a>.</span></p>
<p><img loading="lazy" decoding="async" class=" wp-image-97780 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2026/01/figure1.png" alt="" width="448" height="306" /></p>
<p><img loading="lazy" decoding="async" class="wp-image-97787 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2026/01/figure2-1.png" alt="" width="798" height="350" /></p>
<h3><b>Example System with Two Parallel Memories</b></h3>
<p><span style="font-weight: 400;">Let’s start simple. Consider the hardware depicted in Figure 1 with High Bandwidth Memory (HBM) with bandwidth 16 TB/s </span><b>in parallel with</b><span style="font-weight: 400;"> an LPDDR memory with bandwidth 0.5 TB/s. Assume for now that there are no transfers between memories, e.g., to cache. </span></p>
<p><span style="font-weight: 400;">Using the PipeOrgan math from the next section, Figure 2’s blue line shows how system performance changes depending on what percentage of data comes from LPDDR memory. (The orange line comes later when we add caching.) </span>Performance is highest when LPDDR provides exactly 3% of the data <span style="font-weight: 400;">(arrow 1)</span>, which matches its 3% bandwidth <span style="font-weight: 400;">(0.5/(16.0+0.5))</span>. At this point, both LPDDR and HBM memories finish transferring data at the same time, so they act as co-bottlenecks and the system runs at peak efficiency.</p>
<p>When less than 3% of data is from LPDDR (left of the peak), <span style="font-weight: 400;">HBM finishes last and limits performance. When LPDDR sources more than 3% (right of the peak), it is</span> the bottleneck. LPDDR might have to source more data, because <span style="font-weight: 400;">HBM&#8217;s limited capacity, currently 48-64GB per stack, may preclude it from being able to source its share (97%). If so, </span><span style="font-weight: 400;">performance drops quickly: 4% from LPDDR gives 76% of peak (arrow 2), and 20% yields just 15% (arrow 3).</span></p>
<p><span style="font-weight: 400;">However, future AI systems will feature <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/pdf/2407.00079">multiple memory and storage levels</a>, using HBM, LPDDR, host DDR, pooled DDR, and attached or pooled FLASH storage</span><span style="font-weight: 400;">.</span></p>
<p><img loading="lazy" decoding="async" class=" wp-image-97782 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2026/01/figure3.png" alt="" width="570" height="314" /></p>
<h3><b>PipeOrgan Model of Systems with N Parallel Memories</b></h3>
<p><span style="font-weight: 400;">The above result generalizes to an N-level memory/storage hierarchy with each level feeding compute in parallel. Optimal performance is achieved when all parallel memories complete a workload phase simultaneously, leading to this PipeOrgan principle:</span></p>
<p><b><i>Memory-bandwidth-bound workloads perform best when data is sourced from each memory level in proportion to its bandwidth.</i></b></p>
<p><strong>Proof: </strong></p>
<ol>
<li>Let each memory provide bandwidth b_i TB/s in parallel for total bandwidth B = b_1 + … + b_N.</li>
<li><span style="font-weight: 400;">For a workload, let each source d_i bytes in parallel for total data transferred D = d_1 + … + d_N.</span></li>
<li><span style="font-weight: 400;">By assumption, the workload is limited by data transfer time with compute hidden.</span></li>
<li><span style="font-weight: 400;">Time for each memory to finish its data transfer is d_i/b_i  = TB/(TB/s) = seconds.</span></li>
<li><span style="font-weight: 400;">Workload Time is the maximum of all memories finishing: MAX [d_1/b_1, …, d_N/b_N].</span></li>
<li><span style="font-weight: 400;">Workload Performance = 1/ Time = MIN[b_1/d_1, …, b_N/d_N].</span></li>
<li><span style="font-weight: 400;">Set each d_i = (D/B)*b_i = proportional to its bandwidth b_i.</span></li>
<li><span style="font-weight: 400;">Performance = MIN[b_1/((D/B)*b_1), …, b_N/((D/B)*b_N)].</span></li>
<li><span style="font-weight: 400;">Performance = MIN[(B/D), …, (B/D)] = B/D and Time = 1/Performance = D/B. </span></li>
</ol>
<p><span style="font-weight: 400;">This makes sense: PipeOrgan shows that best performance occurs when one moves all the data using all the bandwidth with no bandwidth idling.</span></p>
<p><img loading="lazy" decoding="async" class="wp-image-97783 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2026/01/figure4.png" alt="" width="555" height="263" /></p>
<h3><b>But Caching Is Critical</b></h3>
<p><span style="font-weight: 400;">The PipeOrgan version above assumes all data goes directly to compute, without transfers among memories. In reality, systems move data from lower- to higher-bandwidth memories, caching it for reuse. For a two-level system (see Figure 4), assume the entire fraction of the workload’s data used from LPDDR is first transferred to HBM for caching (orange arrow). Let the data used from LPDDR be f*D where f ranges from 0 to 1.</span></p>
<ul>
<li><span style="font-weight: 400;">Performance with caching = MIN[(b_1/D)/(f+1), b_2/(f*D)] = MIN[limited by HBM BW, limited by LPDDR BW].</span></li>
</ul>
<p><span style="font-weight: 400;">Figure 2 shows an orange curve for caching that is hidden under the original blue curve when more than 3% of data is sourced from LPDDR. At more than 3% from LPDDR, performance&#8211;without and with caching&#8211;is limited by the time to transfer needed data with the same limited LPDDR bandwidth.</span></p>
<p><span style="font-weight: 400;">While it might look like caching </span><span style="font-weight: 400;">doesn&#8217;t matter, caching is actually important. </span><span style="font-weight: 400;">This is because caching can greatly shift a workload’s x-axis operating point. For example, sourcing 20% of data from LPDDR yields 15% of peak performance (arrow 3). If LPDDR data is cached in HBM and reused five times, then–as the orange dashed arrow shows–only 4% comes from LPDDR and performance gets boosted to 76% of peak—a ~5x improvement (arrow 2).</span></p>
<p><span style="font-weight: 400;">Consequently, caching remains critical. Moreover, PipeOrgan and its N parallel memory principle also applies bandwidth-bound workloads once caching&#8217;s more complex information flows are accounted for.</span></p>
<h3><b>Implications, Limitations and Future Work</b></h3>
<p><span style="font-weight: 400;">Statistician George Box famously said, “</span><i><span style="font-weight: 400;">Essentially, all models are wrong, but some are useful.</span></i><span style="font-weight: 400;">” </span></p>
<p><span style="font-weight: 400;">We conjecture that the PipeOrgan model is useful for AI codesign, especially in the early stages and with software people having less hardware understanding. </span><b>Its key implication is that bandwidth-bound workloads must carefully manage bandwidth from larger, slower memories and storage. </b><span style="font-weight: 400;">While vast data can be stored statically, dynamic use from low-bandwidth memories should remain modest.</span></p>
<p><span style="font-weight: 400;">Three PipeOrgan limitations motivate future work. First, most workloads aren’t bandwidth bound throughout, and PipeOrgan doesn’t address other phases. Modeling these requires more parameters, increasing accuracy but also complexity.</span></p>
<p><span style="font-weight: 400;">Second, the caching model variant only covers two memory levels and always transfers data first to the higher-bandwidth level before use. Future work should extend this to N memory levels and more advanced caching policies. Modeling the many options for caching may be challenging.</span></p>
<p><span style="font-weight: 400;">Third, PipeOrgan may need to be extended for systems that do some processing in or near the memories themselves rather than moving all data to a segregated compute unit.</span></p>
<p><i><span style="font-weight: 400;"><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.cs.princeton.edu/courses/archive/fall13/cos375/Burks.pdf">Burks, Goldstine, &amp; von Neumann, 1946</a>: We are therefore forced to recognize the possibility of constructing a hierarchy of memories, each of which has greater capacity than the preceding but which is less quickly accessible.</span></i></p>
<p><span style="font-weight: 400;">In sum, after eight decades of memory hierarchies focused mostly on latency, we are now at the exciting early stages of codesigning bandwidth-focused memory/storage hierarchies for more flexible AI software.</span></p>
<p><b>About the Author:</b><span style="font-weight: 400;"><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://pages.cs.wisc.edu/~markhill/"> Mark D. Hill</a> is John P. Morgridge Professor and Gene M. Amdahl Professor Emeritus of Computer Sciences at the University of Wisconsin-Madison and consultant to industry. He initiated the PipeOrgan model consulting for Microsoft and was given permission to release it. He is a fellow of AAAS, ACM, and IEEE, as well as recipient of the 2019 Eckert-Mauchly Award.</span></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/940049756/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/940049756/0/sigarch-cat~PipeOrgan-Modeling-MemoryBandwidthBound-Executions-for-AI-and-Beyond/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">97568</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/in-memoriam-remembering-mike-flynn/</feedburner:origLink>
		<title>In Memoriam: Remembering Mike Flynn</title>
		<link>https://feeds.feedblitz.com/~/939763391/0/sigarch-cat~In-Memoriam-Remembering-Mike-Flynn/</link>
		<comments>https://feeds.feedblitz.com/~/939763391/0/sigarch-cat~In-Memoriam-Remembering-Mike-Flynn/#respond</comments>
		<pubDate>Tue, 06 Jan 2026 21:00:08 +0000</pubDate>
		<dc:creator><![CDATA[Ruby B. Lee, Charlie Neuhauser, Timothy M. Pinkston]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Memoriam]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=97727</guid>
		<description><![CDATA[<div><img width="257" xheight="300" src="https://www.sigarch.org/wp-content/uploads/2026/01/Picture1.jpg" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>Michael J. Flynn is a widely respected contributor—indeed a giant—in the field of Computer Architecture.  He made highly significant and impactful contributions throughout his career, both in industry and in academia.  Sadly, he passed away peacefully December 24, 2025, having lived a long and full life. Born May 20, 1934, in New York, NY, Flynn [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="257" height="300" src="https://www.sigarch.org/wp-content/uploads/2026/01/Picture1.jpg" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p style="font-weight: 400;">Michael J. Flynn is a widely respected contributor—indeed a <em>giant</em>—in the field of Computer Architecture.  He made highly significant and impactful contributions throughout his career, both in industry and in academia.  Sadly, he passed away peacefully December 24, 2025, having lived a long and full life.</p>
<p style="font-weight: 400;">Born May 20, 1934, in New York, NY, Flynn earned his Bachelor’s, Master’s, and Ph.D. degrees in Electrical Engineering from Manhattan College (1955), Syracuse University (1960), and Purdue University (1961), respectively, and he received an honorary Doctor of Science degree from the University of Dublin (1998).  After ten years as a design engineer and project manager at IBM (1955-65, in Endicott and Poughkeepsie, NY), he became a member of the faculty at the University of Illinois at Chicago (1965-1966), Northwestern University (1966-1970), and Johns Hopkins University (1970-1975) before joining Stanford University in 1975 as Professor of Electrical Engineering.  He taught internationally, in Ireland, other places in Europe, Singapore, and Japan.</p>
<p style="font-weight: 400;">As a young project manager at IBM, Flynn was responsible for the design of the well-known <em>IBM System 360 (Models 91/92/95 series)</em>, the first computer to implement the sophisticated Tomasulo algorithm, along with many other groundbreaking high-performance architectural techniques.  As the first family of general-purpose computer mainframes that featured <em>architectural compatibility</em> for both commercial and scientific applications, the System 360 is widely recognized as revolutionizing computing during that time—and in many ways persisting even today.  Indeed, many of the high-performance computing techniques developed by Flynn and his IBM colleagues are used throughout the industry today, having migrated from barn-sized mainframes to finger-nail sized microprocessor chips.  Flynn also was the first to shed light on the performance potential and limitations of parallel computers with what’s become known as <em>Flynn’s classification </em>(or <em>Flynn’s taxonomy</em>), a pioneering framework for categorizing parallelism in computer architectures based on the number of simultaneous instruction streams and data streams they handle, e.g., SISD, SIMD, MISD, and MIMD.  His original taxonomy is still used widely today, with various extensions derived from it, to distinguish between different kinds of parallel processor computer systems.</p>
<p style="font-weight: 400;">In 1972, together with some colleagues from IBM, Flynn co-founded Palyn Associates which provided consulting services in the field of high-performance computer architecture and design.  For more than 30 years, he and his colleagues advised nearly every major computer company in Japan, Europe and the United States, including IBM, CDC, Fujitsu, Hitachi, Honeywell Bull, and ICL.  Later, he played a prominent role in Maxeler, products of which made use of advanced dataflow techniques to provide high performance processing for specific applications, such as automated trading.  As a renowned professor at Stanford until his retirement in 1999 and transition to emeritus status, Flynn made seminal contributions to instruction set architecture (ISA), computer arithmetic, advanced floating-point design, multimedia, parallel processors and interconnects, emulation, and performance evaluation, to name a few.  He (co-)authored several textbooks, including <u>Introduction to Arithmetic for Digital Systems Designers</u>, <u>Computer Architecture: Pipelined and Parallel Processor Design</u>, and <u>Advanced Computer Arithmetic Design</u>. An IEEE Fellow, ACM Fellow, and Fellow of the Institution of Engineers of Ireland, Flynn received numerous other honors and awards for his impactful technical contributions, including the ACM/IEEE Eckert-Mauchly Award (1992), IEEE Computer Society’s (CS) Harry Goode Memorial Award and Medal (1995), the Tesla Award and Medal from the International Tesla Society in Belgrade (1998), IEEE CS Charles Babbage Award, IEEE CS Computer Pioneer Award (2015, his acceptance speech video is <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.youtube.com/watch?v=xAhRYUPSZKM">here</a>), and many others.</p>
<p style="font-weight: 400;">Notably, when the field of computer architecture was still in its infancy more than fifty years ago, Flynn founded the IEEE CS Technical Committee on Computer Architecture (TCCA) and ACM’s Special Interest Group on Computer Architecture (SIGARCH); he also started the ACM/IEEE International Symposium on Computer Architecture (ISCA), co-sponsored by both, which is among the most prestigious flagship computer architecture conferences in the world.  At ISCA’s 50<sup>th</sup> anniversary conference at FCRC 2023, Flynn was invited to give a “<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://u.pcloud.link/publink/show?code=XZro5EVZFqb2N6FIwjuqKDkbaDonqJzVoXzX">50 Year Retrospective Lecture</a>” and was given an honorary plaque with these words inscribed: <em>&#8220;In recognition, with tremendous gratitude, of your lifetime dedication and leadership to the computer architecture community on this the 50<sup>th</sup> anniversary of your founding of ISCA, SIGARCH, and TCCA.&#8221;</em></p>
<p style="font-weight: 400;">Even more than his impressive technical contributions, which are many, Flynn is remembered fondly by the many dozens of doctoral graduate students he advised—for his unending kindness, wealth of wisdom, caring tutelage, gentle encouragement, constant motivation, and enduring support, especially when most needed.  He treated each and every student as if they were a member of his own family, and he was viewed by them not only as their academic “father,” but referred to affectionately as “the Great Man.”  Many of his former mentees returned to Stanford several times each year for luncheons to enjoy his company and reminisce about exciting times working with him in tackling some of the most compelling technical issues of the day.</p>
<p style="font-weight: 400;">Flynn was an equally generous mentor to his junior faculty colleagues, helping them establish their careers and providing sage advice as they made their way.  Kunle Olukotun attests to this: <em>“Meeting Mike Flynn near the end of my Ph.D. at the University of Michigan changed the trajectory of my career.  At the time, I was firmly on a path toward industry, but Mike believed that I could be a strong academic, and he encouraged me to apply to Stanford.  Mike saw something in me that I did not yet see in myself, and that confidence made an enduring difference. Once I arrived at Stanford, Mike served as my mentor. He helped me navigate the academic waters with thoughtful and wise advice, provided opportunities to showcase my research, and supported me through nominations for awards and professional recognition.  I am deeply grateful to Mike for all he did to help establish my career, and for the role he played in the success of so many other junior colleagues whom he mentored with the same generosity and vision.  I am deeply saddened by his passing.”</em>  Similar sentiments are echoed by Bill Dally, who shares the following: <em>“I first met Mike as a graduate student at Stanford in 1980.  I was awed by his accomplishments and his understanding of parallel computing. He kindled my interest in parallel computing which launched me on a very successful career.  Later, when I came to Stanford as a faculty member in 1997, I found Mike to be a great source of advice about Stanford, being a faculty member, research strategy, and many other topics.  I am deeply saddened to hear of Mike&#8217;s passing.  He will be greatly missed.”</em>  Another of his faculty colleagues at Stanford, Christos Kozyrakis, recalls the following: <em>“One of the most memorable moments of my early teaching years was hosting him in class to discuss the Flynn taxonomy of computer architecture—a special experience for both the students and myself and a vivid reminder of the lasting impact of his work.”</em>  Indeed, Mike Flynn was highly respected and revered by fellow colleagues all throughout his professional career.  Solemnly noted by John L. Hennessy, <em>“Mike was the person who hired me at Stanford, gave me some of my first research funding, jointly published an early paper with me, and gave me my first consulting opportunity.  Sadly, his passing marks the end of an important era in computing: Mike was the last of the great System 360 pioneers—Gene Amdahl, Bob Evans, Fred Brooks, Eric Bloch, Gerry Blaauw, and Robert Tomasulo—all are now gone.”</em></p>
<p style="font-weight: 400;">He was a wonderful human being.</p>
<p style="font-weight: 400;">Professor Michael J. Flynn will be sorely missed by his loving family as well as by his extended academic family and all those whose lives he has indelibly touched over his blessed ninety-one plus years.  May he rest blissfully in peace, and may his venerable legacy be inspirational and long lasting.  Fittingly, through Mike Flynn’s final public words to all of us in the computer architecture community in his ISCA 50<sup>th</sup> Anniversary Lecture, he exhorted us all by saying: <em>“Now it’s your turn!”</em></p>
<p style="font-weight: 400;"><em><strong>About the Authors:</strong> </em></p>
<p><strong>Ruby B. Lee</strong> is the Forest G. Hamrick Professor Emeritus in the ECE department at Princeton University, and chief architect at Hewlett-Packard in Silicon valley before that. She is a Fellow of the IEEE, ACM and the American Academy of Arts and Sciences, and recipient of awards such as the most Influential Paper award in 20 years at ISCA 2025 and the Test of Time award at the ACSAC 2024 security conference. Her research combines cyber security, computer architecture and deep learning, including secure processor and cache architectures, attacks and defenses, low-cost AI and multimedia.</p>
<p><strong>Charlie Neuhauser</strong> is now retired after more than 50 years in the field of computer design and analysis.  During the latter half of his career, he provided technical insight to attorneys and companies in the area of intellectual property.  He is currently the registration chair for the IEEE Hot Chips Symposium.</p>
<p><strong>Timothy M. Pinkston</strong> is the George Pfleger Chaired Professor of Electrical and Computer Engineering at the University of Southern California and also is a Vice Dean in USC’s Viterbi School of Engineering.  A Fellow of AAAS, ACM, and IEEE, and recipient of the ACM SIGARCH Alan D. Berenbaum Distinguished Service Award, Timothy’s research contributions mainly are in the area of interconnection networks and efficient data movement in parallel computing systems.</p>
<p style="font-weight: 400;">All three authors are former  Ph.D. students of Mike Flynn at Stanford (Lee and Pinkston) and Johns Hopkins (Neuhauser).</p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/939763391/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/939763391/0/sigarch-cat~In-Memoriam-Remembering-Mike-Flynn/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">97727</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/microarchitectural-modeling-in-the-era-of-accelerator-rich-systems-and-computing-at-scale/</feedburner:origLink>
		<title>Microarchitectural Modeling in the Era of Accelerator-Rich Systems and Computing at Scale</title>
		<link>https://feeds.feedblitz.com/~/932315468/0/sigarch-cat~Microarchitectural-Modeling-in-the-Era-of-AcceleratorRich-Systems-and-Computing-at-Scale/</link>
		<comments>https://feeds.feedblitz.com/~/932315468/0/sigarch-cat~Microarchitectural-Modeling-in-the-Era-of-AcceleratorRich-Systems-and-Computing-at-Scale/#respond</comments>
		<pubDate>Mon, 08 Dec 2025 15:00:26 +0000</pubDate>
		<dc:creator><![CDATA[Dimitris Gizopoulos]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[AI accelerator]]></category>
		<category><![CDATA[Microprocessor]]></category>
		<category><![CDATA[Modeling]]></category>
		<category><![CDATA[Reliability]]></category>
		<category><![CDATA[Simulators]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=96032</guid>
		<description><![CDATA[<div><img width="300" xheight="168" src="https://www.sigarch.org/wp-content/uploads/2025/12/AdobeStock_1023668818-300x168.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>Microarchitecture simulators have been conceived and implemented to be valuable tools for the design of computing chips of all types (SimpleScalar, gem5, SMTSIM, Sniper, Qflex, Scarab, GPGPU-sim, Accel-Sim, Multi2Sim, NaviSim, SCALE-sim, gem5-Salam, TAO, PyTorchSim – the list is neither historically complete nor updated). In essence, microarchitecture simulators have an “impossible” objective: to model and measure [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="168" src="https://www.sigarch.org/wp-content/uploads/2025/12/AdobeStock_1023668818-300x168.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p><strong>Microarchitecture simulators</strong> have been conceived and implemented to be valuable tools for the design of computing chips of all types (<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/982917">SimpleScalar</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/10.1145/2024716.2024718"> gem5</a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cseweb.ucsd.edu/~tullsen/smtsim.html">SMTSIM</a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/snipersim/snipersim">Sniper</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://infoscience.epfl.ch/entities/publication/a829a51b-bc59-445e-9777-57ee89c83bce"> Qflex</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/litz-lab/scarab"> Scarab</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/4919648"> GPGPU-sim</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/9138922"> Accel-Sim</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.multi2sim.org/"> Multi2Sim</a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/abs/10.1145/3559009.3569666">NaviSim</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/9238602"> SCALE-sim</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/9251937"> gem5-Salam</a>,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/10.1145/3656012"> TAO</a>, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dl.acm.org/doi/full/10.1145/3725843.3756045">PyTorchSim</a> – <em>the list is neither historically complete nor updated</em>). In essence, microarchitecture simulators have an “impossible” objective: to model and measure properties of a chip (and a complete system built on it) both quickly and accurately; a contradiction in terms. They try to approach this goal operating at an abstraction layer which allows fast exploration of a large space of options but, still, remaining relatively close to the actual hardware they model; at least at a cycle-level. The objective is the same across different computing units, CPUs, GPUs, DSAs (Domain-Specific Accelerators), including, of course, AIAs (AI accelerators).</p>
<p>Trace based simulation, event-driven simulation, and, more recently, machine-learning based simulation approaches are employed for the same purpose: explore and rank-order the designers’ ideas about the microarchitecture in terms of important system properties – primarily performance, but also power and energy consumption, reliability, and security. Among hundreds or thousands of different microarchitecture and software combinations, the architects and designers are suffocating. Microarchitecture simulation narrows down this space to a handful of configurations that can subsequently be analyzed at lower (finer) abstraction layers.</p>
<p>This short blog post aims to simplistically quantify the challenges of the role of microarchitecture simulators in today’s landscape.</p>
<h4><b>The Simulation Throughput Bet</b></h4>
<p>Assuming that simulation at any abstraction layer can be parallelized similarly, we compare single simulation runs at each level. A single workload run on a simulated or real CPU, GPU, or AIA has a throughput of approximately (IPS = instructions/operations per second):</p>
<ul>
<li>      &lt; 1 Kilo-IPS simulated at the gate level</li>
<li>      10 – 50 Kilo-IPS simulated at the register-transfer (boosted to 5 – 10 Mega-IPS when<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://fires.im/"> FPGA-accelerated simulation</a> is employed)</li>
<li>      0.3 – 1 Mega-IPS simulated at the microarchitecture level</li>
<li>      1 – 3 Giga-IPS running at real silicon</li>
</ul>
<p>(<em>Absolute numbers may vary depending on the host system running simulations, but relative differences are close to the above</em>.)</p>
<p>The closest representation of the real physical system in the list is the gate level (if one wants to descend further down, it can be the transistor level). Using the above approximate numbers, a short 10-second runtime workload on final silicon needs about 1 year of gate-level simulation. The same run at the microarchitecture level takes less than a week!</p>
<p>Let’s now assume a simple design space exploration case, which only involves:</p>
<ul>
<li>10 workloads of about the above short duration each</li>
<li>20 different microarchitecture points (counts, sizes, and organizations of registers, buffers, queues, caches, arithmetic units)</li>
<li>5 compilation options</li>
</ul>
<p>Exploring the design space using gate-level simulation (bravely assuming the entire workloads can run at this level) would require 1000 years of simulation time. If 1000 servers were available for simulation, this time would be reduced again to only (!) 1 year. At the microarchitecture level, it is again a matter of only a few days!</p>
<p>The space that architects and designers need to explore does not consist of only 10 workloads, 20 microarchitecture points, and 5 compilation options. Microarchitecture-level simulation reduces years of simulation studies for design exploration down to days.</p>
<h4><b>Do Accelerator-rich Designs Make a Microarchitecture Simulator’s Life Easier?</b></h4>
<p>Domain-specific accelerators (DSAs) or particularly Artificial intelligence accelerators (AIAs) are very fast and energy-efficient in the few tasks they are dedicated to serve. Does the small number of tasks mean that the design space to explore is smaller or at least more manageable than CPUs or GPUs? Most probably not. The AI accelerators design space is very large because of the rapidly evolving and expanding set of ML algorithms. As a result, the accelerator chip designs often undergo major changes. For example, in systolic array based AIAs, at least the following design parameters must be evaluated in simulation at design time:</p>
<ul>
<li>Dimensions of the systolic array</li>
<li>Data flow of the systolic array (input, output, weight stationary)</li>
<li>Data type of processing elements (short or long integers or floating-point numbers)</li>
<li>Level of memory hierarchy the AIA is connected to</li>
</ul>
<p>Because of the rapid changes in ML/AI algorithms used by AIAs for training and inference, their “useful” life span is and will likely remain much shorter than that of established general purpose CPU and GPU architectures, where the microarchitecture knobs are also plentiful, but well understood. Therefore, the effort put into an AIA design may see a much shorter useful production time compared to a CPU or a GPU. A new design space exploration phase will need to be performed for the new generation of AIAs designed for the new generation of ML/AI algorithms.</p>
<h4><b>Beyond Performance Exploration – Resilience at Scale</b></h4>
<p>Microarchitecture simulators have been employed for resilience analysis of<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/7482075"> CPUs</a> and<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/7482077"> GPUs</a> for almost a decade now. Recently, they have been proven very effective in Silent Data Corruption (<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/11014262">SDCs</a>) research to automatically generate<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/10609719"> functional test programs for CPUs</a>, to <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/abstract/document/10946774">demystify the actual rate</a> of <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.sigarch.org/sdcs-a-b-c/">SDCs</a> generated by CPUs at datacenter scale,<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/11117132"> accurately analyze SDCs for AIAs</a> and contribute to<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~micro58.tutorial.di.uoa.gr/"> decision making for fault protection</a> of AIA-based systems.</p>
<p>Resilience (and SDCs) analysis for large-scale AI systems (recently<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.opencompute.org/documents/sdc-in-ai-ocp-whitepaper-final-pdf"> openly recognized</a> by the OCP group of companies and the <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.computer.org/digital-library/magazines/mi/cfp-silent-data-corruptions">research community</a> as a critical problem) adds extra dimensions to the design space exploration arena (these dimensions exist for CPUs and GPUs resilience analysis too):</p>
<ul>
<li>Type of silicon defects to analyze (fault models)</li>
<li>Methods to enhance the resilience of the systems (protection)</li>
<li>Scale of AI systems</li>
</ul>
<p>For modeling defects and faults, the main challenge is the enhancement of microarchitecture simulators for AIAs with mechanisms for the accurate representation of silicon defect roots and behaviors (such as transient, permanent, or delay faults) and the necessary physical conditions that excite them (temperature, droops, etc.).</p>
<p>Protection schemes represent yet another design knob. When a redundancy technique (at space, time, or information) is employed to detect and correct hardware faults, it alters the design of the microarchitecture, software, or both, adding overheads in terms of area, power and performance. Applying  simple calculations presented above in this context points to further expansion of the design space that the simulator is called to explore.</p>
<p>Increasing system scale of datacenters or HPC clusters (datacenters for “usual” or ML/AI workloads, HPC systems for scientific computing or ML/AI workloads) exacerbates the problem of device variability – chips that are designed to be equal but operate with variations. The scale of deployment of AIA chips, CPUs and GPUs, coupled with diversity of workloads, increases the number of failure mechanisms and scenarios that can happen in the field, but were never imagined at chip design or manufacturing time. Yet another dimension to analyze.</p>
<p>Microarchitecture modeling and simulation for joint performance and resilience exploration is now an integral part of brave development activities around the globe; for one, the<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://dare-riscv.eu/"> DARE project</a>, Europe’s most ambitious endeavor of all times in HPC and AI computing, heavily relies on microarchitectural simulation for all three different computing engines it builds (a high-performance general-purpose microprocessor, an aggressive vector processor, and an AI inference engine) to make design decisions for performance, power, and resilience. The expected FIT (failures-in-time) and SDCs rates when the designed chips are deployed at scale, will be estimated using microarchitecture simulation and protection schemes will be diligently implemented.</p>
<h4><b>The Need for Validation</b></h4>
<p>Like any abstraction, microarchitecture-level modeling is constantly questioned for the validity of its findings vs. simulation of more detailed layers it abstracts (there are examples for gem5 validation for <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/6844457">performance measurements</a> and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/8809532">reliability measurements</a> compared to physical systems and experiments). Validation of simulators is among the most important and useful results expected by the computer architecture and systems research community as our computing systems increase in complexity and scale everywhere. Simulators are extremely important, but we need to continuously tune them to model, as accurately as possible, the properties of physical computing systems and behaviors of modern software stacks.</p>
<p><b>About the Author: </b>Dimitris Gizopoulos is Professor of Computer Architecture at the University of Athens. His research team (Computer Architecture Lab) focuses on modeling, evaluating, and improving the performance, dependability, and energy-efficiency of computing systems based on CPUs, GPUs, and AIAs.</p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/932315468/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/932315468/0/sigarch-cat~Microarchitectural-Modeling-in-the-Era-of-AcceleratorRich-Systems-and-Computing-at-Scale/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">96032</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/the-hitchhikers-guide-to-coherent-fabrics-5-programming-rules-for-cxl-nvlink-and-infinityfabric/</feedburner:origLink>
		<title>The Hitchhiker&#8217;s Guide to Coherent Fabrics: 5 Programming Rules for CXL, NVLink, and InfinityFabric</title>
		<link>https://feeds.feedblitz.com/~/930748850/0/sigarch-cat~The-Hitchhikers-Guide-to-Coherent-Fabrics-Programming-Rules-for-CXL-NVLink-and-InfinityFabric/</link>
		<comments>https://feeds.feedblitz.com/~/930748850/0/sigarch-cat~The-Hitchhikers-Guide-to-Coherent-Fabrics-Programming-Rules-for-CXL-NVLink-and-InfinityFabric/#respond</comments>
		<pubDate>Mon, 01 Dec 2025 14:59:17 +0000</pubDate>
		<dc:creator><![CDATA[Zixuan Wang, Suyash Mahar, Luyi Li, Jangseon Park, Jinpyo Kim, Theodore Michailidis, Yue Pan, Mingyao Shen, Tajana Rosing, Dean Tullsen, Steven Swanson, Jishen Zhao]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[AlphaFold]]></category>
		<category><![CDATA[CXL]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Heterogeneous Systems]]></category>
		<category><![CDATA[Memory]]></category>
		<category><![CDATA[Profiling]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=95299</guid>
		<description><![CDATA[<div><img width="300" xheight="153" src="https://www.sigarch.org/wp-content/uploads/2025/11/image7-300x153.png" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>This is the second article in the series, following our first blog in Dec 2023: https://www.sigarch.org/tuning-the-symphony-of-heterogeneous-memory-systems/ Modern applications are increasingly memory hungry. Applications like Large-Language Models (LLM), in-memory databases, and data analytics platforms often demand more memory bandwidth and capacity than what a standard server CPU can provide. This leads to the development of coherent [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="153" src="https://www.sigarch.org/wp-content/uploads/2025/11/image7-300x153.png" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><blockquote><p>This is the second article in the series, following our first blog in Dec 2023:
<br>
<a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.sigarch.org/tuning-the-symphony-of-heterogeneous-memory-systems/">https://www.sigarch.org/tuning-the-symphony-of-heterogeneous-memory-systems/</a></p></blockquote>
<p>Modern applications are increasingly memory hungry. Applications like Large-Language Models (LLM), in-memory databases, and data analytics platforms often demand more memory bandwidth and capacity than what a standard server CPU can provide. This leads to the development of coherent fabrics to interconnect more memory with cache coherence support, such that workloads can benefit from large memory capacity with ideally not much effort to modify the code.</p>
<p>But is it truly trivial for workloads to adopt such heterogeneous memory systems? Are there any hidden caveats behind the pipedreams of future memory systems? And if one decided to build such a heterogeneous memory system, what are some important factors to consider before building a cluster of such systems?</p>
<p>To answer such questions, <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2411.02814">we studied a wide range of coherent fabrics</a>, namely Compute Express Link (CXL), NVLink Chip-to-Chip (NVLink-C2C), and AMD’s InfinityFabric. In terms of architecture reverse engineering, performance characterizations, and performance implications to emerging workloads.</p>
<p>This work is made possible with 8 PhD students from 4 UCSD research groups working over 1.5 years, with wide industrial collaborations. Together we measured 13 server systems, across 3 CPU vendors with a wide range of CPU generations, 3 types coherent fabric links, 5 device vendors, and multiple system configurations (including local vs. remote NUMA, Sub-NUMA clustering mode, interleaved mode, etc.).</p>
<p><img loading="lazy" decoding="async" class="alignnone size-full wp-image-95303" src="https://www.sigarch.org/wp-content/uploads/2025/11/image1.png" alt="" width="1734" height="630" srcset="https://www.sigarch.org/wp-content/uploads/2025/11/image1.png 1734w, https://www.sigarch.org/wp-content/uploads/2025/11/image1-1280x465.png 1280w, https://www.sigarch.org/wp-content/uploads/2025/11/image1-980x356.png 980w, https://www.sigarch.org/wp-content/uploads/2025/11/image1-480x174.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) and (max-width: 1280px) 1280px, (min-width: 1281px) 1734px, 100vw" /></p>
<p>We have <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://arxiv.org/abs/2411.02814">shared our detailed observations on ArXiV</a>, and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/awesome-cxl/heimdall">open sourced our benchmarking suite on GitHub</a> which runs on all the above-mentioned systems even including non-CXL systems like NVIDIA GH200. This blog article focuses on CXL, which has recently witnessed a broad attention. Please refer to our paper for more details on  other types of coherent fabrics.</p>
<h2>1. Compute Express Link (CXL)</h2>
<p><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://computeexpresslink.org/">CXL today is widely available</a>, with latest server CPUs from Intel and AMD supporting memory expanders from various vendors. But given such options, one needs to decide if they need CXL, and how to build a CXL system.</p>
<h3>1.1 Who Should Use CXL?</h3>
<p>The core value propositions of CXL-based memory expansion are:</p>
<ol>
<li><b>Massive Capacity Expansion:</b> CXL allows for the addition of hundreds of gigabytes, or even terabytes, of extra memory through the server&#8217;s PCIe slots.</li>
<li><b>Targeted Bandwidth Expansion:</b> In scenarios where the main memory channels are saturated but PCIe lanes sit idle, CXL can increase the total system bandwidth. And CXL is superior to directly-attached DIMMs in terms of bandwidth per CPU pin. This is particularly beneficial for workloads that are not acutely sensitive to latency.</li>
</ol>
<p>If any of these are currently limiting your workload, then CXL might be worth looking into. <b>And the rest of this article may help you with more detailed suggestions. </b></p>
<h3>1.2 How to architect a CXL-based Machine?</h3>
<p>Before you start to spec your machine with CXL, there are a few things to consider:</p>
<h4>1.2.1 Latency Tax</h4>
<p>The first principle of CXL is that it is slower than the Local DRAM with accesses being 50% to 300% slower. Measurements show that while a typical local DIMM access might take around 100 ns, an access to a modern ASIC-based CXL device falls in the 200-300 ns range. Early-generation FPGA-based CXL prototypes are even slower, with latencies around 400 ns.</p>
<p>CXL has high latency tax, however alternatives to CXL are even worse. To put CXL’s latency in context, here is a comparison of different memory/storage technologies:</p>
<table style="height: 238px;" width="696">
<tbody>
<tr>
<td><b>Metric</b></td>
<td><b>Local DRAM (DIMM)</b></td>
<td><b>Local CXL Memory (ASIC)</b></td>
<td><b>NVMe SSD</b></td>
</tr>
<tr>
<td><b>Typical Latency</b></td>
<td>Lowest
<br>
(~80-120 ns)</td>
<td>Medium
<br>
(~200-300 ns)</td>
<td>Highest
<br>
(10,000+ ns)</td>
</tr>
<tr>
<td><b>Peak Bandwidth</b></td>
<td>Highest
<br>
(200+ GB/s)</td>
<td>Medium
<br>
(~30 GB/s per device with 16 lanes)</td>
<td>Lower
<br>
(~7-14 GB/s)</td>
</tr>
<tr>
<td><b>Max Capacity</b></td>
<td>Limited by DIMM slots</td>
<td>High (Terabytes)</td>
<td>Highest (Many TBs)</td>
</tr>
<tr>
<td><b>Recommended Use Case</b></td>
<td>Hot Data, Performance</td>
<td>Warm Data, Capacity Tier</td>
<td>Cold Data, Storage</td>
</tr>
</tbody>
</table>
<p><b>Rule of the thumb: It is important to remember that CXL is a new tier of memory, not a DRAM replacement!</b> <b>Don&#8217;t think of CXL as a way to get more of fast memory. Think of it as a way to get faster access to massive capacity.</b></p>
<h4>1.2.2 Bandwidth: Scalable but with Limits</h4>
<p>A single CXL memory expander using 8 CXL 2.0 lanes (each CXL lane takes one PCIe lane) provides up to 32 GiB/s of additional memory bandwidth. Modern AMD servers such as AMD Turin provide up to 64 CXL lanes, offering up to 250 GiB/s of additional memory bandwidth.</p>
<p>In practice, we measured about 25-30 GiB/s of memory bandwidth of a single 8-lane CXL memory expander on real systems. Thus a 64 CXL-lanes CPU can support 200~240 GiB/s of CXL bandwidth.</p>
<h2>2. The 5  Essential Rules of CXL Programming</h2>
<h3>Rule #1: Pin Your Workloads, Especially on Earlier Intel CPUs</h3>
<p><img loading="lazy" decoding="async" class="wp-image-95322 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2025/11/image2-1.png" alt="" width="440" height="219" /></p>
<p>To avoid limiting the workloads to a fraction of the available Last-Level Cache (LLC), Intel CPU users should strongly consider pinning workloads that access CXL memory to the local socket using tools like <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://github.com/numactl/numactl">numactl</a>.</p>
<p>The reason is that on the tested Intel <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.intel.com/content/www/us/en/products/docs/processors/xeon-accelerated/4th-gen-xeon-scalable-processors.html">Sapphire Rapids</a> (SPR) and <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.intel.com/content/www/us/en/products/docs/processors/xeon/5th-gen-xeon-scalable-processors.html">Emerald Rapids</a> (EMR) CPUs, an application accessing CXL memory remotely is restricted to only a fraction of its local CPU&#8217;s LLC which is as little as 1/8th of the total cache on SPR and 1/4th on EMR. In contrast, remote access to standard DIMM memory can utilize the full LLC. However, the tested <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.amd.com/en/products/processors/server/epyc/4th-generation-architecture.html">AMD Zen4</a> (we only have a single socket <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.amd.com/en/products/processors/server/epyc/9005-series.html">Zen5</a> so cannot verify whether Zen5 is affected) and the new Intel <a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.intel.com/content/www/us/en/content-details/845771/intel-xeon-6-processor-family-product-brief.html">Granite Rapids</a> (GNR) systems did not exhibit this behavior, showing symmetric cache utilization for both local and remote CXL access. For system architects, it means services must be designed with an explicit awareness of this physical topology to avoid catastrophic performance degradation.</p>
<table>
<tbody>
<tr>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95307" src="https://www.sigarch.org/wp-content/uploads/2025/11/image5.png" alt="" width="315" height="339" /></td>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95310" src="https://www.sigarch.org/wp-content/uploads/2025/11/image8.png" alt="" width="316" height="339" /></td>
</tr>
<tr>
<td>
<p style="text-align: center;">Zen4-1-ASIC-CXL-1: DRAM bandwidth.</p>
</td>
<td>
<p style="text-align: center;">Zen4-1-ASIC-CXL-1: CXL bandwidth.</p>
</td>
</tr>
<tr>
<td colspan="2">
<p style="text-align: center;">Running two threads of bandwidth test on different cores, using Zen4-1 with ASIC-CXL-1 (refer to Table 1 for hardware details).</p>
</td>
</tr>
</tbody>
</table>
<p>Workload pinning should also take chiplet performance into consideration. As illustrated in the above figure, chiplet architecture impacts the bandwidth of accessing not just DRAM but also CXL; accesses originating from cores within the same chiplet group have a lower bandwidth. While the above example is from Zen4, other chiplet-based CPUs generally have similar performance characteristics.</p>
<h3>Rule #2: Asymmetric Read/Write Performance</h3>
<p>We observed a critical performance asymmetry on the tested AMD platforms. While load (read) bandwidth scaled with thread count, the store (write) bandwidth for ASIC CXL devices remained flat and low, regardless of the number of threads. This indicates a potential &#8220;performance asymmetry&#8221; for certain workloads on these platforms, where a very small number of cores can achieve the peak bandwidth with stores vs. loads.</p>
<table>
<tbody>
<tr>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95312" src="https://www.sigarch.org/wp-content/uploads/2025/11/image10-e1764109518176.png" alt="" width="340" height="199" /></td>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95306" src="https://www.sigarch.org/wp-content/uploads/2025/11/image4-e1764109565249.png" alt="" width="340" height="200" /></td>
</tr>
<tr>
<td>
<p style="text-align: center;">Store performance DIMMs vs CXL memory expander.</p>
</td>
<td>
<p style="text-align: center;">Load performance DIMMs vs CXL memory expander.</p>
</td>
</tr>
<tr>
<td colspan="2">
<table>
<tbody>
<tr>
<td colspan="2">
<p style="text-align: center;">Load and store bandwidth scaling for Intel EMR using an ASIC-based CXL memory expander. Store BW is saturated with a lot fewer cores than the load BW.</p>
</td>
</tr>
</tbody>
</table>
</td>
</tr>
</tbody>
</table>
<h3>Rule #3: Introducing CXL Memory Reduces the Overall Memory Access Latency</h3>
<table>
<tbody>
<tr>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95305" src="https://www.sigarch.org/wp-content/uploads/2025/11/image3.png" alt="" width="362" height="219" /></p>
<p style="text-align: center;">Average load access latency scaling with threads.</p>
</td>
<td><img loading="lazy" decoding="async" class="alignnone wp-image-95311" src="https://www.sigarch.org/wp-content/uploads/2025/11/image9.png" alt="" width="338" height="205" /></p>
<p style="text-align: center;">Total load bandwidth scaling with threads.</p>
</td>
</tr>
<tr>
<td colspan="2">
<p style="text-align: center;">Latency and bandwidth scaling with threads shows that DIMMs+CXL not only increased the total available memory bandwidth, but also decreased the average memory access latency.</p>
</td>
</tr>
</tbody>
</table>
<p>Adding CXL memory alongside DDR DIMMs lowers overall system memory latency even though CXL memory itself is slower. This happens because the extra bandwidth from CXL prevents DRAM channels from saturating, reducing queuing delays and shortening average access time.</p>
<p>In experiments on an Intel SPR system with two CXL memory expanders on the same socket, we found:</p>
<ul>
<li><b>Latency improvement:</b> Once CXL devices were active, total memory latency dropped compared to using only DIMMs. E.g., the “Local-DIMM + 1xCXL” has lower read latencies than using solely local DIMMs.</li>
<li><b>Bandwidth extension:</b> A single CXL device added roughly 17 GiB/s of bandwidth, while two devices offered about 25 GiB/s. This is below their combined theoretical 50 GiB/s peak, indicating partial utilization.</li>
</ul>
<p>On the remote socket, CXL memory did not increase bandwidth. This is likely due to UPI saturation, but it still reduces latency when used with DIMMs.</p>
<h3>Rule #4: Which CPU Architecture to Pick?</h3>
<p>The bandwidth and latency performance of CXL Memory Expanders are significantly influenced by the CPU microarchitecture. We find AMD CPUs in general can saturate the CXL device bandwidth, while Intel’s earlier generations (SPR and EMR) are sub-optimal; but  the recent GNR generation reaches parity with AMD.</p>
<p>Intel and AMD CPUs demonstrate a similar bandwidth accessing local memory. However, Intel&#8217;s SPR and EMR processors exhibit relatively lower bandwidth in remote memory access compared to local access, while the latest GNR fixed the issue. In terms of latency characteristics, Intel CPUs generally offer a lower latency than AMD processors.</p>
<p>Both CPU architectures exhibit a common phenomenon whereby latency increases dramatically when approaching a maximum bandwidth utilization. This represents a typical trade-off characteristic that occurs during memory system saturation and constitutes an important performance consideration when implementing CXL Memory Expander solutions.</p>
<h3>Rule #5: Capacity Expansion Enables Complex AI-based Scientific Discovery Like AlphaFold3</h3>
<p>CXL memory makes it tempting for many memory-hungry workloads, and among which we find AI-based scientific workloads an interesting use case. Such workloads consume a large amount of memory capacity and require good enough bandwidth, while the end-to-end execution time is dominated by CPU-side operations rather than GPU. CXL is the drop-in solution for them.</p>
<p><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cseweb.ucsd.edu/~jzhao/files/kim-iiswc2025.pdf">In our experiments with AlphaFold3 accessing CPU DIMM memory</a>: even when most inputs run within system memory capacity, heavy inputs including RNA will require hundreds of gigabytes. On DIMM-only systems these inputs fail with out-of-memory errors, but with CXL memory expanders they complete successfully. Although CXL introduces additional latency, its impact was secondary; capacity expansion was the decisive factor for enabling these workloads.</p>
<p><img loading="lazy" decoding="async" class="wp-image-95308 aligncenter" src="https://www.sigarch.org/wp-content/uploads/2025/11/image6.png" alt="" width="488" height="331" srcset="https://www.sigarch.org/wp-content/uploads/2025/11/image6.png 488w, https://www.sigarch.org/wp-content/uploads/2025/11/image6-480x325.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 488px, 100vw" /></p>
<p>This is a clear example of a broader pattern: scientific workloads are often deployed on HPC systems already tuned for expected requirements, yet unusually large inputs can break those assumptions. In such situations, CXL memory provides a flexible alternative by expanding capacity on demand without rebuilding servers or modifying applications.</p>
<h2>Conclusions</h2>
<p>We are entering the era of heterogeneous computing, where the memory system is getting more heterogeneous with the addition of coherent fabrics like CXL. Our research revealed that such systems have unconventional performance characteristics, and programmers have to carefully consider such characteristics to achieve optimal performance.</p>
<h2>Acknowledgement</h2>
<p>This work was supported by the PRISM and ACE centers, two of the seven centers in JUMP 2.0, a Semiconductor Research Corporation (SRC) program. We would also like to acknowledge Microsoft, Giga Computing, Samsung Memory Research Center, and National Research Platform for providing access and support for various servers.</p>
<p><b>About the authors:</b></p>
<p><i>Zixuan Wang got his PhD from the Computer</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research spans across architecture and system, with a focus on heterogeneous memory system performance and security.</i></p>
<p><i>Suyash Mahar got his PhD from the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research focuses on cloud computing with focus on memory and storage systems.</i></p>
<p><i>Luyi Li is a PhD student in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research focuses on exploring vulnerabilities, designing protections and optimizing performance for CPU architecture and memory systems.</i></p>
<p><i>Jangseon Park is a PhD student in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research interests are in heterogeneous memory system architecture for AI systems with emerging memories</i><b><i>.</i></b></p>
<p><i>Jinpyo Kim is a PhD student in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research focuses on memory-centric optimization and energy-efficient computing for heterogeneous AI/ML architectures.</i></p>
<p><i>Theodore Michailidis is a PhD candidate in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research interests lie at the intersection of memory, operating and datacenter systems.</i></p>
<p><i>Yue Pan is a PhD student in the </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research interest lies in computer architecture, memory, and system designs for high-performance applications. </i></p>
<p><i>Mingyao Shen got his PhD from the Computer</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research is in storage and memory system performance. This work was done while Mingyao was with UCSD.</i></p>
<p><i>Tajana Rosing is a Fratamico Endowed Chair of </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i>Computer Science and Engineering </i></a><i>and </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ece.ucsd.edu/"><i>Electrical Engineering </i></a><i>at </i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i>University of California, San Diego. </i></a><i>Her research spans energy-efficient computing, computer architecture, neuromorphic computing, and distributed embedded systems. She is an ACM and IEEE Fellow.</i></p>
<p><i>Dean Tullsen is a Distinguished Professor in the</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. His research focuses on computer architecture, with contributions spanning on-chip parallelism (multithreading, multicore), architectures for secure execution, software and hardware techniques for parallel speedup, low-power and energy-efficient processors, servers, datacenters, etc.</i></p>
<p><i>Steven Swanson is a Professor in the</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i> and holds the Halicioglu Chair in Memory Systems.  His research focuses on understanding the implications of emerging technology trends on computing systems.</i></p>
<p><i>Jishen Zhao is a Professor in the</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://cse.ucsd.edu/"><i> Computer Science and Engineering Department</i></a><i> at</i><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~www.ucsd.edu/"><i> University of California, San Diego</i></a><i>. Her research spans and stretches the boundary across computer architecture, system software, and machine learning, with an emphasis on memory systems, machine learning and systems codesign, and system support for smart applications.</i></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/930748850/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/930748850/0/sigarch-cat~The-Hitchhikers-Guide-to-Coherent-Fabrics-Programming-Rules-for-CXL-NVLink-and-InfinityFabric/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">95299</post-id></item>
<item>
<feedburner:origLink>https://www.sigarch.org/ieee-computer-architecture-letters-cal-an-update-and-faqs/</feedburner:origLink>
		<title>IEEE Computer Architecture Letters (CAL) &#8211; An Update and FAQs</title>
		<link>https://feeds.feedblitz.com/~/928607168/0/sigarch-cat~IEEE-Computer-Architecture-Letters-CAL-An-Update-and-FAQs/</link>
		<comments>https://feeds.feedblitz.com/~/928607168/0/sigarch-cat~IEEE-Computer-Architecture-Letters-CAL-An-Update-and-FAQs/#respond</comments>
		<pubDate>Fri, 21 Nov 2025 15:00:30 +0000</pubDate>
		<dc:creator><![CDATA[Sudhanva Gurumurthi and Mattan Erez]]></dc:creator>
		<category><![CDATA[ACM SIGARCH]]></category>
		<category><![CDATA[Journal]]></category>
		<category><![CDATA[Peer-review]]></category>
		<guid isPermaLink="false">https://www.sigarch.org/?p=94674</guid>
		<description><![CDATA[<div><img width="300" xheight="200" src="https://www.sigarch.org/wp-content/uploads/2025/11/AdobeStock_616008509-300x200.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="max-width:100% !important;height:auto !important;margin-bottom:15px;margin-left:15px;float:right;"  loading="lazy" /></div>CAL has held a unique place in the computer architecture community for well over two decades as a periodical for publishing early and exciting results. CAL papers are only four pages long and undergo rigorous peer review to select those with novel ideas and/or insights that are of interest to the computer architecture community and [&#8230;]]]>
</description>
				<content:encoded><![CDATA[<div><img width="300" height="200" src="https://www.sigarch.org/wp-content/uploads/2025/11/AdobeStock_616008509-300x200.jpeg" class="attachment-medium size-medium wp-post-image" alt="" style="margin-bottom:15px;margin-left:15px;float:right;" decoding="async" loading="lazy" /></div><p><span style="font-weight: 400;">CAL has held a unique place in the computer architecture community for well over two decades as a periodical for publishing early and exciting results. CAL papers are only four pages long and undergo rigorous peer review to select those with novel ideas and/or insights that are of interest to the computer architecture community and may have high impact. Another unique attribute of CAL is that it has historically provided a shorter turnaround for reviews compared to conferences and journals. As you will see below, the turnaround time to the first decision is now typically about 30 days</span></p>
<p><span style="font-weight: 400;">We began our EIC/AEIC terms on January 1, 2025. It has been a busy and exciting year! We grew the </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.computer.org/csdl/journal/ca/about/107318?title=Editorial%20Board&amp;periodical=IEEE%20Computer%20Architecture%20Letters"><span style="font-weight: 400;">editorial board</span></a><span style="font-weight: 400;"> to include several academic and industry experts from across the world. We worked with IEEE to update the scope of the periodical to bring it in line with major architecture conferences and made numerous process changes to improve the submission-to-publication turnaround time, building on </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.sigarch.org/save-ieee-computer-architecture-letters/"><span style="font-weight: 400;">work done by prior EICs</span></a><span style="font-weight: 400;">. We also updated CAL’s publicity strategy to encourage the use of arXiv and social media per IEEE’s guidelines and best practices.</span></p>
<p><span style="font-weight: 400;">Over the course of the year, we’ve communicated with many members of the architecture community and identified a few questions that we frequently get asked. We’re capturing those here. </span></p>
<p><b>FAQ1: Can you share any stats about how CAL is doing?  </b></p>
<p><span style="font-weight: 400;">We have been tracking two metrics at a monthly cadence:</span></p>
<ol>
<li style="font-weight: 400;"><span style="font-weight: 400;">Acceptance rate of submitted manuscripts</span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">Turnaround time from submission to first decision</span></li>
</ol>
<p><span style="font-weight: 400;">The graph below plots these metrics as an average over a 12-month window. This data was gathered from the IEEE ScholarOne Manuscripts™ EIC dashboard. The bars correspond to the acceptance rate and the line graph to the turnaround time from submission to first decision.</span></p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-94677 size-full" src="https://www.sigarch.org/wp-content/uploads/2025/11/IEEE-Computer-Architecture-Letters-Monthly-Statistics.png" alt="" width="600" height="371" srcset="https://www.sigarch.org/wp-content/uploads/2025/11/IEEE-Computer-Architecture-Letters-Monthly-Statistics.png 600w, https://www.sigarch.org/wp-content/uploads/2025/11/IEEE-Computer-Architecture-Letters-Monthly-Statistics-480x297.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 600px, 100vw" /></p>
<p><span style="font-weight: 400;">We can see that the acceptance rate initially dropped and has been fluctuating around 35% for the past few months. When a submission arrives to the EIC queue, we have workflows in place to check a manuscript for scope and minimum technical substance. Papers that pass these steps are then sent out for review by the Associate Editor (AE). We’ve found that most manuscripts that eventually get accepted at CAL go through one or more revision rounds (more details on this in the next FAQ). While this acceptance rate is higher than many architecture conferences, we think this is reflective of CAL receiving many good papers and reviewers being more welcoming of early-stage results and new directions of work so long as they are novel, technically sound, and the authors take into account reviewer feedback through the revision rounds. </span></p>
<p><span style="font-weight: 400;">We can also see that, after an initial couple of months, the average turnaround time of manuscripts has been steadily decreasing over the course of the year, with the most recent datapoint being about 30 days. For all manuscripts submitted to CAL this year, we’ve found that approximately 85% received their first decision within 30 days and the remainder within 60 days. These stats are in line with CAL’s turnaround trends in </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://ieeexplore.ieee.org/document/5648421"><span style="font-weight: 400;">the past</span></a><span style="font-weight: 400;">. This fast turnaround is the result of the collective hard work put in by the CAL AEs, reviewers, and the IEEE support staff. Reviewing is hard work and the fast-paced nature of CAL adds a layer of expediency to the process. </span></p>
<p><span style="font-weight: 400;">IEEE sends out a quarterly report to all EICs and the VP of Publications on the turnaround times of all journals. In the most recent report that was sent out in the first week of November, we were happy to see CAL designated as a </span><b>high performer</b><span style="font-weight: 400;">! </span></p>
<p><b>FAQ 2: Why don’t you track the turnaround time from submission to publication or set any goals for it? </b></p>
<p><span style="font-weight: 400;">Once the first decision is sent out, there are many variables that impact when a paper is fully decided as an Accept or Reject. The first decision could be an Accept, a Revise&amp;Resubmit, a Minor Revision, or a Reject. It is very rare for a manuscript to receive an Accept as a first decision. Most papers that eventually get published at CAL go through a Revise&amp;Resubmit, where authors get 6 weeks to submit their revision, and/or a Minor Revision, whether the authors get one week. A Revise&amp;Resubmit is analogous to a “Major Revision” at other journals and requires a complete second review round, with sufficient time given to the reviewers to evaluate the new version and enter their reviews. For a Minor Revision, the AE can, if they so choose, provide a recommendation to the EIC without seeking additional inputs from the reviewers. Given these possibilities and all the parties involved, it could be two months or more for the final decision. </span></p>
<p><b>FAQ 3: Does CAL guarantee a specific turnaround for the first decision?</b></p>
<p><span style="font-weight: 400;">Sorry but there are no guarantees. We routinely keep an eye on all in-flight CAL manuscripts and the review system tracks turnaround and sends out automated reminders to the AEs and reviewers. The AEs and EIC also send out personal reminders, as needed. While we’ve found all of these help move things along (as evidenced by the data shown in FAQ 1), we can’t promise or enforce any specific turnaround. </span></p>
<p><b>FAQ 4: I have a paper submitted to CAL that I’d like to expand for an upcoming conference. Is it okay to submit that paper while the CAL paper is still under review?</b></p>
<p><span style="font-weight: 400;">While CAL welcomes papers on early results that could eventually be expanded to a conference paper, </span><a href="https://feeds.feedblitz.com/~/t/0/0/sigarch-cat/~https://www.computer.org/publications/author-resources#concurrent-duplicate-submissions"><span style="font-weight: 400;">IEEE Computer Society rules</span></a><span style="font-weight: 400;"> disallow a CAL paper that is not fully decided to be submitted to a conference. CAL is an independent periodical with review timelines that are independent of other venues. Also, if your expanded work overlaps with the overall turnaround time of CAL, please consider if the initial submission truly represents early research. We feel a good heuristic for “early” is about 12 months or more from when the work is ready for submission to a conference or Transactions-style journal. Please plan accordingly.</span></p>
<p><b>FAQ 5: How can I help CAL?</b></p>
<p><span style="font-weight: 400;">There are many ways to do this:</span></p>
<ol>
<li style="font-weight: 400;"><span style="font-weight: 400;">Please consider submitting your early research results to CAL. CAL also welcomes papers that provide novel and insightful learnings from an industry context. Please keep FAQs 3 and 4 in mind when planning a submission. </span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">If you are asked to review a paper for CAL, please agree or suggest alternative reviewers.  </span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">If you are an author of a paper accepted at CAL, please publicize your work.</span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">Attend the Best of CAL session at HPCA.</span></li>
</ol>
<p><span style="font-weight: 400;">We thank the computer architecture community for your support of CAL in your roles as authors, reviewers, and editorial board members! Your help has been invaluable in strengthening CAL and retaining its special place as a forum to publish early, novel, and exciting results.</span></p>
<h3><span style="font-weight: 400;">About the Authors:</span></h3>
<p><b>Sudhanva Gurumurthi</b><span style="font-weight: 400;"> is a Fellow at AMD, where he is responsible for research and advanced development in RAS. His work has impacted numerous AMD products, multiple industry standards, and external research in the field. Before joining industry, Sudhanva was an Associate Professor in the Computer Science Department at the University of Virginia. Sudhanva is the recipient of an NSF CAREER Award, a Google Focused Research Award, and is named to the ISCA Hall of Fame. Sudhanva received his BE in Computer Science and Engineering from the College of Engineering Guindy, Anna University, and his PhD in Computer Science and Engineering from Penn State.</span></p>
<p><b>Mattan Erez</b><span style="font-weight: 400;"> is a Professor in the Department of Electrical &amp; Computer Engineering at The University of Texas at Austin, where he holds the Cullen Trust for Higher Education Endowed Professorship in Engineering #7. He has received several best paper awards at international conferences and is named to the Hall of Fame of ISCA and HPCA. Mattan is the recipient of many research awards, including the NSF CAREER Award, the DOE Early Career Research Award, and the Presidential Early Career Research Award for Scientists and Engineers awarded by President Obama. Mattan received a BSc in Electrical Engineering and a BA in Physics from the Technion, Israel Institute of Technology, and his MS and PhD in Electrical Engineering from Stanford University. </span></p>
<p class="disclaim"><strong>Disclaimer:</strong> <em>These posts are written by individual contributors to share their thoughts on the Computer Architecture Today blog for the benefit of the community. Any views or opinions represented in this blog are personal, belong solely to the blog author and do not represent those of ACM SIGARCH or its parent organization, ACM.</em></p><Img align="left" border="0" height="1" width="1" alt="" style="border:0;float:left;margin:0;padding:0;width:1px!important;height:1px!important;" hspace="0" src="https://feeds.feedblitz.com/~/i/928607168/0/sigarch-cat">
]]>
</content:encoded>
			<wfw:commentRss>https://feeds.feedblitz.com/~/928607168/0/sigarch-cat~IEEE-Computer-Architecture-Letters-CAL-An-Update-and-FAQs/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
	<post-id xmlns="com-wordpress:feed-additions:1">94674</post-id></item>
</channel></rss>

