<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></title><description><![CDATA[Abhinav Madahar (अभिनव ਮਦਾਹਰ) is an independent computer scientist. His research focuses on advancing the reasoning capabilities of frontier artificial intelligence models.]]></description><link>https://abhinavmadahar.com</link><image><url>https://substackcdn.com/image/fetch/$s_!thXZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2bd3875-1ece-444f-ae16-8ae902b26b59_1022x1022.png</url><title>Abhinav Madahar · अभिनव ਮਦਾਹਰ</title><link>https://abhinavmadahar.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 06 Apr 2026 06:13:34 GMT</lastBuildDate><atom:link href="https://abhinavmadahar.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[abhinavmadahar@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[abhinavmadahar@substack.com]]></itunes:email><itunes:name><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></itunes:name></itunes:owner><itunes:author><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></itunes:author><googleplay:owner><![CDATA[abhinavmadahar@substack.com]]></googleplay:owner><googleplay:email><![CDATA[abhinavmadahar@substack.com]]></googleplay:email><googleplay:author><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Why Do We Pursue Science?]]></title><description><![CDATA[A Reflection on the Scientific Endeavour]]></description><link>https://abhinavmadahar.com/p/why-do-we-pursue-science</link><guid isPermaLink="false">https://abhinavmadahar.com/p/why-do-we-pursue-science</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Tue, 03 Feb 2026 11:23:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!thXZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2bd3875-1ece-444f-ae16-8ae902b26b59_1022x1022.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>My life is very, very unusual.</p><p>I spend my time attempting to construct a mathematical structure which can reason generally. To do so, I need to examine what intelligence itself is, abstracting away from any particular local or general realization.</p><p>Conceptually, this is very far removed from finding berries and hunting gazelle, the behavioural envelope against which humans are optimized. Science, generally, is far removed from direct importance.</p><p>Why do we pursue science? Why do we examine the zeros of functions, the behaviour of enzymes, and the motion of particles? I do not ask this question while hoping for a vaguely self-congratulatory answer, e.g. that we pursue science to satisfy a sense of curiosity inherent to humans. <em>Really, why do we pursue science?</em> At the fundamental level, science is pursued because its results are useful. The criteria of usefulness may change, but this structure is steady.</p><p>Science was first used because its results were theologically useful. Religious dynamics turned on the visible behaviour of celestial objects, so improvements in the ability to predict these behaviours <em>ex ante</em> improved the ability to predict religious activities. Because the complexity of celestial behaviour was sufficient that studying it through its utility to religion overly dragged progress, astronomers emerged who studied celestial behaviour directly without individually considering downstream theological utility of the knowledge they produce. Consequently, astronomy was the first science to emerge in the sense of being the first field which saw intergenerationally-sustained growth in an understanding of a matter which was not itself applied. We can contrast this with, for example, improvements to agricultural techniques in the millennia preceding the third-millennium BCE emergence of astronomy; although intergenerational growth was certainly present, it was on an applied technique, not on a body of knowledge whose value derived from having that knowledge itself.</p><p>Astronomy in this context was funded in the hopes that its findings would eventually be theologically useful, but individual astronomers and individual findings were not evaluated by directly forecasting their eventual theological utility. Instead, they were evaluated on the extent to which they advanced astronomical knowledge itself, and the eventual theological utility of this knowledge and its producers were considered separately by examining the extent to which astronomical advancement imparted theological utility.</p><p>This structure has persisted since its first emergence of science in the third millennium BCE, and it emerged in all other contexts where science emerged and lasted: individual scientists and individual scientific findings were evaluated on the extent to which they advanced science itself while scientific fields as a whole were evaluated on the extent to which they were useful for matters important. This was necessary: as bodies of scientific knowledge grow, it becomes increasingly difficult to evaluate how useful individual findings are for downstream uses, forcing an evaluative decoupling.</p><p>This decoupling is where I want to focus my attention, and it lets us sharpen our earlier question. Replacing our earlier query of why science is pursued, we can ask whether advances in science should be evaluated exclusively on the extent to which they advance their respective fields, evaluated exclusively on the extent to which they impact downstream utility, or something else.</p><p>This is a difficult question to answer. When scientific funding was limited, the difficulty of the question was immaterial as the answer to the question itself had minimal society-wide impact. However, with the rapid increase in scientific funding with the transition to the modern era, this question became increasingly salient as states and societies became compelled to justify their increasingly large expenditure on basic scientific research.</p><p>The scientific community in the nineteenth century attempted to evade this question by constructing an understanding of how basic scientific work is eventually translated into practical utility. Explicit attention was systematically paid to the chain through which findings in a given field would eventually lead to utility inherently useful, arranged into what were networks in substances even if not explicitly articulated as such. For example, it became recognized that discoveries in how specific chemical reactions proceed could plausibly lead to improvements in the collective understanding of certain biological processes, whose improved understanding could itself plausibly lead to improved medicine. As such, it became structurally routine and central to aim to articulate the pathway by which basic research could be eventually translated into practical utility.</p><p>These translational pathways have grown in scope and nuance, and for many areas of science, they obviate the need to ask the question. If a field is sufficiently mature that the form of its findings can be reliably predicted <em>ex ante</em>, then it is possible to identify translational pathways for its findings, obviating the need for the field to answer this question for findings which have an identified translational pathway. This is not an obscure observation&#8212;far from it. In many fields, identifying such translation pathways to avoid needing to answer this question is a major research activity, and engaging in it is not seen as defensive justification but as responsible consideration of downstream utility. Given the difficulty of this question and the possible benefits of certain scientific advances, this is less a cowardly evasion and more a case of individuals needing to conduct important work before philosophers finish answering a question.</p><p>This question does not need to be answered given the construction of an appropriate translational pathway.</p><p>My own research has such a translation pathway. Advances in our understanding of the nature of intelligence improve our ability to construct an artificial intelligence which can answer questions we ask and conduct intellectual labour we consider useful. The translational pathway is so simple as to be nearly comical: as long as there are questions and problems worth answering and solving, respectively, advances in intelligence are useful.</p><p>My expectation is that, over a very long horizon, continued economic growth will lead to this question becoming increasingly less important. If increases in scientific funding at the onset of the modern era forced this question, increases in the economic base freeing more funding for science would lessen its significance. Perhaps humanity in the distant future will look back at our current attempts to justify science as quaint endeavours when scientific funding was limited and in need of rationing.</p><p>My education was in mathematics with my education in artificial intelligence fitted atop. As the reader can expect, I deeply internalized the prevailing norm in mathematics that advances in the field should be evaluated on their respective contributions to mathematics, not on their practical utility. More than most fields, mathematics is structurally forced to address this question as its translational pathways are arguably the most complex of any field in the sciences. Its continued existence is not the result of articulations of translational pathways but through an exceptional risk&#8211;reward structure: even if almost all mathematical findings are not useful, the few which are justify the field.</p><p>I will most likely continue thinking about this question for the remainder of my life, assuming the philosophers of science do not provide an unexpected answer, even if its answer is immaterial for my research. If there does exist a satisfying answer to this question, my research in intelligence may even lead to the production of an artificial intelligence which answers it for us. My hope is that a solution is in our stars.</p>]]></content:encoded></item><item><title><![CDATA[Artificial Intelligence Is Arriving in the World We Already Live In]]></title><description><![CDATA[Why Artificial Intelligence Must Be Addressed at the International Level]]></description><link>https://abhinavmadahar.com/p/artificial-intelligence-is-arriving</link><guid isPermaLink="false">https://abhinavmadahar.com/p/artificial-intelligence-is-arriving</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Fri, 23 Jan 2026 02:34:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!thXZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2bd3875-1ece-444f-ae16-8ae902b26b59_1022x1022.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Discussions of advanced artificial intelligence often assume a comforting distance: that systems with genuinely transformative capabilities belong to a future society&#8212;one that has already resolved today&#8217;s political, legal, and institutional challenges. That assumption is increasingly untenable.</p><p>The pace of progress in artificial intelligence has already exceeded what many experts expected only a few years ago. While it is possible that this period of rapid improvement will slow, the more responsible assumption is that capability growth will continue, and may even accelerate. We should therefore seriously consider the possibility that highly capable AI systems will emerge not in a hypothetical future, but in the world as it exists now: fragmented, unequal, and governed by institutions that were not designed with such systems in mind.</p><p>This matters because the benefits and risks of artificial intelligence scale together. Systems capable of solving harder problems can materially improve medicine, make infrastructure safer, and expand access to education for people who are currently under-served. At the same time, those same systems can cause harm&#8212;either through intentional misuse or through unintended consequences. Preventing such harm is not a problem we can defer until after these systems exist. It is a problem that must be addressed in advance.</p><p>Crucially, many of the most serious risks posed by advanced AI cannot be managed by individual actors acting alone. They are international in structure, cross-border in impact, and coordination-dependent by nature.</p><h2>Preventing Misuse Requires International Coordination</h2><p>Highly advanced AI systems require enormous upfront investment to develop. As a result, only a small number of states and well-resourced non-state actors are capable of creating them. This concentration is often overlooked, but it is important: while the <em>use</em> of AI systems can be widely distributed, the <em>creation</em> of frontier systems is not.</p><p>That distinction creates a narrow surface on which safety measures can be applied. Regulating individual users of AI is a highly dispersed and often intractable task. But the fact that only a limited number of entities can build the most capable systems means that meaningful safeguards are, in principle, feasible&#8212;if approached at the right level.</p><p>The difficulty is that misuse can be deliberately obscured through decentralization. A malicious actor need not commit an overtly harmful act in any single jurisdiction. Instead, harmful activity can be partitioned across systems, locations, or intermediaries, such that each individual use appears benign in isolation. The malicious intent becomes visible only when the full pattern is examined.</p><p>For example, consider a hypothetical actor attempting to develop a novel biological weapon. Rather than using a single AI system to pursue this goal directly, they might distribute the process across multiple systems located in different jurisdictions&#8212;using each system for tasks that appear innocuous on their own. No single state, acting alone, would necessarily detect the misuse. The risk emerges only at the global level.</p><p>This is not an argument that international coordination is preferable. It is an argument that, for the most difficult and dangerous forms of misuse, international coordination is <em>necessary</em>. Individual states and organizations can prevent some misuse on their own. But the forms of misuse that are hardest to detect and most consequential to prevent are precisely those that evade unilateral oversight.</p><p>It is therefore appropriate to say that only the international community is positioned to prevent the most challenging cases of AI misuse&#8212;not because it holds unique authority, but because it is the only level at which the relevant patterns are visible.</p><p>Importantly, this does not require prescribing specific legal mechanisms or enforcement strategies. The point is not to dictate how prevention must occur, but to recognize where responsibility ultimately lies. Once that responsibility is acknowledged, different states can pursue different implementation paths consistent with their legal and political systems.</p><p>At the same time, the scientific community has made substantial progress in understanding how to constrain AI systems so that they behave in ways humans consider acceptable. The question of <em>how</em> to influence system behaviour is increasingly a technical one. What remains unresolved&#8212;and cannot be answered by engineers alone&#8212;is <em>which behaviours should be permitted and which should not</em>. That is a normative question, and it belongs to society and its institutions, not to laboratories.</p><h2>Accidents Are a Different Kind of Risk</h2><p>Not all harm from artificial intelligence arises from malicious intent. As AI systems become more capable, they can engage in increasingly complex behaviour. In some cases, this complexity leads to decisions or actions that were not intended by their human users, and that nonetheless cause harm.</p><p>Unlike intentional misuse, such accidents cannot always be predicted in advance. They arise from interactions between system capabilities, deployment contexts, and real-world environments that are difficult to model exhaustively. As AI systems are deployed at global scale, the consequences of such failures are unlikely to remain confined within national borders.</p><p>A single state may be able to regulate how AI systems are used within its own jurisdiction. But it is far more difficult for any state to ensure that systems deployed elsewhere do not cause harm within its borders. When systems operate across digital and physical infrastructure that spans countries, impact&#8212;not jurisdiction&#8212;becomes the relevant unit of analysis.</p><p>Containing the risk of unintended harm therefore requires governance frameworks that assume cross-border spillovers as the default, not as an exception. At that scale, bespoke bilateral treaties or narrowly scoped agreements are unlikely to be sufficient. What is required instead is coordinated international effort that treats accidental harm as a shared problem rather than a series of isolated failures.</p><p>This, again, is not a call for centralized control or uniform regulation. It is a recognition that when impact crosses borders by design, governance must do so as well.</p><h2>The Role of the International Community</h2><p>Artificial intelligence presents challenges that cannot be fully addressed by individual states or private actors acting independently. As systems grow in capability and reach, so too does their potential to help and to harm. While many concerns can and should be handled locally, those involving AI systems deployed and used across jurisdictions quickly become difficult&#8212;often intractable&#8212;for any single actor to manage alone.</p><p>At the same time, the scientific community continues to advance rapidly in its understanding of intelligence and in its ability to realize it in machine form. Institutional progress must occur in parallel. The international community has a responsibility to examine how these systems can be constrained, how risks can be anticipated, and how harms that were never intended can nonetheless be prevented.</p><p>The question is not whether advanced AI will arrive. It is whether our institutions will be prepared when it does.</p>]]></content:encoded></item><item><title><![CDATA[Toward 'Theory of Reasoning' Reasoning Architectures]]></title><description><![CDATA[A cognitive-scientific framework for unifying reasoning architectures under a theory of reasoning]]></description><link>https://abhinavmadahar.com/p/toward-theory-of-reasoning-reasoning</link><guid isPermaLink="false">https://abhinavmadahar.com/p/toward-theory-of-reasoning-reasoning</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Fri, 24 Oct 2025 03:32:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!thXZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2bd3875-1ece-444f-ae16-8ae902b26b59_1022x1022.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Introduction</h3><p>Recent years have seen the rise of <em>reasoning architectures</em>&#8212;frameworks such as Chain-of-Thought (CoT), Tree-of-Thoughts (ToT), and their successors&#8212;that enable large models to reason through structured intermediate steps. Yet despite impressive empirical results, these systems remain heuristic and cognitively ungrounded.</p><p>This white paper introduces <strong>Theory-of-Reasoning Reasoning Architectures (ToR-RAs)</strong>, a programme to formalise the <em>cognitive architecture</em> of reasoning itself: how coherence, metacognitive control, and resource allocation jointly give rise to structured thought.</p><div><hr></div><h3>Why This Work</h3><p>Reasoning architectures today resemble early computing before Turing&#8212;powerful, but lacking a theory of what they actually <em>are</em>. Just as the theory of computation defined &#8216;what it means to compute&#8217;, a theory of reasoning should define &#8216;what it means to reason&#8217;.</p><p>The ToR-RA framework treats reasoning as a cognitive architecture composed of interacting subsystems for inference, evaluation, and metareasoning operating under bounded resources. The aim is not merely to engineer better reasoning algorithms, but to <strong>found a science of reasoning architectures</strong> linking cognitive-scientific principles with computational design.</p><div><hr></div><h3>Conceptual Foundations</h3><p>A reasoning architecture, as defined in the paper, satisfies four invariants:</p><ul><li><p><strong>Coherence:</strong> Local inferences support global consistency.</p></li><li><p><strong>Causality of thought:</strong> Representational changes have identifiable precursors.</p></li><li><p><strong>Introspectability:</strong> The system can form meta-representations of its own reasoning.</p></li><li><p><strong>Economy:</strong> Inference optimises progress per unit of cognitive cost.</p></li></ul><p>These invariants align with long-standing frameworks in cognitive science&#8212;Marr&#8217;s three levels of analysis, rational metareasoning, and dual-process models&#8212;and together they provide a bridge between biological and artificial reasoning.</p><div><hr></div><h3>Relation to Prior Work</h3><p>ToR-RAs extend my earlier developments in reasoning architectures:</p><ul><li><p><strong>Lateral Tree-of-Thoughts (LToT)</strong> &#8212; reasoning breadth and cognitive economy.</p></li><li><p><strong>Natural Language Edge Labelling (NLEL)</strong> &#8212; semantic self-instruction and metacognitive control.</p></li></ul><p>Both suggested that reasoning architectures could serve as <em>models of reasoning</em> rather than merely as engineering tools. ToR-RAs make that synthesis explicit.</p><div><hr></div><h3>Significance</h3><p>ToR-RAs aim to unify cognitive science and reasoning-architecture research under a shared theoretical language, integrating interpretability, efficiency, and scientific clarity.</p><p>They represent a step toward understanding reasoning not as a black-box capability but as an <em>architectural phenomenon</em>&#8212;something that can be studied, designed, and improved with the same rigour once brought to computation itself.</p><div><hr></div><p><strong>Abhinav Madahar &#183; &#2309;&#2349;&#2367;&#2344;&#2357; &#2606;&#2598;&#2622;&#2617;&#2608;</strong><br>Independent Computer Scientist<br><a href="https://abhinavmadahar.com">abhinavmadahar.com</a>&#8195;|&#8195;abhinav@abhinavmadahar.com</p>]]></content:encoded></item><item><title><![CDATA[Toward a General Cure for Cancer]]></title><description><![CDATA[Why We Might Need an AGI Researcher]]></description><link>https://abhinavmadahar.com/p/toward-a-general-cure-for-cancer</link><guid isPermaLink="false">https://abhinavmadahar.com/p/toward-a-general-cure-for-cancer</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Wed, 08 Oct 2025 19:33:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!thXZ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2bd3875-1ece-444f-ae16-8ae902b26b59_1022x1022.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>One-sentence summary:</strong><br>Rather than relying solely on incremental, subtype-by-subtype progress, the world should fund and govern a globally accessible <em>AGI cancer researcher</em> &#8212; a language-based, tool-using system coupled to autonomous laboratories &#8212; explicitly tasked with discovering general interventions that restore growth control across tumors.</p><div><hr></div><h2>What unites cancers &#8212; and why a general strategy is thinkable</h2><p>Cancer&#8217;s common denominator is loss of growth control: cells sustain proliferative signaling, evade suppressors, resist death, and bypass intrinsic and extrinsic checkpoints. These capabilities recur across diverse tissues even as the initiating lesions vary; the <em>hallmarks of cancer</em> framework distills this convergence and has been repeatedly refined with new dimensions, including phenotypic plasticity, non-mutational epigenetics, and the tumor microenvironment [1&#8211;3]. Large-scale pan-cancer efforts show conserved disruptions to pathways across histologies, supporting the premise that cross-tumor interventions can be reasoned about without denying heterogeneity [4, 5].</p><p>Clinical practice already acknowledges shared vulnerabilities. Tumor-agnostic approvals &#8212; pembrolizumab for MSI-H/dMMR tumors and for TMB-high disease; larotrectinib for <em>NTRK</em> fusions &#8212; demonstrate that targeting a common failure mode (mismatch-repair deficiency, hypermutation, or an oncogenic fusion) can cut across tissue of origin. These are not &#8220;general cures,&#8221; but they prove the concept that lineage-agnostic levers exist [6&#8211;9].</p><div><hr></div><h2>Two routes to a general cure</h2><p><strong>The direct route</strong> is today&#8217;s default: mobilize human expertise to map mechanisms, integrate multi-omic and clinical data, run trials, and &#8212; through cumulative advances &#8212; approach broadly applicable interventions.</p><p><strong>The indirect route</strong> is technically plausible now: build an artificial general intelligence (AGI) capable of cross-domain scientific reasoning at the effective scale of the global research enterprise and assign it the problem <em>&#8220;discover a general cure for cancer.&#8221;</em> Such a system would read and synthesize literature; generate and prune hypotheses; design, schedule, and interpret experiments; integrate multi-modal data; and orchestrate long-horizon programs with automated labs and human clinicians.</p><p>Importantly, this augments rather than displaces scientists: by coupling models to <em>self-driving laboratories</em>, hypotheses must survive experiment, and error-prone reasoning is bounded by closed-loop validation and independent safety gates. The constituent technologies exist and are improving: &#8220;robot scientist&#8221; systems in biology and modern autonomous experimentation in chemistry, materials, and protein engineering have demonstrated end-to-end loops from hypothesis to result [10, 11].</p><p>The practice is not without precedent. When a concise, hand-checkable route was out of reach in mathematics, the Four-Colour Theorem was settled via a computer-assisted proof, later formalized in a proof assistant. A comparable modus operandi &#8212; computational procedures that scale beyond unaided human capacity, paired with verification &#8212; could apply to discovering pan-cancer interventions [12, 13].</p><div><hr></div><h2>Why a language-centric AGI is a plausible substrate for science</h2><p>Skeptics doubt language-model-based systems can reliably &#8220;do science.&#8221; Yet a substantial literature treats language as a scaffold for abstract thought and self-regulation, with &#8220;inner speech&#8221; coordinating cognition. If much high-level reasoning is conducted <em>through</em> language, scaling language-native models &#8212; augmented with tools, external memory, and verification &#8212; becomes a plausible route to general scientific competence (not the only route, but a powerful one) [14&#8211;16].</p><p>Just as relevant, contemporary reasoning models improve when afforded more <em>test-time compute</em>: structured deliberation, multiple solution paths, and self-consistency checks raise accuracy on hard problems. For frontier queries &#8212; e.g., <em>&#8220;design a multi-component regimen that durably restores growth control across contexts&#8221;</em> &#8212; reliability will hinge as much on inference-time thinking and verification as on pretraining scale [17&#8211;19].</p><div><hr></div><h2>A pragmatic program sketch</h2><p>A credible AGI-for-cancer effort should be practical and staged:</p><ol><li><p><strong>Knowledge integration.</strong> Unified corpora (literature, protocols, negative results) with governed access to clinical and genomic datasets; privacy-preserving analytics (e.g., federated learning) to respect data sovereignty and patient consent while enabling cross-site modeling. Frameworks from the Global Alliance for Genomics and Health (GA4GH) and large biobanks illustrate workable governance [20, 21].</p></li><li><p><strong>Hypothesis generation and curation.</strong> Multi-omic integration to surface lineage-agnostic vulnerabilities (DNA-damage response, cell-cycle checkpoints, apoptotic priming, telomere maintenance), systematically cross-validated against pan-cancer resources [3, 5].</p></li><li><p><strong>Automated experimentation.</strong> High-throughput perturbation (CRISPR/RNAi, small-molecule/protein modalities) run by self-driving labs to test and refine mechanistic claims, with pre-registered plans and independent replication nodes [10, 11].</p></li><li><p><strong>Design against evolution.</strong> From the outset, incorporate adaptive-therapy principles &#8212; explicitly optimizing combinations and schedules to delay or prevent resistance &#8212; so candidates are evaluated not just for response but for evolutionary stability [22, 23].</p></li><li><p><strong>Translational pathways.</strong> Use tumor-agnostic regulatory precedents as a template for trials centered on shared biomarkers or pathway states, then iterate toward progressively broader indications [6&#8211;9].</p></li></ol><div><hr></div><h2>The economics: train once, think many times</h2><p>Public estimates indicate that compute for recent frontier models already costs in the tens to low hundreds of millions of dollars, with several analyses projecting over $1 billion training runs within the next few years if current trends persist. The Stanford <strong>AI Index</strong> and Epoch AI both document rapid growth in training cost and compute; importantly, the total cost to <em>reason</em> on hard scientific questions will also include substantial inference-time computation (deliberation, retrieval, simulation, verification). Planning should assume wide error bars but a clear directional trend: training may be expensive, and sustained &#8220;thinking&#8221; on a problem this hard will be, too [24, 25].</p><div><hr></div><h2>Who pays &#8212; and how to keep it fair, safe, and sustainable</h2><p>At this scale, durable funding and legitimate governance exceed the remit of any single philanthropist, firm, or nation. Analogues exist: CERN&#8217;s council and shared capital facilities; ITER&#8217;s cost-sharing treaty; pooled health-financing and access mechanisms in the Global Fund and Gavi; and UNESCO&#8217;s <em>Open Science Recommendation</em> [26&#8211;30].</p><p>A treaty-backed, UN-anchored vehicle could (i) pool contributions, (ii) procure compute, power, and data at utility scale, (iii) embed auditability, safety testing, and biosecurity review, (iv) guarantee open-science outputs proportionate to public funding, and (v) allocate inference budgets to a short list of humanity-scale questions &#8212; with <em>a general cure for cancer</em> as a flagship.</p><p>Because public investors will expect safeguards, such a body should align with the WHO&#8217;s guidance on artificial intelligence in health and with emerging international safety norms (for example, the Bletchley Declaration). Concretely: staged capability evaluations, independent red-teaming, strict separation between <em>in silico</em> design and wet-lab actuation, tiered access to high-risk tools, and human-in-the-loop oversight for any experimental transition &#8212; especially in biology [31, 32].</p><p>Data stewardship should be designed in, not bolted on. Federated learning and federated analysis let models learn from distributed datasets without moving patient data; GA4GH&#8217;s frameworks provide principled governance; and biobank access regimes supply working templates for consent, oversight, and transparency [20, 21].</p><p>Biosecurity must be non-negotiable. Wet-lab integration should comply with existing nucleic-acid synthesis-screening frameworks (for example, the U.S. HHS guidance for providers of synthetic double-stranded DNA) and the International Gene Synthesis Consortium&#8217;s <em>Harmonized Screening Protocol</em>, with procurement requirements that restrict access to screened providers and benchtop devices. Independent review boards should have veto authority over any agent-design capabilities, and audit logs should be mandatory for all tool calls that could touch sequence design or synthesis [33, 34].</p><p>Environmental externalities must be managed up front. Training and large-scale inference consume energy and, depending on siting and cooling, water. Best-practice engineering &#8212; efficient hardware and algorithms, carbon-free energy procurement, low-water cooling, circularity targets &#8212; materially reduces impact. Credible social license will require binding carbon and water budgets, transparent accounting, and independent auditing [35, 36].</p><div><hr></div><h2>What success would mean</h2><p>If such an undertaking succeeded, it would culminate nearly two centuries of pathology and oncology &#8212; from Virchow&#8217;s cellular pathology, through the molecular and genomic eras, to a general restoration of growth control delivered with machine reasoning. The field&#8217;s <em>telos</em> &#8212; durably ending malignant proliferation &#8212; would be realized not by displacing human scientists, but by completing their project with a new kind of colleague.</p><div><hr></div><h2><strong>References</strong></h2><p>[1] Hanahan D. &amp; Weinberg R.A. <em>The Hallmarks of Cancer.</em> Cell 100, 57&#8211;70 (2000).<br>[2] Hanahan D. &amp; Weinberg R.A. <em>Hallmarks of Cancer: The Next Generation.</em> Cell 144, 646&#8211;674 (2011).<br>[3] Hanahan D. <em>Hallmarks of Cancer: New Dimensions.</em> Cancer Discovery 12, 31&#8211;46 (2022).<br>[4] Vogelstein B. <em>et al.</em> <em>Cancer Genome Landscapes.</em> Science 339, 1546&#8211;1558 (2013).<br>[5] ICGC/TCGA Pan-Cancer Analysis of Whole Genomes Consortium. <em>Pan-Cancer Analysis of Whole Genomes.</em> Nature 578, 82&#8211;93 (2020).<br>[6] Le D.T. <em>et al.</em> <em>Mismatch-repair deficiency predicts response of solid tumors to PD-1 blockade.</em> N. Engl. J. Med. 372, 2509&#8211;2520 (2015).<br>[7] Marabelle A. <em>et al.</em> <em>Tumour mutational burden and outcomes with pembrolizumab (KEYNOTE-158).</em> Ann. Oncol. 31, 131&#8211;141 (2020).<br>[8] Drilon A. <em>et al.</em> <em>Larotrectinib in TRK fusion-positive cancers.</em> N. Engl. J. Med. 378, 731&#8211;739 (2018).<br>[9] U.S. FDA. <em>Pembrolizumab for TMB-H solid tumors.</em> News Release (2020).<br>[10] King R.D. <em>et al.</em> <em>The Automation of Science.</em> Science 324, 85&#8211;89 (2009).<br>[11] MacLeod B.P. <em>et al.</em> <em>Self-driving laboratory for accelerated materials discovery.</em> Matter 3, 613&#8211;628 (2020).<br>[12] Appel K. &amp; Haken W. <em>Every planar map is four colorable.</em> Illinois J. Math. 21, 429&#8211;567 (1977).<br>[13] Gonthier G. <em>Formal Proof &#8212; The Four-Color Theorem.</em> Notices Amer. Math. Soc. 55, 1382&#8211;1393 (2008).<br>[14] Alderson-Day B. &amp; Fernyhough C. <em>Inner Speech: Development, Cognitive Functions, Phenomenology, and Neurobiology.</em> Psychol. Bull. 141, 931&#8211;965 (2015).<br>[15] Carruthers P. <em>The Cognitive Functions of Language.</em> Behav. Brain Sci. 25, 657&#8211;726 (2002).<br>[16] Lupyan G. &amp; Clark A. <em>Words and the World.</em> Curr. Dir. Psychol. Sci. 24, 279&#8211;284 (2015).<br>[17] Wang X. <em>et al.</em> <em>Self-Consistency Improves Chain-of-Thought Reasoning.</em> arXiv:2203.11171 (2022).<br>[18] Yao S. <em>et al.</em> <em>Tree of Thoughts: Deliberate Problem Solving with LLMs.</em> arXiv:2305.10601 (2023).<br>[19] OpenAI. <em>Learning to Reason with LLMs (o1).</em> Technical Report (2024).<br>[20] Global Alliance for Genomics and Health. <em>Framework for Responsible Sharing of Genomic and Health-Related Data.</em> (2018).<br>[21] Bycroft C. <em>et al.</em> <em>The UK Biobank Resource with Deep Phenotyping and Genomic Data.</em> Nature 562, 203&#8211;209 (2018).<br>[22] Gatenby R.A. <em>et al.</em> <em>Adaptive Therapy.</em> Cancer Res. 69, 4894&#8211;4903 (2009).<br>[23] Gatenby R.A. &amp; Brown J.S. <em>Evolution and Ecology of Resistance in Cancer Therapy.</em> Cold Spring Harb. Perspect. Med. 8, a033415 (2018).<br>[24] Stanford HAI. <em>AI Index Report 2024.</em> (2024).<br>[25] Epoch AI. <em>Trends in AI Training Compute and Costs.</em> (2024).<br>[26] CERN. <em>Convention and Governance.</em> (1953 &#8211; present).<br>[27] ITER Organization. <em>Agreement on the Establishment of the ITER Project.</em> (2006).<br>[28] The Global Fund. <em>Results Report.</em> (2025).<br>[29] Gavi, the Vaccine Alliance. <em>Governance and Board Resources.</em> (2025).<br>[30] UNESCO. <em>Recommendation on Open Science.</em> (2021).<br>[31] World Health Organization. <em>Ethics and Governance of Artificial Intelligence for Health.</em> (2021).<br>[32] UK Government. <em>The Bletchley Declaration.</em> (2023).<br>[33] U.S. HHS. <em>Screening Framework Guidance for Providers of Synthetic Double-Stranded DNA.</em> (2010).<br>[34] International Gene Synthesis Consortium. <em>Harmonized Screening Protocol v2.0.</em> (2017).<br>[35] Patterson D. <em>et al.</em> <em>The Carbon Footprint of Machine Learning Training Will Plateau, Then Shrink.</em> IEEE Computer 55, 18&#8211;28 (2022).<br>[36] Masanet E. <em>et al.</em> <em>Recalibrating Global Data Center Energy-Use Estimates.</em> Science 367, 984&#8211;986 (2020).</p><div><hr></div><p><strong>Note:</strong> This essay is adapted from a scholarly draft intended for submission as a <em>Nature Comment</em>; it has not been peer-reviewed.</p>]]></content:encoded></item><item><title><![CDATA[If AI is a GPT, Are We Missing a Complement?]]></title><description><![CDATA[Examining Current AI Discourse from a History-Informed Approach]]></description><link>https://abhinavmadahar.com/p/if-ai-is-a-gpt-are-we-missing-a-complement</link><guid isPermaLink="false">https://abhinavmadahar.com/p/if-ai-is-a-gpt-are-we-missing-a-complement</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Thu, 03 Jul 2025 09:07:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/92ce1d8f-d8a1-4659-a2e0-e01d0492ffce_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Abstract</h2><p>Artificial intelligence is plausibly a general&#8209;purpose technology (GPT), and the current AI discourse overlooks at least one potentially decisive complement: the ability to scale continuously-available, low&#8209;carbon electricity.  Drawing on nineteenth&#8209;century mis&#8209;predictions&#8212;Qing vs.&#8239;Tokugawa and Ottoman/Turkish vs.&#8239;Meiji&#8212;this paper argues that <strong>rapidly deployable nuclear&#8209;fission baseload, especially via small&#8209;modular reactors (SMRs), may determine which economies can translate frontier AI into broad productivity gains</strong>.  Empirical power&#8209;curve data show frontier&#8209;model training demand doubling roughly every year and inference demand doubling every three years, implying a plausible multi&#8209;hundred&#8209;terawatt&#8209;hour baseload gap by the early 2030s&#8212;far beyond the scaling potential of renewables&#8209;plus&#8209;storage.  Mainland China&#8217;s 150&#8239;GW nuclear pipeline illustrates how overlooked complements can invert the rankings suggested by GPU or talent metrics alone, outpacing the United States despite its semiconductor lead.</p><p>The analysis concludes that states should hedge complement uncertainty by fast&#8209;tracking SMR licensing and finance alongside other policy initiatives.  Otherwise, today&#8217;s AI&#8209;readiness dashboards may prove as myopic as nineteenth&#8209;century trade&#8209;balance tables that failed to foresee Japan&#8217;s industrial leapfrogging.</p><h2>1. Introduction</h2><p>A <strong>general&#8209;purpose technology&#8239;(GPT)</strong> is a breakthrough that, like the steam engine or electrification, finds applications across many sectors, improves rapidly, and demands complementary investments&#8212;skills, infrastructure, business models&#8212;before its full economic impact is felt.  Artificial intelligence&#8212;especially its prospective culmination in artificial general intelligence&#8239;(AGI)&#8212;fits this definition: it is already seeping into code writing, molecular design, and logistics, its capabilities double on an annual cadence, and it calls for costly complements.</p><p>Current AI&#8209;readiness conversations typically underline a familiar constellation of complements: frontier&#8209;class GPU clusters and specialised accelerators; vast, well&#8209;curated proprietary data sets; deep pools of machine&#8209;learning talent (often buttressed by liberal immigration regimes); clear regulatory frameworks for data privacy, liability, and safety evaluation; and ample venture or sovereign capital capable of underwriting billion&#8209;dollar training runs. These are crucial&#8212;but they share an unstated assumption: <strong>electricity supply can scale up as necessary and does not form a bottleneck.</strong>  This paper argues that <em>nuclear fission capacity</em>, especially via small&#8209;modular reactors, may prove a more decisive complement&#8212;yet it is largely absent from current readiness indices and investment road&#8209;maps.</p><h2>2. General&#8209;Purpose Technologies</h2><p>General&#8209;purpose technologies (GPTs) are foundational innovations whose impact radiates far beyond any one sector.  Economists Bresnahan and Trajtenberg (1995) define a GPT by three traits:</p><ol><li><p><strong>Pervasiveness</strong>&#8212;the technology can be applied across many industries (e.g. steam engines powered mines, factories, and railways).</p></li><li><p><strong>Continuous technical improvement</strong>&#8212;the core technology keeps getting cheaper and more capable, making new uses viable (electric motors from Edison dynamos to micro&#8209;drives).</p></li><li><p><strong>Wide&#8209;ranging complementarities</strong>&#8212;real benefits emerge only when firms and governments invest in co&#8209;inventions such as new skills, business models, or infrastructure.</p></li></ol><p>Classic GPTs include the Watt steam engine, electrification, and digital computing.  Each triggered growth waves only after decades of complementary investment: factory re&#8209;layout for electric motors, interstate highways for automobiles, organisational IT restructuring for computers.</p><p>Artificial intelligence&#8212;especially its potential culmination in artificial general intelligence (AGI)&#8212;exhibits the same signature.  Large language models already play roles in code generation, drug design, and logistics; capabilities improve on an annual cadence; and they demand costly complements ranging from specialised chips to data&#8209;governance regimes.  Whether AGI becomes transformative therefore hinges less on raw algorithmic progress than on which societies assemble the right <strong>hidden complements</strong>.  The sections that follow show how past observers repeatedly mis&#8209;identified those complements&#8212;and why energy&#8209;system readiness may prove the overlooked GPT enabler today.</p><h2>3. Historical Mis&#8209;Specification</h2><p><strong>Tokugawa Japan vs. High&#8209;Qing Mainland China (c.&#8239;1800)</strong><br>The transformative GPT on the horizon in 1800 was the <strong>steam&#8209;powered factory system</strong>&#8212;initially textile spindles and iron rolling mills.  European merchants assumed the Qing Empire, with its gigantic domestic market and entrenched silver flows, would adopt industrial steam first.  Yet two <em>unseen complements</em> were brewing in Tokugawa Japan:</p><ul><li><p><strong>Mass literacy and numeracy.</strong>  By 1800 as many as <strong>40 % of commoners</strong> attended <em>terakoya</em> schools; printed <em>kanbun</em> primers circulated through commercial lending libraries.  That literacy pool became essential for running and maintaining imported steam machinery and for double&#8209;entry cloth&#8209;mill accounting.</p></li><li><p><strong>Proto&#8209;financial depth.</strong>  Osaka&#8217;s rice&#8209;futures exchange&#8212;arguably the world&#8217;s first standardised commodities market&#8212;provided credit instruments, forward contracts and risk&#8209;transfer mechanisms that lowered capital costs for mechanised mills.</p></li></ul><p>Qing China, despite larger factor endowments, lacked both complements: broad&#8209;based literacy remained confined to the exam&#8209;elite; merchant remittance networks were powerful but regionally siloed.  When Western iron steamers arrived, <strong>Japan could mobilise talent and finance in a decade; the Qing could not.</strong> By 1890 Meiji cotton&#8209;spindle capacity exceeded China&#8217;s threefold, and Japan fielded steam&#8209;powered warships that defeated the Beiyang Fleet.</p><p><strong>Meiji Japan vs. Ottoman/Turkish Reforms (1868&#8211;1923)</strong><br>By the 1880s the next GPT wave&#8212;<strong>late&#8209;steam railways, telegraphy and early electrification</strong>&#8212;required both engineering talent and centralised fiscal capacity.  European economists wagered the Ottoman Empire would industrialise alongside Japan: it enjoyed access to London capital markets and was already importing British locomotives.  Two mis&#8209;specified complements upended that forecast:</p><ul><li><p><strong>Universal elementary schooling.</strong>  Japan&#8217;s 1872 <em>Gakusei</em> law pushed enrolment above <strong>90 % by 1905</strong>, creating a broad technician and clerk class that could absorb telegraph codes, blueprint reading and dynamo maintenance.  Ottoman primary enrolment remained under <strong>20 %</strong> until the Young Turk era, limiting the skilled workforce.</p></li><li><p><strong>Cohesive fiscal centralisation.</strong>  The Meiji state overhauled tax collection, created the Bank of Japan (1882), and channelled surplus into <em>zaibatsu</em> industrial combines.  Ottoman finance remained fragmented across provinces and subject to European debt control, choking railway extensions and delaying electrified streetcar concessions.</p></li></ul><p>By 1913 Japan&#8217;s factory output was growing <strong>~9 % annually</strong>, while Ottoman industrial share stagnated.  Agrarian exports, not electrified factories, dominated Turkish GDP through the 1920s.</p><p><strong>Lesson.</strong> Historically, observers have repeatedly failed to identify the <em>actual</em> complements a GPT would require.  That historical blind spot warns us that today&#8217;s overlooked complement may be something as unglamorous&#8212;but indispensable&#8212;as a bunch of merchants in Osaka talking rice. </p><div><hr></div><h2>4. Complement Uncertainty Today</h2><blockquote><p><em>&#8220;If you look at Chinese power production versus U.S. power production, the Chinese graph is straight up and to the right &#8230; AI capabilities ultimately depend on incredible amounts of energy.&#8221;</em> &#8212; <strong>Alexandr Wang</strong> (CSIS 2025)</p></blockquote><h3>4.1 Why electricity moves from cost variable to hard constraint</h3><p>The transformative nature of AI makes decade&#8209;long forecasts especially treacherous: language&#8209;model <strong>inference demand depends on how deeply AGI capabilities become woven into the wider economy&#8212;an adoption curve that is itself hard to forecast ex&#8239;ante&#8212;so its decade&#8209;scale growth is intrinsically uncertain</strong>, whereas the <strong>training cost of frontier models follows a clearer hardware&#8209;scaling trajectory that analysts have tracked for half a decade.</strong>  That distinction shapes the stress&#8209;test below. Recent studies separate AI workloads into two very different energy trajectories:</p><div id="datawrapper-iframe" class="datawrapper-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://datawrapper.dwcdn.net/8mqiz/1/&quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c008890-ff78-44f1-a1e2-2eb07feb9b09_1260x660.png&quot;,&quot;thumbnail_url_full&quot;:&quot;&quot;,&quot;height&quot;:274,&quot;title&quot;:&quot;Forecasting AI Energy Demand a Decade from Today&quot;,&quot;description&quot;:&quot;Create interactive, responsive &amp; beautiful charts &#8212; no code required.&quot;}" data-component-name="DatawrapperToDOM"><iframe id="iframe-datawrapper" class="datawrapper-iframe" src="https://datawrapper.dwcdn.net/8mqiz/1/" width="730" height="274" frameborder="0" scrolling="no"></iframe><script type="text/javascript">!function(){"use strict";window.addEventListener("message",(function(e){if(void 0!==e.data["datawrapper-height"]){var t=document.querySelectorAll("iframe");for(var a in e.data["datawrapper-height"])for(var r=0;r<t.length;r++){if(t[r].contentWindow===e.source)t[r].style.height=e.data["datawrapper-height"][a]+"px"}}}))}();</script></div><p>Mainstream forecasts&#8212;such as the IEA&#8217;s 2024 <em>Electricity</em> outlook, which projects data&#8209;centre use merely doubling by 2030&#8212;rely on <strong>linear extrapolations of efficiency gains</strong> and assume limited adoption breadth.  Our stress&#8209;test suggests those models may <strong>severely underestimate</strong> long&#8209;run electricity demand because they omit two compounding drivers: (i) scaling laws that push every new frontier model up by an order of magnitude and (ii) second&#8209;order economic diffusion once AGI becomes a genuinely general&#8209;purpose technology.  If either factor materialises at half the historical rate observed in early LLM deployments, aggregate load would already exceed IEA&#8217;s &#8216;high&#8217; scenario by the early 2030s.</p><p>By 2035 AI could plausibly require a continuous electricity flow equivalent to <strong>30 % of all 2024 global generation</strong>&#8212;most of it 24&#8239;/&#8239;7 baseload for multi&#8209;week training runs.</p><h3>4.2 Why only nuclear fission can close the 24&#8239;/&#8239;7 gap in the 2030s window</h3><ul><li><p><strong>Intermittent renewables</strong> require massive over&#8209;build plus long&#8209;duration storage to deliver round&#8209;the&#8209;clock power.  Delivering a steady 1&#8239;GW load from solar&#8209;wind hybrids typically demands 3&#8209;5&#8239;GW of nameplate capacity <strong>plus</strong> 8&#8209;10&#8239;hours of battery or pumped&#8209;hydro storage, pushing levelised costs above $120&#8239;MWh&#8315;&#185; (IEA 2024).  Scaling that to multi&#8209;gigawatt AI campuses compounds mineral bottlenecks&#8212;lithium, nickel, rare&#8209;earth magnets&#8212;already stressed by EV demand.</p></li><li><p><strong>Fossil generation</strong> can deliver 24&#8239;/&#8239;7 power today, but doing so at scale clashes with tightening carbon&#8209;pricing regimes and statutory net&#8209;zero targets.  At a CO&#8322; price of <strong>US&#8239;$60&#8211;100&#8239;t&#8315;&#185;</strong> (already legislated in the EU ETS and proposed in several U.S. states), unabated gas&#8209;turbine electricity exceeds <strong>US&#8239;$110&#8239;MWh&#8315;&#185;</strong>&#8212;on par with or costlier than SMRs, while adding &gt;&#8239;400&#8239;g&#8239;CO&#8322;&#8239;kWh&#8315;&#185;.  Carbon&#8209;capture retrofits can cut emissions by 90&#8239;% but add another $25&#8209;40&#8239;MWh&#8315;&#185; and consume ~15&#8239;% of plant output, further eroding economic viability.  Consequently, most G20 decarbonisation road&#8209;maps leave little headroom for new unabated fossil capacity in the 2030s, making it a politically and financially risky bet for long&#8209;term AI baseload.</p></li><li><p><strong>Hydro&#8239;&amp;&#8239;geothermal</strong> can offer true baseload in a handful of locations&#8212;think Iceland&#8217;s geothermal datacentres or Norway&#8217;s surplus hydro&#8212;but their global expansion potential is severely limited.  The best hydro sites are already dammed; new large dams face ecological&#8209;social pushback, and prospective capacity adds only ~70&#8239;GW worldwide by 2035 (IEA&#8239;2024).  Geothermal is constrained by favourable geology; even generous DOE estimates put scalable flash&#8209;steam potential below 200&#8239;GW globally&#8212;&lt;&#8239;3&#8239;% of the stress&#8209;test AI load.  Crucially, most of the regions where <em>large&#8209;scale AI compute centres</em> are clustering&#8212;Pacific Northwest cloud zones, Northern&#8239;Virginia&#8217;s "Data&#8209;Center Alley," Phoenix&#8217;s solar&#8209;fed campuses, Dublin and Frankfurt in the EU, and the Seoul&#8209;Incheon and Shenzhen&#8209;Guangzhou megaregions in Asia&#8212;possess little untapped hydro headroom and minimal high&#8209;enthalpy geothermal resources.  The compute hubs cannot count on local renewable baseload, sharpening the case for deploy&#8209;anywhere fission.</p></li><li><p><strong>Small&#8209;modular reactors (SMRs)</strong> offer a fundamentally different nuclear&#8209;build paradigm:</p><ul><li><p><strong>Factory&#8209;fabricated modules.</strong> Pressure vessels, steam generators, and containment systems are built on dedicated production lines, shipped by rail or barge, and installed &#8220;plug&#8209;and&#8209;play&#8221; on site&#8212;cutting quality&#8209;control risks that plagued gigawatt&#8209;scale stick&#8209;built reactors.</p></li><li><p><strong>Shorter schedules.</strong> Standardised modules plus reduced on&#8209;site civil works target <strong>36&#8211;48&#8239;months</strong> from first concrete to grid&#8209;sync for FOAK units; clone copies could drop to <strong>&lt; 30&#8239;months</strong>.</p></li><li><p><strong>High capacity factor (&#8805;&#8239;90&#8239;%) and load&#8209;following.</strong> Passive safety systems (natural&#8209;circulation core cooling, gravity&#8209;fed ECCS) allow flexible output without costly fossil peakers, making SMRs ideal for variable AI&#8209;training loads that nonetheless require 24&#8239;/&#8239;7 availability.</p></li><li><p><strong>10&#8211;300&#8239;MW electric blocks.</strong> A single 77&#8239;MW NuScale VOYGR or 100&#8239;MW GE&#8209;Hitachi BWRX&#8209;300 module can be co&#8209;located with a hyperscale datacentre; sites can scale by adding more pods as demand grows&#8212;mirroring cloud &#8220;availability&#8209;zone&#8221; doctrine.</p></li><li><p><strong>Minimal exclusion zones.</strong> Integrated containment and passive decay&#8209;heat removal shrink emergency&#8209;planning zones to the plant fence line (&lt;&#8239;1&#8239;km), unlocking brownfield siting at retired coal plants or even industrial parks.</p></li><li><p><strong>Competitive cost trajectory.</strong> DOE Cost Study (2024) projects <strong>US&#8239;$75&#8211;90&#8239;MWh&#8315;&#185;</strong> LCOE for nth&#8209;of&#8209;a&#8209;kind SMRs by early 2030s&#8212;20&#8211;40&#8239;% below dispatchable renewables + storage and ~30&#8239;% below unabated gas at $80&#8239;t&#8209;CO&#8322;.</p></li><li><p><strong>Environmental profile.</strong>  Life&#8209;cycle analyses show SMRs emit <strong>&lt;&#8239;15&#8239;g CO&#8322;&#8209;eq&#8239;kWh&#8315;&#185;</strong>&#8212;on par with wind and an order of magnitude below gas with CCS (UNSCEAR 2016; IPCC 2022).  Land footprint is tiny: a 300&#8239;MW SMR campus fits in ~12 ha versus &gt;&#8239;5&#8239;000 ha of solar&#8209;plus&#8209;storage for the same firm output.  Cooling demand can be air&#8209;based or use recycled municipal water, mitigating freshwater constraints.  High&#8209;level waste volume per kWh is comparable to large reactors, but smaller core inventories enable on&#8209;site dry&#8209;cask storage for decades, buying time for eventual deep&#8209;geologic repositories.</p></li></ul></li></ul><p>These traits map cleanly onto AI&#8209;compute needs: modularity matches datacentre scaling, high uptime suits multi&#8209;week training runs, small footprint eases siting near load, and low life&#8209;cycle carbon aligns with net&#8209;zero mandates.</p><h3>4.3 Fission as a Hidden GPT Complement</h3><p>The examples above imply an unsettling parallel with the 19th&#8209;century cases: observers in 1800 tracked Qing trade volumes yet missed Tokugawa literacy; planners in 1880 prized Ottoman access to European capital yet overlooked universal schooling.  In the same vein, <strong>rapidly deployable fission capacity may constitute the overlooked complement that determines which states can transform AI from laboratory showcase to economy&#8209;wide GPT</strong>.  If the stress&#8209;test electricity gap materialises and only a subset of nations can bring SMRs online at scale&#8212;due to supply&#8209;chain depth, licensing agility, or public&#8209;finance tools&#8212;those nations could enjoy a multi&#8209;year window of compute abundance that accelerates domestic productivity, defence applications, and AI research itself.  The complement is &#8220;hidden&#8221; not because energy is unfamiliar, but because today&#8217;s readiness frameworks treat energy as a price variable, not a binding technical pre&#8209;condition.</p><p>Mainstream projections&#8212;IEA&#8217;s &#8220;data&#8209;centre demand doubles&#8221; case, Goldman Sachs&#8217; 3&#215; scenario&#8212;<strong>anchor policy debates</strong>, yet they almost certainly <strong>low&#8209;ball plausible electricity demand</strong> because they assume linear efficiency gains and limited GPT diffusion.  When forecasters under&#8209;shoot the requirement by an order of magnitude, energy is framed as a <em>cost</em> variable, not a <em>capacity</em> constraint.</p><p>That framing means the ability to <strong>rapidly scale fission baseload&#8212;particularly via SMRs&#8212;plausibly remains a complement whose very </strong><em><strong>status</strong></em><strong> as a complement is under&#8209;recognized ex&#8239;ante.</strong>  The blind spot rhymes with 19th&#8209;century mis&#8209;specifications: observers prized Qing trade volumes yet missed Meiji literacy, or hailed Ottoman capital inflows while overlooking universal schooling.  If nuclear agility turns out to be decisive, today&#8217;s readiness dashboards will look as myopic tomorrow as those earlier benchmarks appear in hindsight.</p><div><hr></div><h2>5. Illustrative 2025 Contrast: Mainland China vs. United States</h2><p>Mainland China and the United States appear neck&#8209;and&#8209;neck on most headline AI&#8209;readiness metrics&#8212;GPU clusters, talent pipelines, venture funding&#8212;but they diverge sharply on their ability to deliver <strong>continuous, low&#8209;carbon baseload</strong> in the crucial 2025&#8209;2035 window.</p><div id="datawrapper-iframe" class="datawrapper-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://datawrapper.dwcdn.net/XNfKw/1/&quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bb24d24d-bae7-4d03-b9a4-601eb999a1c8_1260x660.png&quot;,&quot;thumbnail_url_full&quot;:&quot;&quot;,&quot;height&quot;:538,&quot;title&quot;:&quot;Comparison between Mainland China and the United States&quot;,&quot;description&quot;:&quot;Create interactive, responsive &amp; beautiful charts &#8212; no code required.&quot;}" data-component-name="DatawrapperToDOM"><iframe id="iframe-datawrapper" class="datawrapper-iframe" src="https://datawrapper.dwcdn.net/XNfKw/1/" width="730" height="538" frameborder="0" scrolling="no"></iframe><script type="text/javascript">!function(){"use strict";window.addEventListener("message",(function(e){if(void 0!==e.data["datawrapper-height"]){var t=document.querySelectorAll("iframe");for(var a in e.data["datawrapper-height"])for(var r=0;r<t.length;r++){if(t[r].contentWindow===e.source)t[r].style.height=e.data["datawrapper-height"][a]+"px"}}}))}();</script></div><p><strong>Implications.</strong> If fission baseload turns out to be the hidden complement that gates AGI scale, Beijing could unlock tens of GW of new 24&#8239;/&#8239;7 power by early 2030s, while Washington might still be waiting for FOAK SMRs to clear regulatory and funding hurdles.  That gap mirrors the 19th&#8209;century mis&#8209;specification: outside observers bet on the Ottoman Empire&#8217;s European capital inflows, yet Japan&#8217;s overlooked literacy and fiscal cohesion let it industrialise first.  Today&#8217;s analysts may overweight U.S. semiconductor design dominance while under&#8209;weighting Mainland China&#8217;s energy&#8209;infrastructure surge.</p><h2>6. Conclusion</h2><p>History teaches a sober lesson: <strong>the complements that unlock a general&#8209;purpose technology often remain invisible until after diffusion begins</strong>.  Early&#8209;nineteenth&#8209;century merchants prized Qing silver flows and Ottoman access to London capital, yet mass literacy, fiscal cohesion, and proto&#8209;finance turned out to be the catalytic enablers in Meiji Japan.  Today&#8217;s AI&#8209;readiness dashboards risk the same error.  By treating electricity as a priced&#8209;in commodity rather than a hard technical prerequisite, they under&#8209;estimate a looming constraint that even optimistic energy outlooks do not capture.</p><p>If the empirically observed 12&#8209;month doubling in training power and three&#8209;year doubling in inference load persist&#8212;even at half speed&#8212;AI workloads alone could require <strong>hundreds of terawatt&#8209;hours of continuous, low&#8209;carbon baseload within a decade</strong>.  Intermittent renewables with storage, geographically capped hydro, and politically constrained fossil generation cannot scale quickly or cleanly enough.  <strong>Small&#8209;modular fission reactors are, for now, the only technology offering grid&#8209;agnostic, 90&#8239;%&#8209;uptime power at a levelised cost below dispatchable renewables.</strong>  Yet not one major readiness index tracks SMR licensing agility, supply&#8209;chain depth, or nuclear&#8209;finance instruments.</p><p>The policy implication is straightforward.  States that wish to lead&#8212;or simply remain competitive&#8212;in an AGI world must adopt <strong>portfolio hedging</strong> against complement uncertainty.  Rapid fission build&#8209;out should sit alongside chip subsidies, data&#8209;governance reforms, and talent visas.  Governments that begin permitting, financing, and standardising SMRs today will create an option value that those who wait cannot replicate on short notice.</p><p>In the nineteenth century, literacy and proto&#8209;finance were the hidden complements that let late&#8209;arriving Japan overtake larger polities.  In the twenty&#8209;first, <strong>scalable nuclear baseload may play the same quiet, decisive role</strong>.  </p><div><hr></div><h2>References</h2><p>Bresnahan, Timothy F., and Manuel Trajtenberg. 1995. &#8220;General Purpose Technologies: &#8216;Engines of Growth&#8217;?&#8221; <em>Journal of Econometrics</em> 65 (1): 83&#8211;108.</p><p>Datawrapper. 2025. <em>&#8220;Nuclear-Build Readiness: Mainland China vs. United States.&#8221;</em></p><p>DOE (U.S. Department of Energy). 2024. <em>Advanced Small Modular Reactor Cost Study: nth-of-a-Kind Projections for the Early 2030s.</em> Washington, DC: Office of Nuclear Energy.</p><p>Epoch AI. 2025. <em>Compute and Power Trends Across Large AI Models, 2019-2025.</em> Epoch Policy Brief 11.</p><p>Goldman Sachs Research. 2024. <em>&#8220;The Power Behind the Cloud: AI Electricity Demand Could Triple by 2030.&#8221;</em> Global Energy Note, October 2024.</p><p>IEA. 2024. <em>Electricity 2024 &#8211; Annex: Artificial Intelligence and Data-Centre Outlook.</em> Paris: International Energy Agency.</p><p>IPCC. 2022. <em>Climate Change 2022: Mitigation of Climate Change.</em> Contribution of Working Group III to the Sixth Assessment Report. Geneva: Intergovernmental Panel on Climate Change.</p><p>OECD. 2024. <em>OECD AI Policy Observatory &#8211; Going Digital Indicators.</em> Paris: Organisation for Economic Co-operation and Development.</p><p>Oxford Insights. 2024. <em>Government AI Readiness Index 2024.</em> London: Oxford Insights.</p><p>Stanford Institute for Human-Centered AI. 2025. <em>AI Index Report 2025.</em> Stanford, CA: Stanford University.</p><p>UNSCEAR. 2016. <em>Sources, Effects and Risks of Ionizing Radiation: 2016 Report to the General Assembly.</em> New York: United Nations Scientific Committee on the Effects of Atomic Radiation.</p><p>Wang, Alexandr. 2025. &#8220;Scale AI&#8217;s Alexandr Wang on Securing U.S. AI Leadership.&#8221; Remarks at the Center for Strategic and International Studies, May 1, 2025. Transcript lines 121&#8211;124.</p><p>Wells Fargo Securities. 2024. &#8220;AI Compute Energy Demand: A First Look at the Power Curve.&#8221; Investor Note, March 2024.</p>]]></content:encoded></item><item><title><![CDATA[Constraining Malicious AGI]]></title><description><![CDATA[Exploring the Capacities of States and Non-State Actors to Develop and Use Artificial General Intelligence for Harm, and the Legal Barriers That May Limit Them]]></description><link>https://abhinavmadahar.com/p/constraining-malicious-agi</link><guid isPermaLink="false">https://abhinavmadahar.com/p/constraining-malicious-agi</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Sun, 25 May 2025 12:59:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/245df239-be8e-4e2e-b4dc-c914023083e8_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Author&#8217;s Note:</strong> This article introduces a artificial intelligence policy research question I intend to develop over the coming months: <em>to what extent do states and non-state actors differ in their capacities to create and use artificial general intelligence for malicious purposes, and how effectively can current legal regimes constrain them? </em>This is an increasingly important question to consider, but the current discourse unfortunately inadequately addresses it. My coming publications will partially fill this lacuna.</p><div><hr></div><h2><strong>Introduction</strong></h2><p>As the capabilities of artificial intelligence (AI) systems continue to advance, the prospect of artificial general intelligence (AGI) has transitioned from speculative possibility to plausible medium-term outcome. This transition invites a reconsideration of how the global system might constrain the malicious use of AGI&#8212;not merely in the abstract, but across the concrete capacities of actors who may seek to develop and deploy it. While international discourse often emphasizes state behavior, non-state actors&#8212;from research collectives to organized illicit networks&#8212;may also possess the capabilities necessary for AGI creation and use. Understanding these actors&#8217; comparative capacities, the legal structures constraining them, and the gaps therein, is critical to effective governance in the years ahead.</p><div><hr></div><h2><strong>Background: Artificial General Intelligence</strong></h2><p>Artificial intelligence refers to the field of creating machines capable of solving problems that require reasoning. Since the mid-twentieth century, this field has undergone several paradigm shifts, with the multilayered perceptron marking an early milestone. The modern era of AI began in earnest with the 2012 demonstration by Krizhevsky et al. that increasing model size yields large performance gains, inaugurating the era of scaled deep learning.</p><p>AGI refers to systems capable of solving nearly all problems that humans can solve using reasoning. In contrast to narrow AI&#8212;designed for specialized tasks such as image recognition or translation&#8212;AGI is intended to perform across domains and adapt to new problems. Recent progress toward AGI is largely driven by large language models (LLMs), which integrate symbolic reasoning, memory, and latent concept formation in a single architecture. These models are trained on massive corpora using transformer architectures and have shown generalization capacity beyond initial expectations.</p><p>Developing AGI entails challenges across four domains: scientific (developing new architectures and optimization techniques), engineering (operationalizing and scaling frontier techniques), data (curating, processing, and aligning high-quality training sets), and compute (securing the hardware and energy necessary for training large models). Scientific progress is largely confined to elite research communities. Engineering implementation is more broadly accessible but still requires deep technical skill. Data and compute are limited by institutional capacity.</p><p>In the medium-term future&#8212;defined here as the next three to seven years&#8212;AGI systems are likely to first solve all closed-form expert-level problems (e.g., those in the Humanity&#8217;s Last Exam benchmark), then extend to highly structured intellectual labor (e.g., evaluating weapons design feasibility), and finally encroach upon unstructured problem-solving domains (e.g., developing operational plans for unconventional biological attacks). These capacities will not emerge uniformly, but rather across a distributed ecosystem of actors with differing capabilities and constraints.</p><div><hr></div><h2><strong>Background: Political Science and International Law</strong></h2><h3><strong>What is a state, and what is its capacity?</strong></h3><p>Under the Montevideo Convention (1933), a state is defined as a political entity with a defined territory, permanent population, governing structure, and the capacity to enter into relations with other states. Jurisprudential debate remains on whether this capacity requires institutional potential or actual functional engagement. Although the Convention is binding only on its American signatories, it is often treated as a codification of customary international law.</p><p>Alternative conceptions include the Weberian view, where the state holds a monopoly on legitimate violence, and the recognition-based positivist view, where statehood derives from international recognition. The concept of the state must also be distinguished from that of a polity&#8212;a broader term encompassing any internally governed political community. The administrative, fiscal, and military capacity associated with modern states has expanded considerably since the eighteenth and nineteenth centuries, when centralization and institutional consolidation became core to statecraft.</p><h3><strong>What is a non-state actor, and what is its capacity?</strong></h3><p>Non-state actors include all entities operating outside formal state institutions. These range from multinational corporations to criminal networks, religious movements, philanthropic foundations, and research collectives. Importantly, non-state actors are not inherently benevolent or malicious; M&#233;decins Sans Fronti&#232;res, for example, is widely regarded as a highly benevolent non-state actor.</p><p>Some non-state actors operate with substantial structural coherence and strategic sophistication. While they typically lack sovereign legal status, treaty access, or diplomatic protections (e.g., immunity), they may possess state-like capacities such as territorial control, coercion, and infrastructure. Their ability to develop AGI depends on access to compute, data, engineering talent, scientific insight, and their ability to evade regulatory oversight.</p><h3><strong>Comparative capacity analysis</strong></h3><p>States and non-state actors share several core capacities: the ability to organize labor, allocate resources, and pursue high-risk technical projects. However, states derive these from sovereignty, taxation, recognition, and institutional permanence; non-state actors rely on market position, ideological coherence, or network dynamics. States enjoy superior access to scale-dependent assets&#8212;e.g., national labs, classified datasets, and strategic infrastructure&#8212;but are often hindered by bureaucratic inertia. Non-state actors can iterate more rapidly and evade scrutiny, particularly if operating transnationally.</p><p>Legal exposure also differs: states may be shielded by sovereign immunity or enforcement asymmetries, while non-state actors face more direct coercive constraints&#8212;but also benefit from attributional ambiguity. In practice, state initiatives often depend on non-state contractors, blurring distinctions and creating hybrid dependencies.</p><h3><strong>How are their respective actions externally regulated?</strong></h3><p>International law lacks a centralized enforcement mechanism. Instead, compliance is encouraged through reputational, reciprocal, and institutional means&#8212;treaties, soft law, judicial forums, and multilateral sanctions. The Westphalian system used today, centered on the principle of anarchy&#8212;the absence of any supranational authority above states, and the presumption that no state holds formal legal authority over another&#8212;is similar in this way to other systems of international relations, such as the Sinocentric tributary system and the Islamic jurisprudential <em>siyar</em>, in that all three systems lacked centralized enforcement.</p><p>States are bound by treaties and customary norms, but enforcement is politically contingent. Non-state actors are regulated indirectly&#8212;primarily through state actions (e.g., sanctions, criminalization). AGI-related governance will likely exacerbate these structural enforcement gaps, especially given problems of jurisdiction, attribution, and multi-actor diffusion.</p><h2><strong>Literature Review: International Constraints on Malicious AGI Development and Use</strong></h2><p>International legal and normative frameworks aimed at constraining the malicious development and use of artificial general intelligence (AGI) are presently limited in both scope and enforceability. While efforts to govern artificial intelligence more broadly have accelerated in recent years, few existing instruments directly address the risks posed by general-purpose reasoning systems. This section reviews the primary international treaties, non-binding norms, and emerging institutional mechanisms intended to regulate harmful AGI use by states and non-state actors. Emphasis is placed on both the formal architecture of these regimes and their practical limitations in a world where AGI development is rapidly diffusing.</p><h3><strong>Binding Instruments and Legal Principles</strong></h3><p>The most consequential legally binding instrument to date is the Council of Europe&#8217;s <em>Framework Convention on Artificial Intelligence, Human Rights, Democracy and the Rule of Law</em> (2024). This treaty obliges signatories to ensure that AI development and deployment, including across the full system lifecycle, conforms to human rights obligations and democratic norms. Its jurisdictional reach remains regionally anchored but open to global accession. Crucially, it exempts national security and defense domains&#8212;precisely those where malicious AGI use may be most likely to arise.</p><p>Other binding legal sources derive from existing international humanitarian and criminal law. The Geneva Conventions and their Additional Protocols already govern state conduct in armed conflict, including the deployment of new technologies. Article 36 of Additional Protocol I requires legal review of novel weapons systems, a mandate applicable to AGI-enabled military tools. Nevertheless, compliance is self-administered, and enforcement is rare. Proposals for an AGI-specific arms control treaty remain speculative, though the UN General Assembly&#8217;s 2024 resolution calling for negotiations on lethal autonomous weapons (LAWS) marks a notable development. While not binding, this resolution signals growing consensus on the need for preemptive restriction of AI systems capable of causing mass harm.</p><p>In the realm of international criminal law, no specific offense of &#8220;malicious AGI development&#8221; exists. However, state or non-state actors who deploy AGI in ways that result in mass atrocity or systemic rights violations could, in principle, be held liable under existing war crimes or crimes against humanity statutes. Yet such avenues are reactive, depend on attribution, and lack preventive force.</p><h3><strong>Soft Law and Normative Frameworks</strong></h3><p>Non-binding norms have proliferated in the absence of comprehensive treaty law. UNESCO&#8217;s <em>Recommendation on the Ethics of Artificial Intelligence</em> (2021), adopted unanimously by 193 member states, articulates global principles around transparency, accountability, and non-maleficence. Similarly, the OECD AI Principles (2019), now endorsed by over 40 countries, emphasize human-centered design, safety, and robustness. These frameworks implicitly apply to AGI, although neither uses the term directly. Their persuasive authority lies in norm diffusion, peer review, and reputational pressure, rather than coercive sanction.</p><p>Multilateral declarations&#8212;such as the 2023 <em>Bletchley Park Declaration</em> on AI safety&#8212;extend these efforts into the frontier-AI domain. While voluntary, such documents increasingly reflect consensus that advanced AI systems, including potential AGI, require international coordination and risk mitigation strategies. Proposals from the UN Secretary-General&#8217;s High-Level Advisory Body on AI suggest the eventual emergence of a distributed global governance framework, though institutional design remains underdefined.</p><p>Efforts to regulate lethal autonomous weapons within the Convention on Certain Conventional Weapons (CCW) process have similarly produced guiding principles affirming the necessity of human control over use-of-force decisions. However, treaty negotiations have stalled amid geopolitical disagreement, and the potential for AGI-enhanced LAWS development remains unconstrained.</p><h3><strong>Enforceability and Structural Gaps</strong></h3><p>International legal enforceability remains weak across these regimes. Binding treaties are implemented domestically, with uneven fidelity. Non-binding norms rely on reputational incentives and voluntary compliance. There is no centralized AI oversight body equivalent to the International Atomic Energy Agency, and existing mechanisms&#8212;such as export control regimes or domestic regulatory frameworks&#8212;function in a fragmented and uncoordinated fashion.</p><p>Attribution poses a critical barrier to enforcement. As AGI systems become more autonomous and development more transnational, assigning responsibility for harms becomes increasingly complex. This is particularly true for non-state actors operating with plausible deniability or in permissive jurisdictions. Although international law holds states responsible for activities emanating from their territory, enforcement depends on proof of knowledge and effective control&#8212;both difficult to establish in distributed AGI projects.</p><p>Moreover, jurisdictional diffusion undermines regulatory coherence. An AGI trained across multiple cloud providers in different countries, developed by a decentralized research collective, and deployed via anonymized infrastructure, may elude any single legal regime. Existing doctrines of universal jurisdiction or state responsibility offer limited recourse in such scenarios.</p><h3><strong>Prospects for Future Constraint</strong></h3><p>While existing instruments fall short of comprehensive AGI governance, they lay normative groundwork for future regimes. A precautionary principle&#8212;requiring risk assessment and mitigation prior to deployment&#8212;has begun to take shape across ethical guidelines and summit declarations. Discussions around international safety standards, model evaluation protocols, and cross-border incident response mechanisms suggest emerging pathways for formalization.</p><p>Nonetheless, the current international legal order lacks the institutional density, technical capacity, and political alignment necessary to effectively constrain malicious AGI development and use. Any future regime will need to address enforcement asymmetries, enhance attribution mechanisms, and develop institutional architectures capable of monitoring high-risk AGI trajectories. Without such evolution, the gap between technological capability and legal constraint will only widen.</p><p>This literature review thus identifies both the partial scaffolding and profound insufficiency of current international frameworks for constraining malicious AGI. The burden of anticipatory governance remains unmet. Recognizing this shortfall is a necessary precondition for building effective future constraints.</p><h2><strong>Conclusion</strong></h2><p>As artificial general intelligence transitions from theoretical construct to foreseeable reality, the question of who may develop and wield such systems&#8212;and under what constraints&#8212;gains strategic urgency. This article has examined the comparative capacities of states and non-state actors to maliciously create and use AGI, as well as the external legal frameworks that currently exist to restrain such actions. States possess clear advantages in scale, infrastructure, and regulatory power, while non-state actors benefit from agility, transnational reach, and opacity. Both actor types face significant, though asymmetrical, constraints imposed by international law, norms, and treaty regimes.</p><p>Yet these constraints remain uneven, fragmented, and frequently reactive. The existing patchwork of international instruments&#8212;while foundational&#8212;does not yet rise to the level of a comprehensive AGI governance regime. Enforcement asymmetries, attribution difficulties, and normative ambiguities present systemic vulnerabilities that could be exploited by actors pursuing harmful ends. Bridging these gaps will require deliberate institutional innovation: the development of actor-specific safeguards, the refinement of legal attribution doctrines, and potentially the establishment of new treaty frameworks or monitoring institutions.</p><p>A clear understanding of which actors are most capable of&#8212;and most likely to&#8212;pursue malicious AGI trajectories is a prerequisite to meaningful intervention. Future work must build on this foundation, identifying specific technical thresholds, institutional failure modes, and enforcement architectures that can mitigate cross-actor risk. The regulatory space around AGI remains under-structured, but it need not remain so. By anticipating the shape of its challenges and the actors involved, policy can begin to move at the speed of the technologies it seeks to govern.</p>]]></content:encoded></item><item><title><![CDATA[Artificial Intelligence Pedagogy Must Incorporate Interdisciplinary Exposure to Ensure Safety]]></title><description><![CDATA[Why Training in Economics, Political Science, and Other Fields Is Essential for AI Researchers to Responsibly Anticipate the Societal Consequences of Their Research]]></description><link>https://abhinavmadahar.com/p/artificial-intelligence-pedagogy</link><guid isPermaLink="false">https://abhinavmadahar.com/p/artificial-intelligence-pedagogy</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Fri, 02 May 2025 19:07:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/623466ee-d0e2-4be5-9408-c00e7051bcc0_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I have noticed that, when computer scientists are asked the potential impacts of artificial intelligence, they rarely use frameworks from economics, political science, sociology, etc. Instead, they will give answers roughly comparable to what someone with no academic training in the relevant fields would give, e.g., &#8216;AI models will be able to replace receptionists at paper companies&#8217;, etc.</p><p><em>Prima facie</em>, I would expect exceptionally well-trained computer scientists to be able to situate their advancements into the broader ecosystem of our world, analysing, for example, how artificial general intelligence would impact the environment using a framework informed by political-science-theoretic models of green energy production, e.g., &#8216;AGI models would cause less environmental degradation in a world where states are able to engage in popularly disfavourable, but expert-supported, activities&#8212;fission production in this case.&#8217;</p><p>My guess is that this derives from the currently dominant pedagogical models we have from American-led education [1], where students are siloed into their discipline, with limited exposure to other fields. Without my informal education, for example, I would personally not be able to situate my research in artificial general intelligence within a macroeconomic framework [2].</p><p>I am very inclined towards the current trend for more interdisciplinary work. More specifically, I would strongly encourage computer science departments to require early-stage undergraduates interested in artificial intelligence to take coursework in economics and political science, focusing primarily on the fundamentals of macroeconomics and the fundamentals of democratic institutions, respectively. Do we really want our leading computer scientists to lack an understanding of how their research impacts our broader social ecosystem? Do we not want Dr Frankenstein to understand how his Adam would affect the townspeople?</p><p>[1]: Levine, Emily J. 2024. &#8216;Research &amp; Teaching: Lasting Union or House Divided?&#8217; <em>American Academy of Arts &amp; Sciences</em>, 30 May 2024. <a href="https://amacad.org/publication/daedalus/research-teaching-lasting-union-or-house-divided">https://amacad.org/publication/daedalus/research-teaching-lasting-union-or-house-divided</a>.</p><p>[2]: Though I emphasize that my ability to do so is a fraction of what someone with doctoral training in macroeconomics would be able to accomplish.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://abhinavmadahar.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Are you the type of person who enjoys the idea of being trapped in a seminar hall with a bunch of computer science nerds and political science wonks? I can&#8217;t trap you in that seminar hall, but I can recreate the vibe in your inbox.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Subscribed</p>]]></content:encoded></item><item><title><![CDATA[How Poland and Its Peers Can Advance Their Economies Using Artificial General Intelligence]]></title><description><![CDATA[How Economies on the Low End of High-Income Can Use Artificial General Intelligence to Reach the Frontier]]></description><link>https://abhinavmadahar.com/p/poland-and-agi</link><guid isPermaLink="false">https://abhinavmadahar.com/p/poland-and-agi</guid><dc:creator><![CDATA[Abhinav Madahar · अभिनव ਮਦਾਹਰ]]></dc:creator><pubDate>Tue, 22 Apr 2025 19:18:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/93676d49-172e-45d2-9430-3d024e680835_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A friend recently asked me whether Poland and its peers can meaningfully incorporate artificial general intelligence into their respective economies. This piece is adapted from my response.</p><div><hr></div><p>Yes, the Polish economy would gain from exposure to artificial general intelligence. However, the Polish economy can only expose itself to some areas in the industry of artificial general intelligence. First, let's map out the artificial general intelligence industry, examining different scientific and engineering questions asked as well as which economies are able to insert themselves in different areas.</p><p>I am not very familiar with hardware matters in artificial intelligence, e.g. the creation of more advanced GPUs which are able to perform tensor operations more quickly, etc. I will focus on the areas with which I am familiar&#8212;I recommend learning more about hardware matters from someone better versed in that area.</p><p>Not considering hardware, artificial general intelligence can be divided into three major areas<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>: the field of scientific research into more advanced techniques which will hopefully achieve artificial general intelligence; the field of engineering which applies already-known techniques to create artificial intelligence models; and the field of business, engineering, and design which uses existing artificial intelligence models to create software which solves business problems. Of these three areas, the first two are out of Poland's reach, but the third is arguably not only achievable but necessary.</p><p>The first area, the scientific field which aims to discover the techniques necessary to achieve artificial general intelligence, poses a high barrier to entry. For an economy to be competitive here, it needs a strong academic community, something incredibly difficult to foster. Germany, for example, had arguably the strongest academic community of any economy prior to the Second World War, but the war destroyed the German academic community, especially because German academics were disproportionately Jewish. Germany still has not recovered its academic edge, and there is no indication that it will.</p><p>Currently, only a handful of economies have academic communities able to make meaningful strides towards artificial general intelligence: the United States, mainland China, the United Kingdom, Canada, the Republic of Korea, and the State of Israel. Of these, the United States and mainland China are dramatically ahead of the remaining, with the United States having an edge over mainland China. As previously established, it is almost impossible for an economy to nurture an academic community competitive on the world stage, so it is not feasible for Poland to achieve this.</p><p>A similar story presents itself for the second area. Here, we consider the field which applies existing scientific knowledge about techniques towards artificial general intelligence to engineer artificial-general-intelligence-like models. As in the previous area, this area presents a high barrier to entry, one even higher than the previous. While work in the first area is present in economies such as the United Kingdom, Canada, the Republic of Korea, and the State of Israel, the second area only meaningfully sees two economies: the United States and mainland China, though the United Kingdom adds a small addition through Google DeepMind.</p><p>This barrier is arguably even higher as training frontier models now costs around one billion USD, a figure which I expect to rise rapidly. In particular, I expect that training a frontier model will cost around one hundred billion USD in five years' time. Building a technology industry and ecosystem is already incredibly difficult, but the incredible training costs associated with frontier models turn this from difficult to effectively impossible without continental support. Even if Poland and other European economies are able to attract the necessary engineering talent and nurture their respective business environments, the cost of the compute, data, and electricity necessary would require a non-negligible portion of Europe's GDP to be spent on training frontier models. While the United States and mainland China have governments prepared for this, European institutions generally have not been as friendly. Unless Europe as a whole changes course to facilitate massive, continent-wide endeavours in building frontier models, it is not possible for Poland to establish a domestic industry focused on engineering frontier artificial-general-intelligence-like models.</p><p>This brings me to the third area, which is not only much more reachable for an economy like Poland but, arguably, necessary. The landmark World Bank report <em>China 2030</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> argues that, for an economy to advance from upper-middle income to high-income and, onwards, to the economic frontier, it needs to grow what is often referred to as its 'innovation economy'. The report does not use that terminology, and I generally avoid using that term as well because it derives primarily from applied work in industry rather than from academic sources in economics and political science, but those whose expertise derives primarily from applied fields can certainly use that term if it is helpful.</p><p>The argument raised in <em>China 2030</em>, which is fairly widely held within the academic community, is that advanced economies, especially those at the economic frontier, e.g. the United States and Taiwan, derive their economic strength from their respective scientific and technological industries. That is to say, Taiwan is an advanced economy because there exist organizations on the island which make discoveries and advancements which are not made elsewhere. For example, TSMC is the world's most advanced chipmaker for CPUs; when they discover a novel technique for increasing CPU speed, for example, they are the first in the world to make this discovery, so they issue a surcharge for anyone who wants to purchase a chip they manufacture with this advancement. Through scientific and technological advancement, Taiwanese firms are able to achieve particularly-high revenues, which keeps the Taiwanese economy at the frontier. To advance from the lower levels of high-income, which is where the Polish economy is currently, to the economic frontier, Poland needs to nurture its scientific and technological industry.</p><p>Specifically, it would behoove an economy like Poland's to have a strong industry focused on creating software which applies existing artificial-general-intelligence-like models to solve business problems. The recent few Y Combinator batches have a large portion of their startups focus on applying artificial-general-intelligence-like models to, for example, improve medical insurance processing or facilitate education by streamlining the educators' toil tasks<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. An economy like Poland's is much better suited to address these problems. They require far fewer upfront capital costs, generally only the standard single-digit million USD or equivalent figures expected of a tech startup, in stark contrast to the punch-a-hole-in-your-nation's-GDP figures seen when training frontier artificial-general-intelligence-like models. Poland would benefit from a business ecosystem and industry where startups and enterprises regularly apply already-built artificial-general-intelligence-like models to create software which solves business problems. For example, perhaps a startup in Krak&#243;w will be formed which aims to use artificial-general-intelligence to help large enterprises manage their supply chain, responding to blocks and relieving bottlenecks.</p><p>From here, the next step is to consider the specifics of Poland and its economy to form policy recommendations which can facilitate the growth of a nascent Polish startup ecosystem and industry in this way.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://abhinavmadahar.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://abhinavmadahar.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This is just one useful way to structure the field; others are valid depending on the context.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>World Bank and Development Research Center of the State Council, P. R. China. <em>China 2030: Building a Modern, Harmonious, and Creative Society</em>. Washington, DC: World Bank, 2013. <a href="https://documents.worldbank.org/en/publication/documents-reports/documentdetail/781101468239669951/china-2030-building-a-modern-harmonious-and-creative-society">https://documents.worldbank.org/en/publication/documents-reports/documentdetail/781101468239669951/china-2030-building-a-modern-harmonious-and-creative-society</a>.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>I do not know if there are any examples from the recent few Y Combinator batches which aim to achieve these exact directions. These are just examples I pulled from the aether.</p><p></p></div></div>]]></content:encoded></item></channel></rss>